全文获取类型
收费全文 | 6340篇 |
免费 | 157篇 |
国内免费 | 3篇 |
专业分类
管理学 | 988篇 |
民族学 | 54篇 |
人才学 | 9篇 |
人口学 | 500篇 |
丛书文集 | 49篇 |
理论方法论 | 776篇 |
综合类 | 51篇 |
社会学 | 3273篇 |
统计学 | 800篇 |
出版年
2023年 | 38篇 |
2022年 | 29篇 |
2021年 | 46篇 |
2020年 | 114篇 |
2019年 | 163篇 |
2018年 | 196篇 |
2017年 | 199篇 |
2016年 | 202篇 |
2015年 | 142篇 |
2014年 | 163篇 |
2013年 | 995篇 |
2012年 | 222篇 |
2011年 | 242篇 |
2010年 | 189篇 |
2009年 | 155篇 |
2008年 | 193篇 |
2007年 | 217篇 |
2006年 | 200篇 |
2005年 | 222篇 |
2004年 | 194篇 |
2003年 | 169篇 |
2002年 | 167篇 |
2001年 | 113篇 |
2000年 | 151篇 |
1999年 | 122篇 |
1998年 | 110篇 |
1997年 | 104篇 |
1996年 | 95篇 |
1995年 | 84篇 |
1994年 | 107篇 |
1993年 | 92篇 |
1992年 | 96篇 |
1991年 | 64篇 |
1990年 | 55篇 |
1989年 | 57篇 |
1988年 | 68篇 |
1987年 | 53篇 |
1986年 | 48篇 |
1985年 | 56篇 |
1984年 | 67篇 |
1983年 | 54篇 |
1982年 | 60篇 |
1981年 | 51篇 |
1980年 | 51篇 |
1979年 | 44篇 |
1978年 | 33篇 |
1977年 | 31篇 |
1976年 | 46篇 |
1975年 | 26篇 |
1974年 | 35篇 |
排序方式: 共有6500条查询结果,搜索用时 15 毫秒
91.
David G. T. Denison 《Statistics and Computing》2001,11(2):171-178
Boosting is a new, powerful method for classification. It is an iterative procedure which successively classifies a weighted version of the sample, and then reweights this sample dependent on how successful the classification was. In this paper we review some of the commonly used methods for performing boosting and show how they can be fit into a Bayesian setup at each iteration of the algorithm. We demonstrate how this formulation gives rise to a new splitting criterion when using a domain-partitioning classification method such as a decision tree. Further we can improve the predictive performance of simple decision trees, known as stumps, by using a posterior weighted average of them to classify at each step of the algorithm, rather than just a single stump. The main advantage of this approach is to reduce the number of boosting iterations required to produce a good classifier with only a minimal increase in the computational complexity of the algorithm. 相似文献
92.
The Dirichlet process prior allows flexible nonparametric mixture modeling. The number of mixture components is not specified
in advance and can grow as new data arrive. However, analyses based on the Dirichlet process prior are sensitive to the choice
of the parameters, including an infinite-dimensional distributional parameter G
0. Most previous applications have either fixed G
0 as a member of a parametric family or treated G
0 in a Bayesian fashion, using parametric prior specifications. In contrast, we have developed an adaptive nonparametric method
for constructing smooth estimates of G
0. We combine this method with a technique for estimating α, the other Dirichlet process parameter, that is inspired by an
existing characterization of its maximum-likelihood estimator. Together, these estimation procedures yield a flexible empirical
Bayes treatment of Dirichlet process mixtures. Such a treatment is useful in situations where smooth point estimates of G
0 are of intrinsic interest, or where the structure of G
0 cannot be conveniently modeled with the usual parametric prior families. Analysis of simulated and real-world datasets illustrates
the robustness of this approach. 相似文献
93.
In many domains, simple forms of classification rules are needed because of requirements such as ease of use. A particularly simple form splits each variable into just a few categories, assigns weights to the categories, sums the weights for a new object to be classified, and produces a classification by comparing the score with a threshold. Such instruments are often called scorecards. We describe a way to find the best partition of each variable using a simulated annealing strategy. We present theoretical and empirical comparisons of two such additive models, one based on weights of evidence and another based on logistic regression. 相似文献
94.
95.
96.
Julian P. T. Higgins Simon G. Thompson David J. Spiegelhalter 《Journal of the Royal Statistical Society. Series A, (Statistics in Society)》2009,172(1):137-159
Summary. Meta-analysis in the presence of unexplained heterogeneity is frequently undertaken by using a random-effects model, in which the effects underlying different studies are assumed to be drawn from a normal distribution. Here we discuss the justification and interpretation of such models, by addressing in turn the aims of estimation, prediction and hypothesis testing. A particular issue that we consider is the distinction between inference on the mean of the random-effects distribution and inference on the whole distribution. We suggest that random-effects meta-analyses as currently conducted often fail to provide the key results, and we investigate the extent to which distribution-free, classical and Bayesian approaches can provide satisfactory methods. We conclude that the Bayesian approach has the advantage of naturally allowing for full uncertainty, especially for prediction. However, it is not without problems, including computational intensity and sensitivity to a priori judgements. We propose a simple prediction interval for classical meta-analysis and offer extensions to standard practice of Bayesian meta-analysis, making use of an example of studies of 'set shifting' ability in people with eating disorders. 相似文献
97.
David E. Tyler Frank Critchley Lutz Dümbgen Hannu Oja 《Journal of the Royal Statistical Society. Series B, Statistical methodology》2009,71(3):549-592
Summary. A general method for exploring multivariate data by comparing different estimates of multivariate scatter is presented. The method is based on the eigenvalue–eigenvector decomposition of one scatter matrix relative to another. In particular, it is shown that the eigenvectors can be used to generate an affine invariant co-ordinate system for the multivariate data. Consequently, we view this method as a method for invariant co-ordinate selection . By plotting the data with respect to this new invariant co-ordinate system, various data structures can be revealed. For example, under certain independent components models, it is shown that the invariant co- ordinates correspond to the independent components. Another example pertains to mixtures of elliptical distributions. In this case, it is shown that a subset of the invariant co-ordinates corresponds to Fisher's linear discriminant subspace, even though the class identifications of the data points are unknown. Some illustrative examples are given. 相似文献
98.
The Pacific Rim Library (PRL) is an initiative of the Pacific Rim Digital Library Association (PRDLA). The project began in 2006 using the OAI-PMH paradigm and now holds over 300,000 records harvested from OAI data provider libraries around the Pacific. PRL's goal is to enable the sharing of digital collections amongst PRDLA members and the world, but greater unexpected benefits have been discovered. Through mirroring their metadata, PRL increases the chance that their data will be discovered in Google and other general search engines. With its many disparate collections, PRL is not a repository for traditional information discovery and retrieval. Initially users will bounce from a Google hit to the PRL metadata record in Hong Kong and then begin an intensive search on the original site which hosts the full digital object, in Vancouver, Honolulu, Wuhan, Singapore, or other PRDLA member location. 相似文献
99.
In response surface methodology, one is usually interested in estimating the optimal conditions based on a small number of experimental runs which are designed to optimally sample the experimental space. Typically, regression models are constructed from the experimental data and interrogated in order to provide a point estimate of the independent variable settings predicted to optimize the response. Unfortunately, these point estimates are rarely accompanied with uncertainty intervals. Though classical frequentist confidence intervals can be constructed for unconstrained quadratic models, higher order, constrained or nonlinear models are often encountered in practice. Existing techniques for constructing uncertainty estimates in such situations have not been implemented widely, due in part to the need to set adjustable parameters or because of limited or difficult applicability to constrained or nonlinear problems. To address these limitations a Bayesian method of determining credible intervals for response surface optima was developed. The approach shows good coverage probabilities on two test problems, is straightforward to implement and is readily applicable to the kind of constrained and/or nonlinear problems that frequently appear in practice. 相似文献
100.
In this note we provide a counterexample which resolves conjectures about Hadamard matrices made in this journal. Beder [1998. Conjectures about Hadamard matrices. Journal of Statistical Planning and Inference 72, 7–14] conjectured that if H is a maximal m×n row-Hadamard matrix then m is a multiple of 4; and that if n is a power of 2 then every row-Hadamard matrix can be extended to a Hadamard matrix. Using binary integer programming we obtain a maximal 13×32 row-Hadamard matrix, which disproves both conjectures. Additionally for n being a multiple of 4 up to 64, we tabulate values of m for which we have found a maximal row-Hadamard matrix. Based on the tabulated results we conjecture that a m×n row-Hadamard matrix with m?n-7 can be extended to a Hadamard matrix. 相似文献