全文获取类型
收费全文 | 836篇 |
免费 | 34篇 |
专业分类
管理学 | 11篇 |
人口学 | 9篇 |
丛书文集 | 4篇 |
理论方法论 | 2篇 |
综合类 | 16篇 |
社会学 | 7篇 |
统计学 | 821篇 |
出版年
2022年 | 1篇 |
2021年 | 7篇 |
2020年 | 7篇 |
2019年 | 24篇 |
2018年 | 34篇 |
2017年 | 53篇 |
2016年 | 16篇 |
2015年 | 22篇 |
2014年 | 23篇 |
2013年 | 340篇 |
2012年 | 60篇 |
2011年 | 28篇 |
2010年 | 23篇 |
2009年 | 35篇 |
2008年 | 22篇 |
2007年 | 19篇 |
2006年 | 7篇 |
2005年 | 20篇 |
2004年 | 18篇 |
2003年 | 9篇 |
2002年 | 5篇 |
2001年 | 9篇 |
2000年 | 9篇 |
1999年 | 12篇 |
1998年 | 7篇 |
1997年 | 5篇 |
1996年 | 7篇 |
1995年 | 3篇 |
1994年 | 7篇 |
1993年 | 1篇 |
1992年 | 3篇 |
1991年 | 3篇 |
1987年 | 2篇 |
1986年 | 1篇 |
1985年 | 3篇 |
1984年 | 7篇 |
1983年 | 5篇 |
1982年 | 1篇 |
1981年 | 2篇 |
1980年 | 3篇 |
1979年 | 2篇 |
1978年 | 4篇 |
1977年 | 1篇 |
排序方式: 共有870条查询结果,搜索用时 31 毫秒
81.
This article proposes new methodologies for evaluating economic models’ out-of-sample forecasting performance that are robust to the choice of the estimation window size. The methodologies involve evaluating the predictive ability of forecasting models over a wide range of window sizes. The study shows that the tests proposed in the literature may lack the power to detect predictive ability and might be subject to data snooping across different window sizes if used repeatedly. An empirical application shows the usefulness of the methodologies for evaluating exchange rate models’ forecasting ability. 相似文献
82.
In the present paper, a semiparametric maximum-likelihood-type test statistic is proposed and proved to have the same limit null distribution as the classical parametric likelihood one. Under some mild conditions, the limiting law of the proposed test statistic, suitably normalized and centralized, is shown to be double exponential, under the null hypothesis of no change in the parameter of copula models. We also discuss the Gaussian-type approximations for the semiparametric likelihood ratio. The asymptotic distribution of the proposed statistic under specified alternatives is shown to be normal, and an approximation to the power function is given. Simulation results are provided to illustrate the finite sample performance of the proposed statistical tests based on the double exponential and Gaussian-type approximations. 相似文献
83.
Milton W. Loyer 《The American statistician》2013,67(1):57-59
This article examines some improperly stated but often used textbook probability problems. Moving from a probabilistic to a statistical setting provides insight into group testing (i.e., observing only whether one or more of a group responds and not the response of each individual). Exact methods are used to construct tables showing (i) that group testing n times to estimate p can be more efficient than n individual tests even for small n and large p, (ii) optimal grouping strategies for various (n, p) combinations, and (iii) the efficiencies and biases achieved. 相似文献
84.
《The American statistician》2013,67(3):256-261
Frequency tables are often constructed on intervals of irregular width. When plotted as bar charts, the underlying true density information may be quite distorted. The majority of introductory statistics texts recommend tabulating data into intervals of equal width, but seldom caution the consequences of failing to do so. An occasional introductory text correctly emphasizes that area rather than frequency should be plotted. Nevertheless, the correctly scaled density figure is often visually less informative than one might expect, with wide bins at constant height. In many cases, the right most bin interval has no well-defined end point, making its depiction some what arbitrary. In this note, we introduce a regular histogram approximation that matches the frequencies and also minimizes a roughness criterion for visual and exploratory appeal. The resulting estimate can reveal the density structure much more clearly. We also formulate an alternative criterion that explicitly takes account of the uncertainty in the bin frequencies. 相似文献
85.
The main problem in applying the mean-variance portfolio selection consists of the fact that the first
two moments of the asset returns are unknown. In practice the optimal portfolio weights have to be estimated.
This is usually done by replacing the moments by the classical unbiased sample estimators. We provide a comparison
of the exact and the asymptotic distributions of the estimated portfolio weights as well as a sensitivity
analysis to shifts in the moments of the asset returns. Furthermore we consider several types of shrinkage
estimators for the moments. The corresponding estimators of the portfolio weights are compared with each
other and with the portfolio weights based on the sample estimators of the moments. We show how the uncertainty
about the portfolio weights can be introduced into the performance measurement of trading strategies. The
methodology explains the bad out-of-sample performance of the classical Markowitz procedures. 相似文献
86.
This paper generalizes the tolerance interval approach for assessing agreement between two methods of continuous measurement for repeated measurement data—a common scenario in applications. The repeated measurements may be longitudinal or they may be replicates of the same underlying measurement. Our approach is to first model the data using a mixed model and then construct a relevant asymptotic tolerance interval (or band) for the distribution of appropriately defined differences. We present the methodology in the general context of a mixed model that can incorporate covariates, heteroscedasticity and serial correlation in the errors. Simulation for the no-covariate case shows good small-sample performance of the proposed methodology. For the longitudinal data, we also describe an extension for the case when the observed time profiles are modelled nonparametrically through penalized splines. Two real data applications are presented. 相似文献
87.
S. J. Welham & R. Thompson 《Journal of the Royal Statistical Society. Series B, Statistical methodology》1997,59(3):701-714
Likelihood ratio tests for fixed model terms are proposed for the analysis of linear mixed models when using residual maximum likelihood estimation. Bartlett-type adjustments, using an approximate decomposition of the data, are developed for the test statistics. A simulation study is used to compare properties of the test statistics proposed, with or without adjustment, with a Wald test. A proposed test statistic constructed by dropping fixed terms from the full fixed model is shown to give a better approximation to the asymptotic χ2 -distribution than the Wald test for small data sets. Bartlett adjustment is shown to improve the χ2 -approximation for the proposed tests substantially. 相似文献
88.
Two independent random samples are drawn from two multivariate normal populations with mean vectors μ1 and μ2 and a common
variance-covariance matrix Σ. Ahmed and Saleh (1990) considered preliminary test maximum likelihood estimator (PMLTE) for estimating μ1 based on the Hotelling's T
N
2, when it is suspected that μ1=μ2. In this paper, the PTMLE based on the Wald (W), Likelihood Ratio (LR) and Lagrangian Multiplier (LM) tests are considered. Using the quadratic risk function, the conditions of superiority of the proposed estimator for departure
parameter are derived. A max-min rule for the size of the preliminary test of significance is presented. It is demonstrated
that the PTMLE based on W test produces the highest minimum guaranteed efficiencies compared to UMLE among the three test procedures. 相似文献
89.
Evaluation of trace evidence in the form of multivariate data 总被引:1,自引:0,他引:1
C. G. G. Aitken D. Lucy 《Journal of the Royal Statistical Society. Series C, Applied statistics》2004,53(1):109-122
Summary. The evaluation of measurements on characteristics of trace evidence found at a crime scene and on a suspect is an important part of forensic science. Five methods of assessment for the value of the evidence for multivariate data are described. Two are based on significance tests and three on the evaluation of likelihood ratios. The likelihood ratio which compares the probability of the measurements on the evidence assuming a common source for the crime scene and suspect evidence with the probability of the measurements on the evidence assuming different sources for the crime scene and suspect evidence is a well-documented measure of the value of the evidence. One of the likelihood ratio approaches transforms the data to a univariate projection based on the first principal component. The other two versions of the likelihood ratio for multivariate data account for correlation among the variables and for two levels of variation: that between sources and that within sources. One version assumes that between-source variability is modelled by a multivariate normal distribution; the other version models the variability with a multivariate kernel density estimate. Results are compared from the analysis of measurements on the elemental composition of glass. 相似文献
90.
In this paper we discuss different aspects of long memory behavior and applicable parametric models. We discuss the confusion that can arise when the empirical autocorrelation function decreases in a hyperbolic way. 相似文献