全文获取类型
收费全文 | 677篇 |
免费 | 6篇 |
国内免费 | 1篇 |
专业分类
管理学 | 36篇 |
人口学 | 7篇 |
丛书文集 | 7篇 |
理论方法论 | 4篇 |
综合类 | 127篇 |
社会学 | 17篇 |
统计学 | 486篇 |
出版年
2022年 | 4篇 |
2021年 | 1篇 |
2020年 | 7篇 |
2019年 | 19篇 |
2018年 | 17篇 |
2017年 | 34篇 |
2016年 | 19篇 |
2015年 | 17篇 |
2014年 | 23篇 |
2013年 | 226篇 |
2012年 | 48篇 |
2011年 | 24篇 |
2010年 | 15篇 |
2009年 | 29篇 |
2008年 | 20篇 |
2007年 | 17篇 |
2006年 | 13篇 |
2005年 | 16篇 |
2004年 | 11篇 |
2003年 | 12篇 |
2002年 | 11篇 |
2001年 | 8篇 |
2000年 | 5篇 |
1999年 | 10篇 |
1998年 | 6篇 |
1997年 | 11篇 |
1996年 | 6篇 |
1995年 | 4篇 |
1994年 | 7篇 |
1993年 | 5篇 |
1992年 | 11篇 |
1990年 | 4篇 |
1989年 | 4篇 |
1988年 | 3篇 |
1987年 | 1篇 |
1985年 | 3篇 |
1984年 | 4篇 |
1983年 | 2篇 |
1982年 | 4篇 |
1979年 | 1篇 |
1978年 | 1篇 |
1976年 | 1篇 |
排序方式: 共有684条查询结果,搜索用时 468 毫秒
141.
Daniel Fernández Panos Ipeirotis Tara McAllister 《Journal of applied statistics》2019,46(8):1364-1384
Crowdsourcing has become a major tool for scholarly research since its introduction to the academic sphere in 2008. However, unlike in traditional laboratory settings, it is nearly impossible to control the conditions under which workers on crowdsourcing platforms complete tasks. In the study of communication disorders, crowdsourcing has provided a novel solution to the collection of perceptual ratings of human speech production. Such ratings allow researchers to gauge whether a treatment yields meaningful change in how human listeners' perceive disordered speech. This paper will explore some statistical considerations of crowdsourced data with specific focus on collecting perceptual ratings of human speech productions. Random effects models are applied to crowdsourced perceptual ratings collected in both a continuous and binary fashion. A simulation study is conducted to test the reliability of the proposed models under differing numbers of workers and tasks. Finally, this methodology is applied to a data set from the study of communication disorders. 相似文献
142.
Applied statisticians and pharmaceutical researchers are frequently involved in the design and analysis of clinical trials where at least one of the outcomes is binary. Treatments are judged by the probability of a positive binary response. A typical example is the noninferiority trial, where it is tested whether a new experimental treatment is practically not inferior to an active comparator with a prespecified margin δ. Except for the special case of δ = 0, no exact conditional test is available although approximate conditional methods (also called second‐order methods) can be applied. However, in some situations, the approximation can be poor and the logical argument for approximate conditioning is not compelling. The alternative is to consider an unconditional approach. Standard methods like the pooled z‐test are already unconditional although approximate. In this article, we review and illustrate unconditional methods with a heavy emphasis on modern methods that can deliver exact, or near exact, results. For noninferiority trials based on either rate difference or rate ratio, our recommendation is to use the so‐called E‐procedure, based on either the score or likelihood ratio statistic. This test is effectively exact, computationally efficient, and respects monotonicity constraints in practice. We support our assertions with a numerical study, and we illustrate the concepts developed in theory with a clinical example in pulmonary oncology; R code to conduct all these analyses is available from the authors. 相似文献
143.
《Journal of Statistical Computation and Simulation》2012,82(8):731-745
A general saddlepoint/Monte Carlo method to approximate (conditional) multivariate probabilities is presented. This method requires a tractable joint moment generating function (m.g.f.), but does not require a tractable distribution or density. The method is easy to program and has a third-order accuracy with respect to increasing sample size in contrast to standard asymptotic approximations which are typically only accurate to the first order. The method is most easily described in the context of a continuous regular exponential family. Here, inferences can be formulated as probabilities with respect to the joint density of the sufficient statistics or the conditional density of some sufficient statistics given the others. Analytical expressions for these densities are not generally available, and it is often not possible to simulate exactly from the conditional distributions to obtain a direct Monte Carlo approximation of the required integral. A solution to the first of these problems is to replace the intractable density by a highly accurate saddlepoint approximation. The second problem can be addressed via importance sampling, that is, an indirect Monte Carlo approximation involving simulation from a crude approximation to the true density. Asymptotic normality of the sufficient statistics suggests an obvious candidate for an importance distribution. The more general problem considers the computation of a joint probability for a subvector of random T, given its complementary subvector, when its distribution is intractable, but its joint m.g.f. is computable. For such settings, the distribution may be tilted, maintaining T as the sufficient statistic. Within this tilted family, the computation of such multivariate probabilities proceeds as described for the exponential family setting. 相似文献
144.
Arturo J. Fernández 《Statistical Papers》2009,50(1):119-136
Trimmed samples are commonly used in several branches of statistical methodology, especially when the presence of contaminated
data is suspected. Assuming that certain proportions of the smallest and largest observations from a Weibull sample are unknown
or have been eliminated, a Bayesian approach to point and interval estimation of the scale parameter, as well as hypothesis
testing and prediction, is presented. In many cases, the use of substantial prior information can significantly increase the
quality of the inferences and reduce the amount of testing required. Some Bayes estimators and predictors are derived in closed-forms.
Highest posterior density estimators and credibility intervals can be computed using iterative methods. Bayes rules for testing
one- and two-sided hypotheses are also provided. An illustrative numerical example is included. 相似文献
145.
A singular partitioned linear model, i.e. the singular model comprising the main parameters and the nuisance parameters, can
be reduced, or transformed to the form in which only linear functions concerning main parameters are involved. In the paper
some properties of the best linear unbiased estimators of these functions following from these models are considered. 相似文献
146.
Bing Li 《Scandinavian Journal of Statistics》2001,28(4):577-602
To apply the quasi likelihood method one needs both the mean and the variance functions to determine its optimal weights. If the variance function is unknown, then the weights should be acquired from the data. One way to do so is by adaptive estimation, which involves non-parametric estimation of the variance function. Adaptation, however, also brings in noise that hampers its improvement for moderate samples. In this paper we introduce an alternative method based not on the estimation of the variance function, but on the penalized minimization of the asymptotic variance of the estimator. By doing so we are able to retain a restricted optimality under the smoothness condition, however strong that condition may be. This is important because for moderate sample sizes we need to impose a strong smoothness constraint to damp the noise—often stronger than would be adequate for the adaptive method. We will give a rigorous development of the related asymptotic theory, and provide the simulation evidence for the advantage of this method. 相似文献
147.
K. Henschke 《Statistics》2013,47(2):257-272
Using given significant additional information it is possible to improve different confidence regions for the regression parameters in a linear model. Thereby, the given informations may concern the expectation and (or) the variance of the observations, and an improvement is possible in the sense of the decrease of the confidence regions' size. In particular it is possible to improve the so called confidence ellipsoids which are often used to estimate the considered parameters. 相似文献
148.
We revisit the classic problem of estimation of the binomial parameters when both parameters n,p are unknown. We start with a series of results that illustrate the fundamental difficulties in the problem. Specifically, we establish lack of unbiased estimates for essentially any functions of just n or just p. We also quantify just how badly biased the sample maximum is as an estimator of n. Then, we motivate and present two new estimators of n. One is a new moment estimate and the other is a bias correction of the sample maximum. Both are easy to motivate, compute, and jackknife. The second estimate frequently beats most common estimates of n in the simulations, including the Carroll–Lombard estimate. This estimate is very promising. We end with a family of estimates for p; a specific one from the family is compared to the presently common estimate and the improvements in mean-squared error are often very significant. In all cases, the asymptotics are derived in one domain. Some other possible estimates such as a truncated MLE and empirical Bayes methods are briefly discussed. 相似文献
149.
通过构建包含利率、汇率、房价、股价、货币供应量和第二产业增加值比重共六个指标的FCI,以具有可变参数特征的状态空间模型为基础,拟合具有可变权重的FCI。实证结果显示:FCI在样本期内的拟合效果基本符合经济现实,用于指示金融松紧程度和经济运行方向的功能明显,2005年第3季度、2008年第3季度和2011年第3季度等时段的金融形势最为紧张,2004年第2季度和2011年第1季度的金融形势最为宽松。 相似文献
150.
The statistical model is considered in which the collection of data from several independent populations is available only at random times determined by order statistics of lifetimes of a given number of objects. Each of the populations is distributed according to a general multiparameter exponential family. The problem is to estimate the mean value vector parameter of the multiparameter exponential family of distributions of the forthcoming observations. Under the loss function involving a weighted squared error loss, the cost proportional to the events appeared and a cost of observing the process, a class of optimal sequential procedures is established. The procedures are derived in two situations: when the lifetime distribution is completely known and in the case when it is unknown but assumed to belong to an exponential subfamily with an unknown failure rate parameter. 相似文献