首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   3837篇
  免费   102篇
  国内免费   13篇
管理学   180篇
民族学   1篇
人口学   37篇
丛书文集   21篇
理论方法论   18篇
综合类   325篇
社会学   25篇
统计学   3345篇
  2024年   1篇
  2023年   21篇
  2022年   31篇
  2021年   23篇
  2020年   67篇
  2019年   145篇
  2018年   161篇
  2017年   266篇
  2016年   124篇
  2015年   79篇
  2014年   111篇
  2013年   1146篇
  2012年   344篇
  2011年   96篇
  2010年   115篇
  2009年   131篇
  2008年   117篇
  2007年   90篇
  2006年   91篇
  2005年   88篇
  2004年   74篇
  2003年   60篇
  2002年   68篇
  2001年   61篇
  2000年   57篇
  1999年   59篇
  1998年   53篇
  1997年   42篇
  1996年   23篇
  1995年   20篇
  1994年   26篇
  1993年   19篇
  1992年   23篇
  1991年   8篇
  1990年   15篇
  1989年   9篇
  1988年   17篇
  1987年   8篇
  1986年   6篇
  1985年   4篇
  1984年   12篇
  1983年   13篇
  1982年   6篇
  1981年   5篇
  1980年   1篇
  1979年   6篇
  1978年   5篇
  1977年   2篇
  1975年   2篇
  1973年   1篇
排序方式: 共有3952条查询结果,搜索用时 328 毫秒
141.
This paper addresses the problems of frequentist and Bayesian estimation for the unknown parameters of generalized Lindley distribution based on lower record values. We first derive the exact explicit expressions for the single and product moments of lower record values, and then use these results to compute the means, variances and covariance between two lower record values. We next obtain the maximum likelihood estimators and associated asymptotic confidence intervals. Furthermore, we obtain Bayes estimators under the assumption of gamma priors on both the shape and the scale parameters of the generalized Lindley distribution, and associated the highest posterior density interval estimates. The Bayesian estimation is studied with respect to both symmetric (squared error) and asymmetric (linear-exponential (LINEX)) loss functions. Finally, we compute Bayesian predictive estimates and predictive interval estimates for the future record values. To illustrate the findings, one real data set is analyzed, and Monte Carlo simulations are performed to compare the performances of the proposed methods of estimation and prediction.  相似文献   
142.
In high-dimensional linear regression, the dimension of variables is always greater than the sample size. In this situation, the traditional variance estimation technique based on ordinary least squares constantly exhibits a high bias even under sparsity assumption. One of the major reasons is the high spurious correlation between unobserved realized noise and several predictors. To alleviate this problem, a refitted cross-validation (RCV) method has been proposed in the literature. However, for a complicated model, the RCV exhibits a lower probability that the selected model includes the true model in case of finite samples. This phenomenon may easily result in a large bias of variance estimation. Thus, a model selection method based on the ranks of the frequency of occurrences in six votes from a blocked 3×2 cross-validation is proposed in this study. The proposed method has a considerably larger probability of including the true model in practice than the RCV method. The variance estimation obtained using the model selected by the proposed method also shows a lower bias and a smaller variance. Furthermore, theoretical analysis proves the asymptotic normality property of the proposed variance estimation.  相似文献   
143.
Data collected in various scientific fields are count data. One way to analyze such data is to compare the individual levels of the factor treatment using multiple comparisons. However, the measured individuals are often clustered – e.g. according to litter or rearing. This must be considered when estimating the parameters by a repeated measurement model. In addition, ignoring the overdispersion to which count data is prone leads to an increase of the type one error rate. We carry out simulation studies using several different data settings and compare different multiple contrast tests with parameter estimates from generalized estimation equations and generalized linear mixed models in order to observe coverage and rejection probabilities. We generate overdispersed, clustered count data in small samples as can be observed in many biological settings. We have found that the generalized estimation equations outperform generalized linear mixed models if the variance-sandwich estimator is correctly specified. Furthermore, generalized linear mixed models show problems with the convergence rate under certain data settings, but there are model implementations with lower implications exists. Finally, we use an example of genetic data to demonstrate the application of the multiple contrast test and the problems of ignoring strong overdispersion.  相似文献   
144.
This paper focuses on bivariate kernel density estimation that bridges the gap between univariate and multivariate applications. We propose a subsampling-extrapolation bandwidth matrix selector that improves the reliability of the conventional cross-validation method. The proposed procedure combines a U-statistic expression of the mean integrated squared error and asymptotic theory, and can be used in both cases of diagonal bandwidth matrix and unconstrained bandwidth matrix. In the subsampling stage, one takes advantage of the reduced variability of estimating the bandwidth matrix at a smaller subsample size m (m < n); in the extrapolation stage, a simple linear extrapolation is used to remove the incurred bias. Simulation studies reveal that the proposed method reduces the variability of the cross-validation method by about 50% and achieves an expected integrated squared error that is up to 30% smaller than that of the benchmark cross-validation. It shows comparable or improved performance compared to other competitors across six distributions in terms of the expected integrated squared error. We prove that the components of the selected bivariate bandwidth matrix have an asymptotic multivariate normal distribution, and also present the relative rate of convergence of the proposed bandwidth selector.  相似文献   
145.
In this paper, the kernel density estimator for negatively superadditive dependent random variables is studied. The exponential inequalities and the exponential rate for the kernel estimator of density function with a uniform version, over compact sets are investigated. Also, the optimal bandwidth rate of the estimator is obtained using mean integrated squared error. The results are generalized and used to improve the ones obtained for the case of associated sequences. As an application, FGM sequences that fulfil our assumptions are investigated. Also, the convergence rate of the kernel density estimator is illustrated via a simulation study. Moreover, a real data analysis is presented.  相似文献   
146.
The traditional classification is based on the assumption that distribution of indicator variable X in one class is homogeneous. However, when data in one class comes from heterogeneous distribution, the likelihood ratio of two classes is not unique. In this paper, we construct the classification via an ambiguity criterion for the case of distribution heterogeneity of X in a single class. The separated historical data in each situation are used to estimate the thresholds respectively. The final boundary is chosen as the maximum and minimum thresholds from all situations. Our approach obtains the minimum ambiguity with a high classification accuracy allowing for a precise decision. In addition, nonparametric estimation of the classification region and theoretical properties are derived. Simulation study and real data analysis are reported to demonstrate the effectiveness of our method.  相似文献   
147.
In survival analysis, one way to deal with non-proportional hazards is to model short-term and long-term hazard ratios. The existing model of this nature has no control over how fast the hazard ratio is changing over time. We add a parameter to the existing model to allow the hazard ratio to change over time at different speed. A nonparametric maximum likelihood approach is used to estimate the model parameters. The existing model is a special case of the extended model when the speed parameter is 0, which leads naturally to a way of testing the adequacy of the existing model. Simulation results show that there can be substantial bias in the estimation of the short-term and long-term hazard ratio if the speed parameter is fixed incorrectly at 0 rather than estimated. The extended model is fitted to three real data sets to shed new insights, including the observation that converging hazards does not necessarily imply the odds are proportional.  相似文献   
148.
In this paper, we consider inference of the stress-strength parameter, R, based on two independent Type-II censored samples from exponentiated Fréchet populations with different index parameters. The maximum likelihood and uniformly minimum variance unbiased estimators, exact and asymptotic confidence intervals and hypotheses testing for R are obtained. We conduct a Monte Carlo simulation study to evaluate the performance of these estimators and confidence intervals. Finally, two real data sets are analysed for illustrative purposes.  相似文献   
149.
Benjamin Laumen 《Statistics》2019,53(3):569-600
In this paper, we revisit the progressive Type-I censoring scheme as it has originally been introduced by Cohen [Progressively censored samples in life testing. Technometrics. 1963;5(3):327–339]. In fact, original progressive Type-I censoring proceeds as progressive Type-II censoring but with fixed censoring times instead of failure time based censoring times. Apparently, a time truncation has been added to this censoring scheme by interpreting the final censoring time as a termination time. Therefore, not much work has been done on Cohens's original progressive censoring scheme with fixed censoring times. Thus, we discuss distributional results for this scheme and establish exact distributional results in likelihood inference for exponentially distributed lifetimes. In particular, we obtain the exact distribution of the maximum likelihood estimator (MLE). Further, the stochastic monotonicity of the MLE is verified in order to construct exact confidence intervals for both the scale parameter and the reliability.  相似文献   
150.
In this paper, we consider the estimation of the three determining parameters of the efficient frontier, the expected return, and the variance of the global minimum variance portfolio and the slope parameter, from a Bayesian perspective. Their posterior distribution is derived by assigning the diffuse and the conjugate priors to the mean vector and the covariance matrix of the asset returns and is presented in terms of a stochastic representation. Furthermore, Bayesian estimates together with the standard uncertainties for all three parameters are provided, and their asymptotic distributions are established. All obtained findings are applied to real data, consisting of the returns on assets included into the S&P 500. The empirical properties of the efficient frontier are then examined in detail.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号