首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   2179篇
  免费   65篇
  国内免费   38篇
管理学   442篇
民族学   1篇
人口学   1篇
丛书文集   33篇
理论方法论   4篇
综合类   692篇
社会学   8篇
统计学   1101篇
  2024年   6篇
  2023年   11篇
  2022年   38篇
  2021年   21篇
  2020年   34篇
  2019年   64篇
  2018年   68篇
  2017年   125篇
  2016年   74篇
  2015年   63篇
  2014年   71篇
  2013年   307篇
  2012年   144篇
  2011年   97篇
  2010年   85篇
  2009年   77篇
  2008年   112篇
  2007年   113篇
  2006年   106篇
  2005年   99篇
  2004年   97篇
  2003年   64篇
  2002年   50篇
  2001年   49篇
  2000年   61篇
  1999年   51篇
  1998年   37篇
  1997年   27篇
  1996年   30篇
  1995年   21篇
  1994年   17篇
  1993年   8篇
  1992年   20篇
  1991年   6篇
  1990年   5篇
  1989年   10篇
  1988年   6篇
  1987年   3篇
  1985年   2篇
  1984年   1篇
  1981年   2篇
排序方式: 共有2282条查询结果,搜索用时 31 毫秒
991.
介绍了指纹特征点的匹配原理,提出了一种改进的实时指纹特征点匹配算法,并对算法性能进行了实验研究.给出了错误匹配率(FMR)和错误不匹配率(FNMR)随阈值变化的情况及算法的ROC曲线.得到算法的等错误率(EER)为1.8%,最小FMR(zeroFNMR)为6.8%,平均匹配时间为0.1s.算法在指纹库FVC2004上的实验结果表明,算法性能较好,适合于实时指纹识别系统.  相似文献   
992.
In this study, estimation of the parameters of the zero-inflated count regression models and computations of posterior model probabilities of the log-linear models defined for each zero-inflated count regression models are investigated from the Bayesian point of view. In addition, determinations of the most suitable log-linear and regression models are investigated. It is known that zero-inflated count regression models cover zero-inflated Poisson, zero-inflated negative binomial, and zero-inflated generalized Poisson regression models. The classical approach has some problematic points but the Bayesian approach does not have similar flaws. This work points out the reasons for using the Bayesian approach. It also lists advantages and disadvantages of the classical and Bayesian approaches. As an application, a zoological data set, including structural and sampling zeros, is used in the presence of extra zeros. In this work, it is observed that fitting a zero-inflated negative binomial regression model creates no problems at all, even though it is known that fitting a zero-inflated negative binomial regression model is the most problematic procedure in the classical approach. Additionally, it is found that the best fitting model is the log-linear model under the negative binomial regression model, which does not include three-way interactions of factors.  相似文献   
993.
This article enlarges the covariance configurations, on which the classical linear discriminant analysis is based, by considering the four models arising from the spectral decomposition when eigenvalues and/or eigenvectors matrices are allowed to vary or not between groups. As in the classical approach, the assessment of these configurations is accomplished via a test on the training set. The discrimination rule is then built upon the configuration provided by the test, considering or not the unlabeled data. Numerical experiments, on simulated and real data, have been performed to evaluate the gain of our proposal with respect to the linear discriminant analysis.  相似文献   
994.
In this paper, the semi varying coefficient zero-inflated generalized Poisson model is discussed based on penalized log-likelihood. All the coefficient functions are fitted by penalized spline (P-spline), and Expectation-maximization algorithm is used to drive these estimators. The estimation approach is rapid and computationally stable. Under some mild conditions, the consistency and the asymptotic normality of these resulting estimators are given. The score test statistics about dispersion parameter is discussed based on the P-spline estimation. Both simulated and real data example are used to illustrate our proposed methods.  相似文献   
995.
An automated (Markov chain) Monte Carlo EM algorithm   总被引:1,自引:0,他引:1  
We present an automated Monte Carlo EM (MCEM) algorithm which efficiently assesses Monte Carlo error in the presence of dependent Monte Carlo, particularly Markov chain Monte Carlo, E-step samples and chooses an appropriate Monte Carlo sample size to minimize this Monte Carlo error with respect to progressive EM step estimates. Monte Carlo error is gauged though an application of the central limit theorem during renewal periods of the MCMC sampler used in the E-step. The resulting normal approximation allows us to construct a rigorous and adaptive rule for updating the Monte Carlo sample size each iteration of the MCEM algorithm. We illustrate our automated routine and compare the performance with competing MCEM algorithms in an analysis of a data set fit by a generalized linear mixed model.  相似文献   
996.
This article describes a new Monte Carlo method for the evaluation of the orthant probabilities by sampling first passage times of a non-singular Gaussian discrete time-series across an absorbing boundary. This procedure makes use of a simulation of several time-series sample paths, aiming to record their first crossing instants. Thus, the computation of the orthant probabilities is traced back to the accurate simulation of a non-singular Gaussian discrete-time series. Moreover, if the simulation is also efficient, this method is shown to be speedier than the others proposed in the literature. As example, we make use of the Davies–Harte algorithm in the evaluation of the orthant probabilities associated to the ARFIMA(0, d, 0) model. Test results are presented that compare this method with currently available software.  相似文献   
997.
Recently, several new robust multivariate estimators of location and scatter have been proposed that provide new and improved methods for detecting multivariate outliers. But for small sample sizes, there are no results on how these new multivariate outlier detection techniques compare in terms of p n , their outside rate per observation (the expected proportion of points declared outliers) under normality. And there are no results comparing their ability to detect truly unusual points based on the model that generated the data. Moreover, there are no results comparing these methods to two fairly new techniques that do not rely on some robust covariance matrix. It is found that for an approach based on the orthogonal Gnanadesikan–Kettenring estimator, p n can be very unsatisfactory with small sample sizes, but a simple modification gives much more satisfactory results. Similar problems were found when using the median ball algorithm, but a modification proved to be unsatisfactory. The translated-biweights (TBS) estimator generally performs well with a sample size of n≥20 and when dealing with p-variate data where p≤5. But with p=8 it can be unsatisfactory, even with n=200. A projection method as well the minimum generalized variance method generally perform best, but with p≤5 conditions where the TBS method is preferable are described. In terms of detecting truly unusual points, the methods can differ substantially depending on where the outliers happen to be, the number of outliers present, and the correlations among the variables.  相似文献   
998.
The Buckley–James estimator (BJE) [J. Buckley and I. James, Linear regression with censored data, Biometrika 66 (1979), pp. 429–436] has been extended from right-censored (RC) data to interval-censored (IC) data by Rabinowitz et al. [D. Rabinowitz, A. Tsiatis, and J. Aragon, Regression with interval-censored data, Biometrika 82 (1995), pp. 501–513]. The BJE is defined to be a zero-crossing of a modified score function H(b), a point at which H(·) changes its sign. We discuss several approaches (for finding a BJE with IC data) which are extensions of the existing algorithms for RC data. However, these extensions may not be appropriate for some data, in particular, they are not appropriate for a cancer data set that we are analysing. In this note, we present a feasible iterative algorithm for obtaining a BJE. We apply the method to our data.  相似文献   
999.
We study the nonparametric maximum likelihood estimate (NPMLE) of the cdf or sub-distribution functions of the failure time for the failure causes in a series system. The study is motivated by a cancer research data (from the Memorial Sloan-Kettering Cancer Center) with interval-censored time and masked failure cause. The NPMLE based on this data set suggests that the existing masking models are not appropriate. We propose a new model called the random partition masking model, which does not rely on the commonly used symmetry assumption (namely, given the failure cause, the probability of observing the masked failure causes is independent of the failure time; see Flehinger et al. Inference about defects in the presence of masking, Technometrics 38 (1996), pp. 247–255). The RPM model is easier to implement in simulation studies than the existing models. We discuss the algorithms for computing the NPMLE and study its asymptotic properties. Our simulation and data analysis indicate that the NPMLE is feasible for a moderate sample size.  相似文献   
1000.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号