首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   11607篇
  免费   195篇
管理学   1672篇
民族学   54篇
人才学   3篇
人口学   1190篇
丛书文集   60篇
教育普及   1篇
理论方法论   986篇
现状及发展   1篇
综合类   136篇
社会学   5435篇
统计学   2264篇
  2023年   78篇
  2022年   60篇
  2020年   175篇
  2019年   230篇
  2018年   273篇
  2017年   365篇
  2016年   292篇
  2015年   189篇
  2014年   248篇
  2013年   1996篇
  2012年   391篇
  2011年   320篇
  2010年   262篇
  2009年   255篇
  2008年   279篇
  2007年   278篇
  2006年   251篇
  2005年   259篇
  2004年   250篇
  2003年   218篇
  2002年   251篇
  2001年   287篇
  2000年   287篇
  1999年   274篇
  1998年   183篇
  1997年   176篇
  1996年   170篇
  1995年   170篇
  1994年   139篇
  1993年   152篇
  1992年   170篇
  1991年   148篇
  1990年   164篇
  1989年   186篇
  1988年   154篇
  1987年   140篇
  1986年   150篇
  1985年   170篇
  1984年   175篇
  1983年   162篇
  1982年   126篇
  1981年   110篇
  1980年   104篇
  1979年   141篇
  1978年   89篇
  1977年   90篇
  1976年   89篇
  1975年   83篇
  1974年   83篇
  1973年   75篇
排序方式: 共有10000条查询结果,搜索用时 15 毫秒
991.
This article introduces a new model for transaction prices in the presence of market microstructure noise in order to study the properties of the price process on two different time scales, namely, transaction time where prices are sampled with every transaction and tick time where prices are sampled with every price change. Both sampling schemes have been used in the literature on realized variance, but a formal investigation into their properties has been lacking. Our empirical and theoretical results indicate that the return dynamics in transaction time are very different from those in tick time and the choice of sampling scheme can therefore have an important impact on the properties of realized variance. For RV we find that tick time sampling is superior to transaction time sampling in terms of mean-squared-error, especially when the level of noise, number of ticks, or the arrival frequency of efficient price moves is low. Importantly, we show that while the microstructure noise may appear close to IID in transaction time, in tick time it is highly dependent. As a result, bias correction procedures that rely on the noise being independent, can fail in tick time and are better implemented in transaction time.  相似文献   
992.
There is an emerging consensus in empirical finance that realized volatility series typically display long range dependence with a memory parameter (d) around 0.4 (Andersen et al., 2001; Martens et al., 2004). The present article provides some illustrative analysis of how long memory may arise from the accumulative process underlying realized volatility. The article also uses results in Lieberman and Phillips (2004, 2005) to refine statistical inference about d by higher order theory. Standard asymptotic theory has an O(n-1/2) error rate for error rejection probabilities, and the theory used here refines the approximation to an error rate of o(n-1/2). The new formula is independent of unknown parameters, is simple to calculate and user-friendly. The method is applied to test whether the reported long memory parameter estimates of Andersen et al. (2001) and Martens et al. (2004) differ significantly from the lower boundary (d = 0.5) of nonstationary long memory, and generally confirms earlier findings.  相似文献   
993.
Unreplicated factorial designs pose a difficult problem in analysis because there are no degrees of freedom left to estimate the error. Daniel [Technometrics 1 (1959), pp. 311-341] proposed an ingenious graphical method that does not require σ to be estimated. Here we try to put Daniel's method into a formal framework and lift the subjectiveness that carries. A simulation study has been conducted that shows that the proposed method behaves better than Lenth's [Technometrics 31 (1989), pp. 469-473] popular method.  相似文献   
994.
Summary.  To investigate the variability in energy output from a network of photovoltaic cells, solar radiation was recorded at 10 sites every 10 min in the Pentland Hills to the south of Edinburgh. We identify spatiotemporal auto-regressive moving average models as the most appropriate to address this problem. Although previously considered computationally prohibitive to work with, we show that by approximating using toroidal space and fitting by matching auto-correlations, calculations can be substantially reduced. We find that a first-order spatiotemporal auto-regressive (STAR(1)) process with a first-order neighbourhood structure and a Matern noise process provide an adequate fit to the data, and we demonstrate its use in simulating realizations of energy output.  相似文献   
995.
Summary.  This is a response to Stone's criticisms of the Spottiswoode report to the UK Treasury which was responding to the Treasury's request for improved methods to evaluate the efficiency and productivity of the 43 police districts in England and Wales. The Spottiswoode report recommended uses of data envelopment analysis (DEA) and stochastic frontier analysis (SFA), which Stone critiqued en route to proposing an alternative approach. Here we note some of the most serious errors in his criticism and inaccurate portrayals of DEA and SFA. Most of our attention is devoted to DEA, and to Stone's recommended alternative approach without much attention to SFA, partly because of his abbreviated discussion of the latter. In our response we attempt to be constructive as well as critical by showing how Stone's proposed approach can be joined to DEA to expand his proposal beyond limitations in his formulations.  相似文献   
996.
To capture mean and variance asymmetries and time‐varying volatility in financial time series, we generalize the threshold stochastic volatility (THSV) model and incorporate a heavy‐tailed error distribution. Unlike existing stochastic volatility models, this model simultaneously accounts for uncertainty in the unobserved threshold value and in the time‐delay parameter. Self‐exciting and exogenous threshold variables are considered to investigate the impact of a number of market news variables on volatility changes. Adopting a Bayesian approach, we use Markov chain Monte Carlo methods to estimate all unknown parameters and latent variables. A simulation experiment demonstrates good estimation performance for reasonable sample sizes. In a study of two international financial market indices, we consider two variants of the generalized THSV model, with US market news as the threshold variable. Finally, we compare models using Bayesian forecasting in a value‐at‐risk (VaR) study. The results show that our proposed model can generate more accurate VaR forecasts than can standard models.  相似文献   
997.
The basic assumption underlying the concept of ranked set sampling is that actual measurement of units is expensive, whereas ranking is cheap. This may not be true in reality in certain cases where ranking may be moderately expensive. In such situations, based on total cost considerations, k-tuple ranked set sampling is known to be a viable alternative, where one selects k units (instead of one) from each ranked set. In this article, we consider estimation of the distribution function based on k-tuple ranked set samples when the cost of selecting and ranking units is not ignorable. We investigate estimation both in the balanced and unbalanced data case. Properties of the estimation procedure in the presence of ranking error are also investigated. Results of simulation studies as well as an application to a real data set are presented to illustrate some of the theoretical findings.  相似文献   
998.
The problem considered is that of finding an optimum measurement schedule to estimate population parameters in a nonlinear model when the patient effects are random. The paper presents examples of the use of sensitivity functions, derived from the General Equivalence Theorem for D-optimality, in the construction of optimum population designs for such schedules. With independent observations, the theorem applies to the potential inclusion of a single observation. However, in population designs the observations are correlated and the theorem applies to the inclusion of an additional measurement schedule. In one example, three groups of patients of differing size are subject to distinct schedules. Numerical, as opposed to analytical, calculation of the sensitivity function is advocated. The required covariances of the observations are found by simulation.  相似文献   
999.
The paper deals with the problem of using contours as the basis for defining probability distributions. First, the most general probability densities with given contours are obtained and the particular cases of circular and elliptical contours are dealt with. It is shown that the so-called elliptically contoured distributions do not include all possible cases. Next, the case of contours defined by polar coordinates is analyzed including its simulation and parameter estimation. Finally, the case of cumulative distribution functions with given contours is discussed. Several examples are used for illustrative purposes.  相似文献   
1000.
For asymptotic posterior normality in the one-parameter cases, Weng [2003. On Stein's identity for posterior normality. Statist. Sinica 13, 495–506] proposed to use a version of Stein's Identity to write the posterior expectations for functions of a normalized quantity in a form that is more transparent and can be easily analyzed. In the present paper we extend this approach to the multi-parameter cases and compare our conditions with earlier work. Three examples are used to illustrate the application of this method.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号