首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   119006篇
  免费   3165篇
  国内免费   1296篇
管理学   2192篇
劳动科学   27篇
民族学   1795篇
人才学   8篇
人口学   1580篇
丛书文集   16258篇
理论方法论   5161篇
综合类   87193篇
社会学   3719篇
统计学   5534篇
  2024年   198篇
  2023年   593篇
  2022年   893篇
  2021年   1003篇
  2020年   1233篇
  2019年   1178篇
  2018年   1218篇
  2017年   1525篇
  2016年   1632篇
  2015年   2212篇
  2014年   5478篇
  2013年   6841篇
  2012年   6886篇
  2011年   8237篇
  2010年   6771篇
  2009年   6955篇
  2008年   7331篇
  2007年   9326篇
  2006年   9458篇
  2005年   8797篇
  2004年   8390篇
  2003年   8140篇
  2002年   6610篇
  2001年   5477篇
  2000年   3242篇
  1999年   939篇
  1998年   466篇
  1997年   383篇
  1996年   327篇
  1995年   268篇
  1994年   207篇
  1993年   169篇
  1992年   139篇
  1991年   116篇
  1990年   66篇
  1989年   62篇
  1988年   63篇
  1987年   20篇
  1986年   27篇
  1985年   98篇
  1984年   102篇
  1983年   72篇
  1982年   67篇
  1981年   59篇
  1980年   50篇
  1979年   52篇
  1978年   55篇
  1977年   17篇
  1976年   11篇
  1975年   6篇
排序方式: 共有10000条查询结果,搜索用时 15 毫秒
991.
This article reviews symmetrical global sensitivity analysis based on the analysis of variance of high-dimensional model representation. To overcome the computational difficulties and explore the use of symmetrical design of experiment (SDOE), two methods are presented. If the form of the objective function f is known, we use SDOE to estimate the symmetrical global sensitivity indices instead of Monte Carlo or quasi-Monte Carlo simulation. Otherwise, we use the observed values of the experiment to do symmetrical global sensitivity analysis. These methods are easy to implement and can reduce the computational cost. An example is given by symmetrical design of experiment.  相似文献   
992.
Most existing reduced-form macroeconomic multivariate time series models employ elliptical disturbances, so that the forecast densities produced are symmetric. In this article, we use a copula model with asymmetric margins to produce forecast densities with the scope for severe departures from symmetry. Empirical and skew t distributions are employed for the margins, and a high-dimensional Gaussian copula is used to jointly capture cross-sectional and (multivariate) serial dependence. The copula parameter matrix is given by the correlation matrix of a latent stationary and Markov vector autoregression (VAR). We show that the likelihood can be evaluated efficiently using the unique partial correlations, and estimate the copula using Bayesian methods. We examine the forecasting performance of the model for four U.S. macroeconomic variables between 1975:Q1 and 2011:Q2 using quarterly real-time data. We find that the point and density forecasts from the copula model are competitive with those from a Bayesian VAR. During the recent recession the forecast densities exhibit substantial asymmetry, avoiding some of the pitfalls of the symmetric forecast densities from the Bayesian VAR. We show that the asymmetries in the predictive distributions of GDP growth and inflation are similar to those found in the probabilistic forecasts from the Survey of Professional Forecasters. Last, we find that unlike the linear VAR model, our fitted Gaussian copula models exhibit nonlinear dependencies between some macroeconomic variables. This article has online supplementary material.  相似文献   
993.
Mehmet Caner 《Econometric Reviews》2016,35(8-10):1343-1346
This special issue is concerned with model selection and shrinkage estimators. This Introduction gives an overview of the papers published in this special issue.  相似文献   
994.
Oracle Inequalities for Convex Loss Functions with Nonlinear Targets   总被引:1,自引:1,他引:0  
This article considers penalized empirical loss minimization of convex loss functions with unknown target functions. Using the elastic net penalty, of which the Least Absolute Shrinkage and Selection Operator (Lasso) is a special case, we establish a finite sample oracle inequality which bounds the loss of our estimator from above with high probability. If the unknown target is linear, this inequality also provides an upper bound of the estimation error of the estimated parameter vector. Next, we use the non-asymptotic results to show that the excess loss of our estimator is asymptotically of the same order as that of the oracle. If the target is linear, we give sufficient conditions for consistency of the estimated parameter vector. We briefly discuss how a thresholded version of our estimator can be used to perform consistent variable selection. We give two examples of loss functions covered by our framework.  相似文献   
995.
This article considers in-sample prediction and out-of-sample forecasting in regressions with many exogenous predictors. We consider four dimension-reduction devices: principal components, ridge, Landweber Fridman, and partial least squares. We derive rates of convergence for two representative models: an ill-posed model and an approximate factor model. The theory is developed for a large cross-section and a large time-series. As all these methods depend on a tuning parameter to be selected, we also propose data-driven selection methods based on cross-validation and establish their optimality. Monte Carlo simulations and an empirical application to forecasting inflation and output growth in the U.S. show that data-reduction methods outperform conventional methods in several relevant settings, and might effectively guard against instabilities in predictors’ forecasting ability.  相似文献   
996.
In this paper, we propose two new estimators of treatment effects in regression discontinuity designs. These estimators can aid understanding of the existing estimators such as the local polynomial estimator and the partially linear estimator. The first estimator is the partially polynomial estimator which extends the partially linear estimator by further incorporating derivative differences of the conditional mean of the outcome on the two sides of the discontinuity point. This estimator is related to the local polynomial estimator by a relocalization effect. Unlike the partially linear estimator, this estimator can achieve the optimal rate of convergence even under broader regularity conditions. The second estimator is an instrumental variable estimator in the fuzzy design. This estimator will reduce to the local polynomial estimator if higher order endogeneities are neglected. We study the asymptotic properties of these two estimators and conduct simulation studies to confirm the theoretical analysis.  相似文献   
997.
In many economic models, theory restricts the shape of functions, such as monotonicity or curvature conditions. This article reviews and presents a framework for constrained estimation and inference to test for shape conditions in parametric models. We show that “regional” shape-restricting estimators have important advantages in terms of model fit and flexibility (as opposed to standard “local” or “global” shape-restricting estimators). In our empirical illustration, this is the first article to impose and test for all shape restrictions required by economic theory simultaneously in the “Berndt and Wood” data. We find that this dataset is consistent with “duality theory,” whereas previous studies have found violations of economic theory. We discuss policy consequences for key parameters, such as whether energy and capital are complements or substitutes.  相似文献   
998.
In this paper, we focus on the problem of factor screening in nonregular two-level designs through gradually reducing the number of possible sets of active factors. We are particularly concerned with situations when three or four factors are active. Our proposed method works through examining fits of projection models, where variable selection techniques are used to reduce the number of terms. To examine the reliability of the methods in combination with such techniques, a panel of models consisting of three or four active factors with data generated from the 12-run and the 20-run Plackett–Burman (PB) design is used. The dependence of the procedure on the amount of noise, the number of active factors and the number of experimental factors is also investigated. For designs with few runs such as the 12-run PB design, variable selection should be done with care and default procedures in computer software may not be reliable to which we suggest improvements. A real example is included to show how we propose factor screening can be done in practice.  相似文献   
999.
This paper provides a Bayesian estimation procedure for monotone regression models incorporating the monotone trend constraint subject to uncertainty. For monotone regression modeling with stochastic restrictions, we propose a Bayesian Bernstein polynomial regression model using two-stage hierarchical prior distributions based on a family of rectangle-screened multivariate Gaussian distributions extended from the work of Gurtis and Ghosh [7 S.M. Curtis and S.K. Ghosh, A variable selection approach to monotonic regression with Bernstein polynomials, J. Appl. Stat. 38 (2011), pp. 961976. doi: 10.1080/02664761003692423[Taylor &; Francis Online], [Web of Science ®] [Google Scholar]]. This approach reflects the uncertainty about the prior constraint, and thus proposes a regression model subject to monotone restriction with uncertainty. Based on the proposed model, we derive the posterior distributions for unknown parameters and present numerical schemes to generate posterior samples. We show the empirical performance of the proposed model based on synthetic data and real data applications and compare the performance to the Bernstein polynomial regression model of Curtis and Ghosh [7 S.M. Curtis and S.K. Ghosh, A variable selection approach to monotonic regression with Bernstein polynomials, J. Appl. Stat. 38 (2011), pp. 961976. doi: 10.1080/02664761003692423[Taylor &; Francis Online], [Web of Science ®] [Google Scholar]] for the shape restriction with certainty. We illustrate the effectiveness of our proposed method that incorporates the uncertainty of the monotone trend and automatically adapts the regression function to the monotonicity, through empirical analysis with synthetic data and real data applications.  相似文献   
1000.
With competing risks data, one often needs to assess the treatment and covariate effects on the cumulative incidence function. Fine and Gray proposed a proportional hazards regression model for the subdistribution of a competing risk with the assumption that the censoring distribution and the covariates are independent. Covariate‐dependent censoring sometimes occurs in medical studies. In this paper, we study the proportional hazards regression model for the subdistribution of a competing risk with proper adjustments for covariate‐dependent censoring. We consider a covariate‐adjusted weight function by fitting the Cox model for the censoring distribution and using the predictive probability for each individual. Our simulation study shows that the covariate‐adjusted weight estimator is basically unbiased when the censoring time depends on the covariates, and the covariate‐adjusted weight approach works well for the variance estimator as well. We illustrate our methods with bone marrow transplant data from the Center for International Blood and Marrow Transplant Research. Here, cancer relapse and death in complete remission are two competing risks.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号