首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   3901篇
  免费   121篇
  国内免费   26篇
管理学   137篇
民族学   5篇
人才学   1篇
人口学   15篇
丛书文集   106篇
理论方法论   29篇
综合类   1195篇
社会学   34篇
统计学   2526篇
  2024年   2篇
  2023年   19篇
  2022年   8篇
  2021年   24篇
  2020年   68篇
  2019年   89篇
  2018年   130篇
  2017年   195篇
  2016年   115篇
  2015年   117篇
  2014年   131篇
  2013年   996篇
  2012年   368篇
  2011年   176篇
  2010年   165篇
  2009年   163篇
  2008年   175篇
  2007年   156篇
  2006年   127篇
  2005年   123篇
  2004年   113篇
  2003年   85篇
  2002年   68篇
  2001年   79篇
  2000年   74篇
  1999年   54篇
  1998年   31篇
  1997年   44篇
  1996年   22篇
  1995年   20篇
  1994年   13篇
  1993年   13篇
  1992年   10篇
  1991年   8篇
  1990年   11篇
  1989年   6篇
  1988年   10篇
  1987年   6篇
  1986年   3篇
  1985年   5篇
  1984年   6篇
  1983年   6篇
  1982年   4篇
  1980年   3篇
  1979年   1篇
  1978年   1篇
  1977年   2篇
  1976年   1篇
  1975年   2篇
排序方式: 共有4048条查询结果,搜索用时 625 毫秒
101.
关于上市公司利润操控的相关研究主要是将净利润作为一个事关操控的敏感指标,但在监管与被监管双方努力的博弈下,其可被操纵的潜力已被充分挖掘。从实证角度关注上市公司财务操纵的其他可操作途径,其中就包括一些看似合理的变相财务操纵手段。按照2001年中国证监会公布的《上市公司行业分类指引》,把我国上市企业分为13个行业,对不同行业内企业2003—2013年部分主观可操控的财务指标与其所在行业的平均市盈率进行格兰杰因果检验和回归分析,结果表明,一些行业内企业的部分财务指标与其所在行业的平均市盈率存在领先与滞后关系,并且影响作用显著。  相似文献   
102.
平稳性检验方法的有效性研究   总被引:2,自引:1,他引:1  
平稳性检验是时间序列分析的重要研究内容,现有检验方法的性能缺乏系统的比较分析。文章从样本长度的视角研究平稳性检验方法的性能,采用ADF检验、PP检验、KPSS检验和LMC检验四种方法展开实证研究。仿真实验结果表明:时间序列数据长度会对检验方法的准确率产生明显的影响,数据长度较小时检验准确率偏低;数据长度增大时可以提升检验方法的准确率,但仍未能达到100%的上限值。当样本长度较小时,这些方法的检验统计量的渐进分布难以满足,因此其实际检验效果值得探究。样本长度是有限的,因此渐进分布检验方式的改进空间有限,新的检验方式值得探究。  相似文献   
103.
针对塑件成型过程中体积收缩率对成型塑件尺寸精度的影响,文章选取长条薄壁板件作为研究对象,利用AMI有限元分析软件对长条薄壁板注塑成型过程进行数值模拟,采用多因素交互正交试验的方法获得常用ABS塑料在不同的工艺参数下成型薄壁件的体积收缩率。以体积收缩率为研究目标,对试验结果进行定量分析,对比分析每一个工艺参数对研究目标的贡献率,并计算得到最优的工艺参数组合,依据实验结果优化塑件成型工艺。该方法的应用为注塑成型工艺参数优化提供了定量的数据参考,是一种快速而实用的方法。  相似文献   
104.
A common approach taken in high‐dimensional regression analysis is sliced inverse regression, which separates the range of the response variable into non‐overlapping regions, called ‘slices’. Asymptotic results are usually shown assuming that the slices are fixed, while in practice, estimators are computed with random slices containing the same number of observations. Based on empirical process theory, we present a unified theoretical framework to study these techniques, and revisit popular inverse regression estimators. Furthermore, we introduce a bootstrap methodology that reproduces the laws of Cramér–von Mises test statistics of interest to model dimension, effects of specified covariates and whether or not a sliced inverse regression estimator is appropriate. Finally, we investigate the accuracy of different bootstrap procedures by means of simulations.  相似文献   
105.
Subgroup detection has received increasing attention recently in different fields such as clinical trials, public management and market segmentation analysis. In these fields, people often face time‐to‐event data, which are commonly subject to right censoring. This paper proposes a semiparametric Logistic‐Cox mixture model for subgroup analysis when the interested outcome is event time with right censoring. The proposed method mainly consists of a likelihood ratio‐based testing procedure for testing the existence of subgroups. The expectation–maximization iteration is applied to improve the testing power, and a model‐based bootstrap approach is developed to implement the testing procedure. When there exist subgroups, one can also use the proposed model to estimate the subgroup effect and construct predictive scores for the subgroup membership. The large sample properties of the proposed method are studied. The finite sample performance of the proposed method is assessed by simulation studies. A real data example is also provided for illustration.  相似文献   
106.
A practical problem with large-scale survey data is the possible presence of overdispersion. It occurs when the data display more variability than is predicted by the variance–mean relationship. This article describes a probability distribution generated by a mixture of discrete random variables to capture uncertainty, feeling, and overdispersion. Specifically, several tests for detecting overdispersion will be implemented on the basis of the asymptotic theory for maximum likelihood estimators. We discuss the results of a simulation experiment concerning log-likelihood ratio, Wald, Score, and Profile tests. Finally, some real datasets are analyzed to illustrate the previous results.  相似文献   
107.
In this article, we investigate the relationships among intraday serial correlation, jump-robust volatility, positive and negative jumps based on Shanghai composite index high frequency data. We implement variance ratio test to quantify intraday serial correlation. We also measure the continuous part of realized volatility using jump-robust MedRV estimator and disentangle positive and negative jumps using Realized Downside Risk Measure and Realized Upside Potential Measure proposed by Bi et al., (2013 Bi, T., Zhang, B., Wu, H. (2013). Measuring downside risk using high frequency data–realized downside risk measure. Communications in Statistics–Simulation and Computation 42(4):741754.[Taylor &; Francis Online], [Web of Science ®] [Google Scholar]). We find that intraday serial correlation are positively correlated with jump-robust volatility and negatively correlated with negative jumps which confirm the LeBaron effect.  相似文献   
108.
The paper proposes a new test for detecting the umbrella pattern under a general non‐parametric scheme. The alternative asserts that the umbrella ordering holds while the hypothesis is its complement. The main focus is put on controlling the power function of the test outside the alternative. As a result, the asymptotic error of the first kind of the constructed solution is smaller than or equal to the fixed significance level α on the whole set where the umbrella ordering does not hold. Also, under finite sample sizes, this error is controlled to a satisfactory extent. A simulation study shows, among other things, that the new test improves upon the solution widely recommended in the literature of the subject. A routine, written in R, is attached as the Supporting Information file.  相似文献   
109.
This article proposes a linearly weighted unit root test with a new weighting scheme which reflects the trade-off in power between the ADF and LM tests regarding the initial value of a time series. Simulation results indicate that the proposed test has better power performance and works better than other available tests in the literature for a range of initial conditions.  相似文献   
110.
Data envelopment analysis (DEA) and free disposal hull (FDH) estimators are widely used to estimate efficiency of production. Practitioners use DEA estimators far more frequently than FDH estimators, implicitly assuming that production sets are convex. Moreover, use of the constant returns to scale (CRS) version of the DEA estimator requires an assumption of CRS. Although bootstrap methods have been developed for making inference about the efficiencies of individual units, until now no methods exist for making consistent inference about differences in mean efficiency across groups of producers or for testing hypotheses about model structure such as returns to scale or convexity of the production set. We use central limit theorem results from our previous work to develop additional theoretical results permitting consistent tests of model structure and provide Monte Carlo evidence on the performance of the tests in terms of size and power. In addition, the variable returns to scale version of the DEA estimator is proved to attain the faster convergence rate of the CRS-DEA estimator under CRS. Using a sample of U.S. commercial banks, we test and reject convexity of the production set, calling into question results from numerous banking studies that have imposed convexity assumptions. Supplementary materials for this article are available online.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号