全文获取类型
收费全文 | 22854篇 |
免费 | 216篇 |
国内免费 | 4篇 |
专业分类
管理学 | 3108篇 |
民族学 | 160篇 |
人口学 | 3436篇 |
丛书文集 | 57篇 |
理论方法论 | 1500篇 |
综合类 | 502篇 |
社会学 | 10013篇 |
统计学 | 4298篇 |
出版年
2023年 | 96篇 |
2020年 | 205篇 |
2019年 | 322篇 |
2018年 | 1901篇 |
2017年 | 2078篇 |
2016年 | 1357篇 |
2015年 | 302篇 |
2014年 | 352篇 |
2013年 | 2511篇 |
2012年 | 707篇 |
2011年 | 1437篇 |
2010年 | 1289篇 |
2009年 | 1034篇 |
2008年 | 1109篇 |
2007年 | 1273篇 |
2006年 | 268篇 |
2005年 | 463篇 |
2004年 | 446篇 |
2003年 | 384篇 |
2002年 | 309篇 |
2001年 | 277篇 |
2000年 | 255篇 |
1999年 | 243篇 |
1998年 | 200篇 |
1997年 | 197篇 |
1996年 | 211篇 |
1995年 | 182篇 |
1994年 | 136篇 |
1993年 | 182篇 |
1992年 | 203篇 |
1991年 | 189篇 |
1990年 | 206篇 |
1989年 | 169篇 |
1988年 | 159篇 |
1987年 | 153篇 |
1986年 | 147篇 |
1985年 | 140篇 |
1984年 | 170篇 |
1983年 | 138篇 |
1982年 | 134篇 |
1981年 | 97篇 |
1980年 | 133篇 |
1979年 | 147篇 |
1978年 | 107篇 |
1977年 | 104篇 |
1976年 | 90篇 |
1975年 | 104篇 |
1974年 | 88篇 |
1973年 | 73篇 |
1972年 | 76篇 |
排序方式: 共有10000条查询结果,搜索用时 0 毫秒
11.
Jonathan H. Wright 《Econometric Reviews》2002,21(4):397-417
Many recent papers have used semiparametric methods, especially the log-periodogram regression, to detect and estimate long memory in the volatility of asset returns. In these papers, the volatility is proxied by measures such as squared, log-squared, and absolute returns. While the evidence for the existence of long memory is strong using any of these measures, the actual long memory parameter estimates can be sensitive to which measure is used. In Monte-Carlo simulations, I find that if the data is conditionally leptokurtic, the log-periodogram regression estimator using squared returns has a large downward bias, which is avoided by using other volatility measures. In United States stock return data, I find that squared returns give much lower estimates of the long memory parameter than the alternative volatility measures, which is consistent with the simulation results. I conclude that researchers should avoid using the squared returns in the semiparametric estimation of long memory volatility dependencies. 相似文献
12.
If a population contains many zero values and the sample size is not very large, the traditional normal approximation‐based confidence intervals for the population mean may have poor coverage probabilities. This problem is substantially reduced by constructing parametric likelihood ratio intervals when an appropriate mixture model can be found. In the context of survey sampling, however, there is a general preference for making minimal assumptions about the population under study. The authors have therefore investigated the coverage properties of nonparametric empirical likelihood confidence intervals for the population mean. They show that under a variety of hypothetical populations, these intervals often outperformed parametric likelihood intervals by having more balanced coverage rates and larger lower bounds. The authors illustrate their methodology using data from the Canadian Labour Force Survey for the year 2000. 相似文献
13.
Sets of relatively short time series arise in many situations. One aspect of their analysis may be the detection of outlying
series. We examine the performance of standard normal outlier tests applied to the means, or to simple functions of the means,
of AR(1) series, not necessarily of equal lengths. Although unequal lengths of series implies that the means have unequal
variances, that are only known approximately, it is shown that nominal significance levels hold good under most circumstances.
Thus a standard outlier test can usefully be applied, avoiding the complication of estimating the time series' parameters.
The test's power is affected by unequal lengths, being higher when the slippage occurs in one of the longer series 相似文献
14.
Craig H. Mallinckrodt Christopher J. Kaiser John G. Watkin Michael J. Detke Geert Molenberghs Raymond J. Carroll 《Pharmaceutical statistics》2004,3(3):171-186
The last observation carried forward (LOCF) approach is commonly utilized to handle missing values in the primary analysis of clinical trials. However, recent evidence suggests that likelihood‐based analyses developed under the missing at random (MAR) framework are sensible alternatives. The objective of this study was to assess the Type I error rates from a likelihood‐based MAR approach – mixed‐model repeated measures (MMRM) – compared with LOCF when estimating treatment contrasts for mean change from baseline to endpoint (Δ). Data emulating neuropsychiatric clinical trials were simulated in a 4 × 4 factorial arrangement of scenarios, using four patterns of mean changes over time and four strategies for deleting data to generate subject dropout via an MAR mechanism. In data with no dropout, estimates of Δ and SEΔ from MMRM and LOCF were identical. In data with dropout, the Type I error rates (averaged across all scenarios) for MMRM and LOCF were 5.49% and 16.76%, respectively. In 11 of the 16 scenarios, the Type I error rate from MMRM was at least 1.00% closer to the expected rate of 5.00% than the corresponding rate from LOCF. In no scenario did LOCF yield a Type I error rate that was at least 1.00% closer to the expected rate than the corresponding rate from MMRM. The average estimate of SEΔ from MMRM was greater in data with dropout than in complete data, whereas the average estimate of SEΔ from LOCF was smaller in data with dropout than in complete data, suggesting that standard errors from MMRM better reflected the uncertainty in the data. The results from this investigation support those from previous studies, which found that MMRM provided reasonable control of Type I error even in the presence of MNAR missingness. No universally best approach to analysis of longitudinal data exists. However, likelihood‐based MAR approaches have been shown to perform well in a variety of situations and are a sensible alternative to the LOCF approach. MNAR methods can be used within a sensitivity analysis framework to test the potential presence and impact of MNAR data, thereby assessing robustness of results from an MAR method. Copyright © 2004 John Wiley & Sons, Ltd. 相似文献
15.
Seismic risk can be reduced by implementing newly developed seismic provisions in design codes. Furthermore, financial protection or enhanced utility and happiness for stakeholders could be gained through the purchase of earthquake insurance. If this is not so, there would be no market for such insurance. However, perceived benefit associated with insurance is not universally shared by stakeholders partly due to their diverse risk attitudes. This study investigates the implied seismic design preference with insurance options for decisionmakers of bounded rationality whose preferences could be adequately represented by the cumulative prospect theory (CPT). The investigation is focused on assessing the sensitivity of the implied seismic design preference with insurance options to model parameters of the CPT and to fair and unfair insurance arrangements. Numerical results suggest that human cognitive limitation and risk perception can affect the implied seismic design preference by the CPT significantly. The mandatory purchase of fair insurance will lead the implied seismic design preference to the optimum design level that is dictated by the minimum expected lifecycle cost rule. Unfair insurance decreases the expected gain as well as its associated variability, which is preferred by risk-averse decisionmakers. The obtained results of the implied preference for the combination of the seismic design level and insurance option suggest that property owners, financial institutions, and municipalities can take advantage of affordable insurance to establish successful seismic risk management strategies. 相似文献
16.
Bram Thuysbaert 《Journal of Economic Inequality》2008,6(1):33-55
Empirical applications of poverty measurement often have to deal with a stochastic weighting variable such as household size.
Within the framework of a bivariate distribution function defined over income and weight, I derive the limiting distributions
of the decomposable poverty measures and of the ordinates of stochastic dominance curves. The poverty line is allowed to depend
on the income distribution. It is shown how the results can be used to test hypotheses concerning changes in poverty. The
inference procedures are briefly illustrated using Belgian data.
An erratum to this article can be found at 相似文献
17.
Singh et al. ([13]) pointed out that the Randomized response (RR) technique proposed by Moors ([9]) is not desirable because
it fails to protect the confidentiality of the respondents and they provided two alternative strategies free from the above
drawback but limited to SRSWOR sampling only. In this paper, generalization of one of the strategies is provided for complex
survey designs, wider class of estimators and for quantitative characteristics. Relative efficiency of the modified strategy
is tested through empirical investigations.
An erratum to this article is available at . 相似文献
18.
19.
20.
Time, Self, and the Curiously Abstract Concept of Agency* 总被引:2,自引:0,他引:2
The term "agency" is quite slippery and is used differently depending on the epistemological roots and goals of scholars who employ it. Distressingly, the sociological literature on the concept rarely addresses relevant social psychological research. We take a social behaviorist approach to agency by suggesting that individual temporal orientations are underutilized in conceptualizing this core sociological concept. Different temporal foci—the actor's engaged response to situational circumstances—implicate different forms of agency. This article offers a theoretical model involving four analytical types of agency ("existential,""identity,""pragmatic," and "life course") that are often conflated across treatments of the topic. Each mode of agency overlaps with established social psychological literatures, most notably about the self, enabling scholars to anchor overly abstract treatments of agency within established research literatures. 相似文献