全文获取类型
收费全文 | 2624篇 |
免费 | 53篇 |
国内免费 | 8篇 |
专业分类
管理学 | 280篇 |
人口学 | 11篇 |
丛书文集 | 11篇 |
理论方法论 | 33篇 |
综合类 | 67篇 |
社会学 | 24篇 |
统计学 | 2259篇 |
出版年
2024年 | 18篇 |
2023年 | 45篇 |
2022年 | 31篇 |
2021年 | 46篇 |
2020年 | 58篇 |
2019年 | 108篇 |
2018年 | 127篇 |
2017年 | 216篇 |
2016年 | 100篇 |
2015年 | 97篇 |
2014年 | 133篇 |
2013年 | 542篇 |
2012年 | 229篇 |
2011年 | 88篇 |
2010年 | 90篇 |
2009年 | 104篇 |
2008年 | 74篇 |
2007年 | 78篇 |
2006年 | 73篇 |
2005年 | 58篇 |
2004年 | 58篇 |
2003年 | 38篇 |
2002年 | 33篇 |
2001年 | 29篇 |
2000年 | 35篇 |
1999年 | 26篇 |
1998年 | 28篇 |
1997年 | 21篇 |
1996年 | 9篇 |
1995年 | 8篇 |
1994年 | 14篇 |
1993年 | 7篇 |
1992年 | 11篇 |
1991年 | 13篇 |
1990年 | 3篇 |
1989年 | 5篇 |
1988年 | 7篇 |
1987年 | 3篇 |
1986年 | 3篇 |
1985年 | 4篇 |
1984年 | 2篇 |
1983年 | 3篇 |
1982年 | 5篇 |
1981年 | 1篇 |
1980年 | 2篇 |
1979年 | 1篇 |
1975年 | 1篇 |
排序方式: 共有2685条查询结果,搜索用时 0 毫秒
991.
《Journal of Statistical Computation and Simulation》2012,82(6):1247-1263
Based on the Bayesian framework of utilizing a Gaussian prior for the univariate nonparametric link function and an asymmetric Laplace distribution (ALD) for the residuals, we develop a Bayesian treatment for the Tobit quantile single-index regression model (TQSIM). With the location-scale mixture representation of the ALD, the posterior inferences of the latent variables and other parameters are achieved via the Markov Chain Monte Carlo computation method. TQSIM broadens the scope of applicability of the Tobit models by accommodating nonlinearity in the data. The proposed method is illustrated by two simulation examples and a labour supply dataset. 相似文献
992.
《Journal of Statistical Computation and Simulation》2012,82(15):3108-3128
For a normal model with a conjugate prior, we provide an in-depth examination of the effects of the hyperparameters on the long-run frequentist properties of posterior point and interval estimates. Under an assumed sampling model for the data-generating mechanism, we examine how hyperparameter values affect the mean-squared error (MSE) of posterior means and the true coverage of credible intervals. We develop two types of hyperparameter optimality. MSE optimal hyperparameters minimize the MSE of posterior point estimates. Credible interval optimal hyperparameters result in credible intervals that have a minimum length while still retaining nominal coverage. A poor choice of hyperparameters has a worse consequence on the credible interval coverage than on the MSE of posterior point estimates. We give an example to demonstrate how our results can be used to evaluate the potential consequences of hyperparameter choices. 相似文献
993.
994.
《Journal of Statistical Computation and Simulation》2012,82(10):2007-2021
Process capability index Cpk has been the most popular one used in the manufacturing industry dealing with problems of measuring reproduction capability of processes to enhance product development with very low fraction of defectives. In the manufacturing industry, lower confidence bound (LCB) estimates the minimum process capability providing pivotal information for quality engineers to monitoring the process and assessing process performance for quality assurance. The main objective of this paper is to compare and contrast the LCBs on Cpk using two approaches, Classical method and Bayesian method. 相似文献
995.
《Journal of Statistical Computation and Simulation》2012,82(6):1183-1202
Quantile regression (QR) models have received increasing attention recently for longitudinal data analysis. When continuous responses appear non-centrality due to outliers and/or heavy-tails, commonly used mean regression models may fail to produce efficient estimators, whereas QR models may perform satisfactorily. In addition, longitudinal outcomes are often measured with non-normality, substantial errors and non-ignorable missing values. When carrying out statistical inference in such data setting, it is important to account for the simultaneous treatment of these data features; otherwise, erroneous or even misleading results may be produced. In the literature, there has been considerable interest in accommodating either one or some of these data features. However, there is relatively little work concerning all of them simultaneously. There is a need to fill up this gap as longitudinal data do often have these characteristics. Inferential procedure can be complicated dramatically when these data features arise in longitudinal response and covariate outcomes. In this article, our objective is to develop QR-based Bayesian semiparametric mixed-effects models to address the simultaneous impact of these multiple data features. The proposed models and method are applied to analyse a longitudinal data set arising from an AIDS clinical study. Simulation studies are conducted to assess the performance of the proposed method under various scenarios. 相似文献
996.
Test statistics for checking the independence between the innovations of several time series are developed. The time series models considered allow for general specifications for the conditional mean and variance functions that could depend on common explanatory variables. In testing for independence between more than two time series, checking pairwise independence does not lead to consistent procedures. Thus a finite family of empirical processes relying on multivariate lagged residuals are constructed, and we derive their asymptotic distributions. In order to obtain simple asymptotic covariance structures, Möbius transformations of the empirical processes are studied, and simplifications occur. Under the null hypothesis of independence, we show that these transformed processes are asymptotically Gaussian, independent, and with tractable covariance functions not depending on the estimated parameters. Various procedures are discussed, including Cramér–von Mises test statistics and tests based on non‐parametric measures. The ranks of the residuals are considered in the new methods, giving test statistics which are asymptotically margin‐free. Generalized cross‐correlations are introduced, extending the concept of cross‐correlation to an arbitrary number of time series; portmanteau procedures based on them are discussed. In order to detect the dependence visually, graphical devices are proposed. Simulations are conducted to explore the finite sample properties of the methodology, which is found to be powerful against various types of alternatives when the independence is tested between two and three time series. An application is considered, using the daily log‐returns of Apple, Intel and Hewlett‐Packard traded on the Nasdaq financial market. The Canadian Journal of Statistics 40: 447–479; 2012 © 2012 Statistical Society of Canada 相似文献
997.
Peter Burgherr 《Risk analysis》2013,33(1):146-160
We analyze the risk of severe fatal accidents causing five or more fatalities and for nine different activities covering the entire oil chain. Included are exploration and extraction, transport by different modes, refining and final end use in power plants, heating or gas stations. The risks are quantified separately for OECD and non‐OECD countries and trends are calculated. Risk is analyzed by employing a Bayesian hierarchical model yielding analytical functions for both frequency (Poisson) and severity distributions (Generalized Pareto) as well as frequency trends. This approach addresses a key problem in risk estimation—namely the scarcity of data resulting in high uncertainties in particular for the risk of extreme events, where the risk is extrapolated beyond the historically most severe accidents. Bayesian data analysis allows the pooling of information from different data sets covering, for example, the different stages of the energy chains or different modes of transportation. In addition, it also inherently delivers a measure of uncertainty. This approach provides a framework, which comprehensively covers risk throughout the oil chain, allowing the allocation of risk in sustainability assessments. It also permits the progressive addition of new data to refine the risk estimates. Frequency, severity, and trends show substantial differences between the activities, emphasizing the need for detailed risk analysis. 相似文献
998.
A new methodology is introduced based on Bayesian network both to model domino effect propagation patterns and to estimate the domino effect probability at different levels. The flexible structure and the unique modeling techniques offered by Bayesian network make it possible to analyze domino effects through a probabilistic framework, considering synergistic effects, noisy probabilities, and common cause failures. Further, the uncertainties and the complex interactions among the domino effect components are captured using Bayesian network. The probabilities of events are updated in the light of new information, and the most probable path of the domino effect is determined on the basis of the new data gathered. This study shows how probability updating helps to update the domino effect model either qualitatively or quantitatively. The methodology is applied to a hypothetical example and also to an earlier‐studied case study. These examples accentuate the effectiveness of Bayesian network in modeling domino effects in processing facility. 相似文献
999.
In recent years analyses of dependence structures using copulas have become more popular than the standard correlation analysis. Starting from Aas et al. ( 2009 ) regular vine pair‐copula constructions (PCCs) are considered the most flexible class of multivariate copulas. PCCs are involved objects but (conditional) independence present in data can simplify and reduce them significantly. In this paper the authors detect (conditional) independence in a particular vine PCC model based on bivariate t copulas by deriving and implementing a reversible jump Markov chain Monte Carlo algorithm. However, the methodology is general and can be extended to any regular vine PCC and to all known bivariate copula families. The proposed approach considers model selection and estimation problems for PCCs simultaneously. The effectiveness of the developed algorithm is shown in simulations and its usefulness is illustrated in two real data applications. The Canadian Journal of Statistics 39: 239–258; 2011 © 2011 Statistical Society of Canada 相似文献
1000.
本文应用贝叶斯方法研究了股价时序的均值和方差双重变点问题。基于后验概率比,我们提出一个类似ICSS算法的快速侦测算法。通过对上证指数时序的实证分析,我们总共发现5处方差突变。其中,3处是均值和方差双重变点,它们都对应中国股市的重大结构变化。 相似文献