共查询到20条相似文献,搜索用时 15 毫秒
1.
Media reports of the dangers of swine flu can leave much to be desired. But then scientific reports can be just as unclear. Diamanto Mamuneas and Maria Viskaduraki have sympathy for two journalists and help a frightened friend. 相似文献
2.
We propose a simple method for evaluating the model that has been chosen by an adaptive regression procedure, our main focus being the lasso. This procedure deletes each chosen predictor and refits the lasso to get a set of models that are “close” to the chosen “base model,” and compares the error rates of the base model with that of nearby models. If the deletion of a predictor leads to significant deterioration in the model's predictive power, the predictor is called indispensable; otherwise, the nearby model is called acceptable and can serve as a good alternative to the base model. This provides both an assessment of the predictive contribution of each variable and a set of alternative models that may be used in place of the chosen model. We call this procedure “Next-Door analysis” since it examines models “next” to the base model. It can be applied to supervised learning problems with ℓ1 penalization and stepwise procedures. We have implemented it in the R language as a library to accompany the well-known glmnet library. The Canadian Journal of Statistics 48: 447–470; 2020 © 2020 Statistical Society of Canada 相似文献
3.
This paper tackles the problem of detecting abrupt changes in the mean of a heteroscedastic signal by model selection, without
knowledge on the variations of the noise. A new family of change-point detection procedures is proposed, showing that cross-validation
methods can be successful in the heteroscedastic framework, whereas most existing procedures are not robust to heteroscedasticity.
The robustness to heteroscedasticity of the proposed procedures is supported by an extensive simulation study, together with
recent partial theoretical results. An application to Comparative Genomic Hybridization (CGH) data is provided, showing that
robustness to heteroscedasticity can indeed be required for their analysis. 相似文献
4.
Sy-mien Chen 《统计学通讯:理论与方法》2013,42(6):1407-1408
First of all, the parameters involved in Chen and Hsu (1995) are unknown. For convenience, we discussed the two possibilities of the unknown parameter μ namely μ =M and μ M, separately. In other words, we do not known wheter μ = M or μ M. Therefore, we must estimate it before drawing any statistical conclusion about the index Cpmk, The conclusion should be applied only after th ehypothesis if μ = M is tested. 相似文献
5.
6.
Raymond Hubbard 《The American statistician》2019,73(1):31-35
ABSTRACTRecent efforts by the American Statistical Association to improve statistical practice, especially in countering the misuse and abuse of null hypothesis significance testing (NHST) and p-values, are to be welcomed. But will they be successful? The present study offers compelling evidence that this will be an extraordinarily difficult task. Dramatic citation-count data on 25 articles and books severely critical of NHST's negative impact on good science, underlining that this issue was/is well known, did nothing to stem its usage over the period 1960–2007. On the contrary, employment of NHST increased during this time. To be successful in this endeavor, as well as restoring the relevance of the statistics profession to the scientific community in the 21st century, the ASA must be prepared to dispense detailed advice. This includes specifying those situations, if they can be identified, in which the p-value plays a clearly valuable role in data analysis and interpretation. The ASA might also consider a statement that recommends abandoning the use of p-values. 相似文献
7.
Giovanna Jona Lasinio Gianluca Mastrantonio Alessio Pollice 《Statistical Methods and Applications》2013,22(1):97-112
When a large amount of spatial data is available computational and modeling challenges arise and they are often labeled as “big n problem”. In this work we present a brief review of the literature. Then we focus on two approaches, respectively based on stochastic partial differential equations and integrated nested Laplace approximation, and on the tapering of the spatial covariance matrix. The fitting and predictive abilities of using the two methods in conjunction with Kriging interpolation are compared in a simulation study. 相似文献
8.
ABSTRACTWe propose a new estimator for the spot covariance matrix of a multi-dimensional continuous semimartingale log asset price process, which is subject to noise and nonsynchronous observations. The estimator is constructed based on a local average of block-wise parametric spectral covariance estimates. The latter originate from a local method of moments (LMM), which recently has been introduced by Bibinger et al.. We prove consistency and a point-wise stable central limit theorem for the proposed spot covariance estimator in a very general setup with stochastic volatility, leverage effects, and general noise distributions. Moreover, we extend the LMM estimator to be robust against autocorrelated noise and propose a method to adaptively infer the autocorrelations from the data. Based on simulations we provide empirical guidance on the effective implementation of the estimator and apply it to high-frequency data of a cross-section of Nasdaq blue chip stocks. Employing the estimator to estimate spot covariances, correlations, and volatilities in normal but also unusual periods yields novel insights into intraday covariance and correlation dynamics. We show that intraday (co-)variations (i) follow underlying periodicity patterns, (ii) reveal substantial intraday variability associated with (co-)variation risk, and (iii) can increase strongly and nearly instantaneously if new information arrives. Supplementary materials for this article are available online. 相似文献
9.
The case–control studies using longitudinal data are cost and time efficient when the disease is rare and assessing the exposure level of risk factors is difficult. Instead of GEE method, the method of using a prospective logistic model for analyzing case–control longitudinal data was proposed and the semiparametric inference procedure was explored by Park and Kim (2004). In this article, we apply an empirical likelihood ratio method to derive limiting distribution of the empirical likelihood ratio and find one likelihood-ratio based confidence region for the unknown vector of regression parameters. We compare empirical likelihood method with normal approximation based method. Simulation results show that the proposed empirical likelihood ratio method performs well in terms of coverage probability and interval width. 相似文献
10.
11.
Cluster analysis is the automated search for groups of homogeneous observations in a data set. A popular modeling approach
for clustering is based on finite normal mixture models, which assume that each cluster is modeled as a multivariate normal
distribution. However, the normality assumption that each component is symmetric is often unrealistic. Furthermore, normal
mixture models are not robust against outliers; they often require extra components for modeling outliers and/or give a poor
representation of the data. To address these issues, we propose a new class of distributions, multivariate t distributions with the Box-Cox transformation, for mixture modeling. This class of distributions generalizes the normal distribution
with the more heavy-tailed t distribution, and introduces skewness via the Box-Cox transformation. As a result, this provides a unified framework to simultaneously
handle outlier identification and data transformation, two interrelated issues. We describe an Expectation-Maximization algorithm
for parameter estimation along with transformation selection. We demonstrate the proposed methodology with three real data
sets and simulation studies. Compared with a wealth of approaches including the skew-t mixture model, the proposed t mixture model with the Box-Cox transformation performs favorably in terms of accuracy in the assignment of observations,
robustness against model misspecification, and selection of the number of components. 相似文献
12.
The joint probability density function, evaluated at the observed data, is commonly used as the likelihood function to compute maximum likelihood estimates. For some models, however, there exist paths in the parameter space along which this density-approximation likelihood goes to infinity and maximum likelihood estimation breaks down. In all applications, however, observed data are really discrete due to the round-off or grouping error of measurements. The “correct likelihood” based on interval censoring can eliminate the problem of an unbounded likelihood. This article categorizes the models leading to unbounded likelihoods into three groups and illustrates the density-approximation breakdown with specific examples. Although it is usually possible to infer how given data were rounded, when this is not possible, one must choose the width for interval censoring, so we study the effect of the round-off on estimation. We also give sufficient conditions for the joint density to provide the same maximum likelihood estimate as the correct likelihood, as the round-off error goes to zero. 相似文献
13.
14.
Conventional methods apply symmetric prior distributions such as a normal distribution or a Laplace distribution for regression coefficients, which may be suitable for median regression and exhibit no robustness to outliers. This work develops a quantile regression on linear panel data model without heterogeneity from a Bayesian point of view, i.e. upon a location-scale mixture representation of the asymmetric Laplace error distribution, and provides how the posterior distribution is summarized using Markov chain Monte Carlo methods. Applying this approach to the 1970 British Cohort Study (BCS) data, it finds that a different maternal health problem has different influence on child's worrying status at different quantiles. In addition, applying stochastic search variable selection for maternal health problems to the 1970 BCS data, it finds that maternal nervous breakdown, among the 25 maternal health problems, contributes most to influence the child's worrying status. 相似文献
15.
Retrospective Reporting of Household Wealth: Evidence From the 1983–1989 Survey of Consumer Finances
One way to obtain panel-like information on household wealth is to ask households about changes in their asset holdings. Yet the reliability of retrospective data is unclear, considering the potential for recall error. This article examines the reliability of retrospective reporting, using data from the 1983–1989 Survey of Consumer Finances. We find substantial inconsistencies between reported net investments in assets with measured changes in holdings. Inconsistencies are less severe for salient transactions like home sales and more severe for aggregated items like financial assets. 相似文献
16.
17.
A. J. Jaffe 《The American statistician》2013,67(1):27-31
A second course in statistics, for nonstatisticians who will use packaged statistical software in their work, is outlined. The course is directed toward the wise choice, use, and evaluation of statistical computer packages. The goal of the course is to train educated consumers of statistical programs. Particular attention is paid to computer-based data analysis, interpretation of output, comparison of competing packages, and statistical problems that arise when computers are employed to analyze large data sets. 相似文献
18.
19.
This note is on two theorems in a paper by Rainer Dyckerhoff (Allg. Stat. Arch. 88:163–190, 2004). We state a missing condition in Theorem 3. On the other hand, Theorem 2 can be weakened. 相似文献