首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
Standard Schwarz information criterion for testing a change-point in regression models is considered and two new test procedures are evolved. The case of small sample size is investigated. Numerical approximations to the power against various alternatives are given and compared with powers of tests based on r-ahead recursive residuals and of the CUSUM of squares test. Application of these procedures to some real data is also provided.  相似文献   

2.
This paper derives a simple ANOVA-F-statistic which tests for random individual effects in a one-way error component model, using recursive residuals. Power comparisons are performed for this F-test when it is computed using true disturbances and recursive residuals from a panel data regression. Under the null, both statistics have an exact F distribution. The standardized version of the Breusch and Pagan (1980) Lagrange Multiplier test (SLM) as well as a fixed effects F-statistic (FE) recommended by Moulton and Randolph (1989), are also included in this comparison. The exact power function can be computed in all cases using Imhof's (1961) procedure. Our results suggest that the F-test based on recursive residuals is inferior to the popular SLM and FE tests based on computational simplicity, power comparisons and its sensitivity to the K observations starting the recursion.  相似文献   

3.
This paper concerns model selection for autoregressive time series when the observations are contaminated with trend. We propose an adaptive least absolute shrinkage and selection operator (LASSO) type model selection method, in which the trend is estimated by B-splines, the detrended residuals are calculated, and then the residuals are used as if they were observations to optimize an adaptive LASSO type objective function. The oracle properties of such an adaptive LASSO model selection procedure are established; that is, the proposed method can identify the true model with probability approaching one as the sample size increases, and the asymptotic properties of estimators are not affected by the replacement of observations with detrended residuals. The intensive simulation studies of several constrained and unconstrained autoregressive models also confirm the theoretical results. The method is illustrated by two time series data sets, the annual U.S. tobacco production and annual tree ring width measurements.  相似文献   

4.
The error contrasts from an experimental design can be constructed from uncorrelated residuals normally associated with the linear model. In this paper uncorrelated residuals are defined for the linear model that has a design matrix which is less than full rank, typical of many experimental design representations. It transpires in this setting, that for certain choices of uncorrelated residuals, corresponding to recursive type residuals, there is a natural partition of information when two variance components are known to be present. Under an assumtion of normality of errors this leads to construction of appropriate F-tests for testing heteroscedasticity. The test, which can be optimal, is applied to two well known data sets to illustrate its usefullness.  相似文献   

5.
We propose a structural change test based on the recursive residuals with the local Fourier series estimators. The statistical properties of the proposed test are derived and the empirical properties are shown via simulation. We also consider other structural change tests based on CUSUM, MOSUM, moving estimates (ME), and empirical distribution functions with the recursive residuals and the ordinary residuals. Empirical powers are calculated in various structural change models for the comparison of those tests. These structural change tests are applied to South Korea's gross domestic product (GDP), South Korean Won to US Dollar currency exchange rates, and South Korea's Okun's law.  相似文献   

6.
Current statistical methods for analyzing epidemiological data with disease subtype information allow us to acquire knowledge not only for risk factor-disease subtype association but also, on a more profound account, heterogeneity in these associations by multiple disease characteristics (so-called etiologic heterogeneity of the disease). Current interest, particularly in cancer epidemiology, lies in obtaining a valid p-value for testing the hypothesis whether a particular cancer is etiologically heterogeneous. We consider the two-stage logistic regression model along with pseudo-conditional likelihood estimation method and design a testing strategy based on Rao's score test. An extensive Monte Carlo simulation study is carried out, false discovery rate and statistical power of the suggested test are investigated. Simulation results indicate that applying the proposed testing strategy, even a small degree of true etiologic heterogeneity can be recovered with a large statistical power from the sampled data. The strategy is then applied on a breast cancer data set to illustrate its use in practice where there are multiple risk factors and multiple disease characteristics of simultaneous concern.  相似文献   

7.
Specification tests for the error distribution are proposed in semi-linear models, including the partial linear model and additive models. The tests utilize an integrated distance involving the empirical characteristic function of properly estimated residuals. These residuals are obtained from an initial estimation step involving a combination of penalized least squares and smoothing techniques. A bootstrap version of the tests is utilized in order to study the small sample behavior of the procedures in comparison with more classical approaches. As an example, the tests are applied on some real data sets.  相似文献   

8.
This article presents a multiple hypothesis test procedure that combines two well known tests for structural change in the linear regression model, the CUSUM test and the recursive t test. The CUSUM test is run through the sequence of recursive residuals as usual; if the CUSUM plot does not violate the critical lines, one more step is taken to perform the t test for hypothesis of zero mean based on all recursive residuals. The asymptotic size of this multiple hypothesis test is derived; power simulation results suggest that it outperforms the traditional CUSUM test and complements other tests that are currently stressed in econometrics.  相似文献   

9.
The techniques for recursive estimation of the general linear model with dependent errors and known second order properties, is generalised to allow for simultaneous addition of an arbitrary number of additional observations. Computational formulae for recursive updating of parameter estimates are derived, together with a sequence of univariate recursive residuals for testing the constancy of the regression relation over time.  相似文献   

10.
This article presents a multiple hypothesis test procedure that combines two well known tests for structural change in the linear regression model, the CUSUM test and the recursive t test. The CUSUM test is run through the sequence of recursive residuals as usual; if the CUSUM plot does not violate the critical lines, one more step is taken to perform the t test for hypothesis of zero mean based on all recursive residuals. The asymptotic size of this multiple hypothesis test is derived; power simulation results suggest that it outperforms the traditional CUSUM test and complements other tests that are currently stressed in econometrics.  相似文献   

11.
This paper considers a linear regression model with regression parameter vector β. The parameter of interest is θ= aTβ where a is specified. When, as a first step, a data‐based variable selection (e.g. minimum Akaike information criterion) is used to select a model, it is common statistical practice to then carry out inference about θ, using the same data, based on the (false) assumption that the selected model had been provided a priori. The paper considers a confidence interval for θ with nominal coverage 1 ‐ α constructed on this (false) assumption, and calls this the naive 1 ‐ α confidence interval. The minimum coverage probability of this confidence interval can be calculated for simple variable selection procedures involving only a single variable. However, the kinds of variable selection procedures used in practice are typically much more complicated. For the real‐life data presented in this paper, there are 20 variables each of which is to be either included or not, leading to 220 different models. The coverage probability at any given value of the parameters provides an upper bound on the minimum coverage probability of the naive confidence interval. This paper derives a new Monte Carlo simulation estimator of the coverage probability, which uses conditioning for variance reduction. For these real‐life data, the gain in efficiency of this Monte Carlo simulation due to conditioning ranged from 2 to 6. The paper also presents a simple one‐dimensional search strategy for parameter values at which the coverage probability is relatively small. For these real‐life data, this search leads to parameter values for which the coverage probability of the naive 0.95 confidence interval is 0.79 for variable selection using the Akaike information criterion and 0.70 for variable selection using Bayes information criterion, showing that these confidence intervals are completely inadequate.  相似文献   

12.
Classification of high-dimensional data set is a big challenge for statistical learning and data mining algorithms. To effectively apply classification methods to high-dimensional data sets, feature selection is an indispensable pre-processing step of learning process. In this study, we consider the problem of constructing an effective feature selection and classification scheme for data set which has a small number of sample size with a large number of features. A novel feature selection approach, named four-Staged Feature Selection, has been proposed to overcome high-dimensional data classification problem by selecting informative features. The proposed method first selects candidate features with number of filtering methods which are based on different metrics, and then it applies semi-wrapper, union and voting stages, respectively, to obtain final feature subsets. Several statistical learning and data mining methods have been carried out to verify the efficiency of the selected features. In order to test the adequacy of the proposed method, 10 different microarray data sets are employed due to their high number of features and small sample size.  相似文献   

13.
Residuals are frequently used to evaluate the validity of the assumptions of statistical models and may also be employed as tools for model selection. In this paper, we consider residuals and their limiting properties in the linear mixed measurement error models. Also, we develop types of residuals for these models and then review some of the residual analysis techniques. Further, by using the definition of generalized leverage, we derive generalized leverage matrices for identification of high-leverage points for these models. Finally, we analyse a real data set.  相似文献   

14.
Often the unknown covariance structure of a stationary, dependent, Gaussian error sequence can be simply parametrised. The error sequence can either be directly observed or observed only through a random sequence containing a deterministic regression model. The method of scoring is used here, in conjunction with recursive estimation techniques, to effect the maximum likelihood estimation of the covariance parameters. Sequences of recursive residuals, useful in model diagnostics and data analysis, are obtained in the estimation procedure.  相似文献   

15.
The operating characteristics (OCs) of an indifference-zone ranking and selection procedure are derived for randomized response binomial data. The OCs include tables and figures to facilitate tradeoffs between sample size and a stated probability of a correct selection, i.e., correctly identifying the binomial population (out of k ≥ 2) characterized by the largest probability of success. Measures of efficiency are provided to assist the analyst in selection of an appropriate randomized response design for the collection of the data. A hybrid randomized response model, which includes the Warner model and the Greenberg et al. model, is introduced to facilitate comparisons among a wider range of statistical designs than previously available. An example comparing failure rates of contraceptive methods is used to illustrate the use of these new results.  相似文献   

16.
对复杂样本进行推断通常有两种体系,一种是传统的基于随机化理论的统计推断,另一种是基于模型的统计推断。传统的抽样理论以随机化理论为基础,将总体取值视为固定,随机性仅体现在样本的选取上,对总体的推断依赖于抽样设计。该方法在大样本情况下具有稳健估计量,但在小样本、数据缺失等情况下失效。基于模型的抽样推断认为总体是超总体模型中抽取的一个随机样本,对总体的推断取决于模型的建立,但在不可忽略抽样设计下估计量是有偏估计。在对这两类推断方法分析的基础上,提出抽样设计辅助的模型推断,并指出该方法在复杂抽样中具有重要的应用价值。  相似文献   

17.
Normal probability plots for a simple random sample and normal probability plots for residuals from linear regression are not treated differently in statistical text books. In the statistical literature, 1 ? α simultaneous probability intervals for augmenting a normal probability plot for a simple random sample are available. The first purpose of this article is to demonstrate that the tests associated with the 1 ? α simultaneous probability intervals for a simple random sample may have a size substantially different from α when applied to the residuals from linear regression. This leads to the second purpose of this article: construction of four normal probability plot-based tests for residuals, which have size α exactly. We then compare the powers of these four graphical tests and a non-graphical test for residuals in order to assess the power performances of the graphical tests and to identify the ones that have better power. Finally, an example is provided to illustrate the methods.  相似文献   

18.
Christensen & Lin ( 2015 ) suggested two lack of fit tests to assess the adequacy of a linear model based on partial sums of residuals. In particular, their tests evaluated the adequacy of the mean function. Their tests relied on asymptotic results without requiring small sample normality. We propose four new tests, find their asymptotic distributions, and propose an alternative simulation method for defining tests that is remarkably robust to the distribution of the errors. To assess their strengths and weaknesses, the Christensen & Lin ( 2015 ) tests and the new tests were compared in different scenarios by simulation. In particular, the new tests include two based on partial sums of absolute residuals. Previous partial sums of residuals tests have used signed residuals whose values when summed can cancel each other out. The use of absolute residuals requires small sample normality, but allows detection of lack of fit that was previously not possible with partial sums of residuals.  相似文献   

19.
Efficient statistical inference on nonignorable missing data is a challenging problem. This paper proposes a new estimation procedure based on composite quantile regression (CQR) for linear regression models with nonignorable missing data, that is applicable even with high-dimensional covariates. A parametric model is assumed for modelling response probability, which is estimated by the empirical likelihood approach. Local identifiability of the proposed strategy is guaranteed on the basis of an instrumental variable approach. A set of data-based adaptive weights constructed via an empirical likelihood method is used to weight CQR functions. The proposed method is resistant to heavy-tailed errors or outliers in the response. An adaptive penalisation method for variable selection is proposed to achieve sparsity with high-dimensional covariates. Limiting distributions of the proposed estimators are derived. Simulation studies are conducted to investigate the finite sample performance of the proposed methodologies. An application to the ACTG 175 data is analysed.  相似文献   

20.
A technique is given for drawing valid inferences in cases where performance characteristics of statistical procedures (e.g. power for a test, or probability of a correct selection for a selection procedure) depend upon unknown parameters (e.g. an unknown variance). The technique is especially useful in situations where sample sizes are small (e.g. in many medical trials); the “usual” approximate procedures are found to be misleading in such cases.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号