首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 703 毫秒
1.
In this paper we evaluate the performance of three methods for testing the existence of a unit root in a time series, when the models under consideration in the null hypothesis do not display autocorrelation in the error term. In such cases, simple versions of the Dickey-Fuller test should be used as the most appropriate ones instead of the known augmented Dickey-Fuller or Phillips-Perron tests. Through Monte Carlo simulations we show that, apart from a few cases, testing the existence of a unit root we obtain actual type I error and power very close to their nominal levels. Additionally, when the random walk null hypothesis is true, by gradually increasing the sample size, we observe that p-values for the drift in the unrestricted model fluctuate at low levels with small variance and the Durbin-Watson (DW) statistic is approaching 2 in both the unrestricted and restricted models. If, however, the null hypothesis of a random walk is false, taking a larger sample, the DW statistic in the restricted model starts to deviate from 2 while in the unrestricted model it continues to approach 2. It is also shown that the probability not to reject that the errors are uncorrelated, when they are indeed not correlated, is higher when the DW test is applied at 1% nominal level of significance.  相似文献   

2.
A number of parametric and non-parametric linear trend tests for time series are evaluated in terms of test size and power, using also resampling techniques to form the empirical distribution of the test statistics under the null hypothesis of no linear trend. For resampling, both bootstrap and surrogate data are considered. Monte Carlo simulations were done for several types of residuals (uncorrelated and correlated with normal and nonnormal distributions) and a range of small magnitudes of the trend coefficient. In particular for AR(1) and ARMA(1, 1) residual processes, we investigate the discrimination of strong autocorrelation from linear trend with respect to the sample size. The correct test size is obtained for larger data sizes as autocorrelation increases and only when a randomization test that accounts for autocorrelation is used. The overall results show that the type I and II errors of the trend tests are reduced with the use of resampled data. Following the guidelines suggested by the simulation results, we could find significant linear trend in the data of land air temperature and sea surface temperature.  相似文献   

3.
In this article, we propose various tests for serial correlation in fixed-effects panel data regression models with a small number of time periods. First, a simplified version of the test suggested by Wooldridge (2002) and Drukker (2003) is considered. The second test is based on the Lagrange Multiplier (LM) statistic suggested by Baltagi and Li (1995), and the third test is a modification of the classical Durbin–Watson statistic. Under the null hypothesis of no serial correlation, all tests possess a standard normal limiting distribution as N tends to infinity and T is fixed. Analyzing the local power of the tests, we find that the LM statistic has superior power properties. Furthermore, a generalization to test for autocorrelation up to some given lag order and a test statistic that is robust against time dependent heteroskedasticity are proposed.  相似文献   

4.
A residual-based test of the null of cointegration in panel data   总被引:2,自引:0,他引:2  
This paper proposes a residual-based Lagrange Multiplier (LM) test for the null of cointegration in panel data. The test is analogous to the locally best unbiased invariant (LBUI) for a moving average (MA) unit root. The asymptotic distribution of the test is derived under the null. Monte Carlo simulations are performed to study the size and power properties of the proposed test.

overall, the empirical sizes of the LM-FM and LM-DOLs are close to the true size even in small samples. The power is quite good for the panels where T ≥ 50, and decent with panels for fewer observation in T. In our fixed sample of N = 50 and T = 50, the presence of a moving average and correlation between the LM-DOLS test seems to be better at correcting these effects, although in some cases the LM-FM test is more powerful.

Although much of the non-stationary time series econometrics has been criticized for having more to do with the specific properties of the data set rather than underlying economic models, the recent development of the cointegration literature has allowed for a concrete bridge between economic long run theory and time series methods. Our test now allows for the testing of the null of cointegration in a panel setting and should be of considerable interest to economists in a wide variety of fields.  相似文献   

5.
A disease prevalence can be estimated by classifying subjects according to whether they have the disease. When gold-standard tests are too expensive to be applied to all subjects, partially validated data can be obtained by double-sampling in which all individuals are classified by a fallible classifier, and some of individuals are validated by the gold-standard classifier. However, it could happen in practice that such infallible classifier does not available. In this article, we consider two models in which both classifiers are fallible and propose four asymptotic test procedures for comparing disease prevalence in two groups. Corresponding sample size formulae and validated ratio given the total sample sizes are also derived and evaluated. Simulation results show that (i) Score test performs well and the corresponding sample size formula is also accurate in terms of the empirical power and size in two models; (ii) the Wald test based on the variance estimator with parameters estimated under the null hypothesis outperforms the others even under small sample sizes in Model II, and the sample size estimated by this test is also accurate; (iii) the estimated validated ratios based on all tests are accurate. The malarial data are used to illustrate the proposed methodologies.  相似文献   

6.
Supremum score test statistics are often used to evaluate hypotheses with unidentifiable nuisance parameters under the null hypothesis. Although these statistics provide an attractive framework to address non‐identifiability under the null hypothesis, little attention has been paid to their distributional properties in small to moderate sample size settings. In situations where there are identifiable nuisance parameters under the null hypothesis, these statistics may behave erratically in realistic samples as a result of a non‐negligible bias induced by substituting these nuisance parameters by their estimates under the null hypothesis. In this paper, we propose an adjustment to the supremum score statistics by subtracting the expected bias from the score processes and show that this adjustment does not alter the limiting null distribution of the supremum score statistics. Using a simple example from the class of zero‐inflated regression models for count data, we show empirically and theoretically that the adjusted tests are superior in terms of size and power. The practical utility of this methodology is illustrated using count data in HIV research.  相似文献   

7.
When the error terms are autocorrelated, the conventional t-tests for individual regression coefficients mislead us to over-rejection of the null hypothesis. We examine, by Monte Carlo experiments, the small sample properties of the unrestricted estimator of ρ and of the estimator of ρ restricted by the null hypothesis. We compare the small sample properties of the Wald, likelihood ratio and Lagrange multiplier test statistics for individual regression coefficients. It is shown that when the null hypothesis is true, the unrestricted estimator of ρ is biased. It is also shown that the Lagrange multiplier test using the maximum likelihood estimator of ρ performs better than the Wald and likelihood ratio tests.  相似文献   

8.

We address the testing problem of proportional hazards in the two-sample survival setting allowing right censoring, i.e., we check whether the famous Cox model is underlying. Although there are many test proposals for this problem, only a few papers suggest how to improve the performance for small sample sizes. In this paper, we do exactly this by carrying out our test as a permutation as well as a wild bootstrap test. The asymptotic properties of our test, namely asymptotic exactness under the null and consistency, can be transferred to both resampling versions. Various simulations for small sample sizes reveal an actual improvement of the empirical size and a reasonable power performance when using the resampling versions. Moreover, the resampling tests perform better than the existing tests of Gill and Schumacher and Grambsch and Therneau . The tests’ practical applicability is illustrated by discussing real data examples.

  相似文献   

9.
We show that for any sample size, any size of the test, and any weights matrix outside a small class of exceptions, there exists a positive measure set of regression spaces such that the power of the Cliff–Ord test vanishes as the autocorrelation increases in a spatial error model. This result extends to the tests that define the Gaussian power envelope of all invariant tests for residual spatial autocorrelation. In most cases, the regression spaces such that the problem occurs depend on the size of the test, but there also exist regression spaces such that the power vanishes regardless of the size. A characterization of such particularly hostile regression spaces is provided.  相似文献   

10.
In this paper, we investigate the properties of the Granger causality test in stationary and stable vector autoregressive models under the presence of spillover effects, that is, causality in variance. The Wald test and the WW test (the Wald test with White's proposed heteroskedasticity-consistent covariance matrix estimator imposed) are analyzed. The investigation is undertaken by using Monte Carlo simulation in which two different sample sizes and six different kinds of data-generating processes are used. The results show that the Wald test over-rejects the null hypothesis both with and without the spillover effect, and that the over-rejection in the latter case is more severe in larger samples. The size properties of the WW test are satisfactory when there is spillover between the variables. Only when there is feedback in the variance is the size of the WW test slightly affected. The Wald test is shown to have higher power than the WW test when the errors follow a GARCH(1,1) process without a spillover effect. When there is a spillover, the power of both tests deteriorates, which implies that the spillover has a negative effect on the causality tests.  相似文献   

11.
We examine the asymptotic and small sample properties of model-based and robust tests of the null hypothesis of no randomized treatment effect based on the partial likelihood arising from an arbitrarily misspecified Cox proportional hazards model. When the distribution of the censoring variable is either conditionally independent of the treatment group given covariates or conditionally independent of covariates given the treatment group, the numerators of the partial likelihood treatment score and Wald tests have asymptotic mean equal to 0 under the null hypothesis, regardless of whether or how the Cox model is misspecified. We show that the model-based variance estimators used in the calculation of the model-based tests are not, in general, consistent under model misspecification, yet using analytic considerations and simulations we show that their true sizes can be as close to the nominal value as tests calculated with robust variance estimators. As a special case, we show that the model-based log-rank test is asymptotically valid. When the Cox model is misspecified and the distribution of censoring depends on both treatment group and covariates, the asymptotic distributions of the resulting partial likelihood treatment score statistic and maximum partial likelihood estimator do not, in general, have a zero mean under the null hypothesis. Here neither the fully model-based tests, including the log-rank test, nor the robust tests will be asymptotically valid, and we show through simulations that the distortion to test size can be substantial.  相似文献   

12.
Nonparametric regression models are often used to check or suggest a parametric model. Several methods have been proposed to test the hypothesis of a parametric regression function against an alternative smoothing spline model. Some tests such as the locally most powerful (LMP) test by Cox et al. (Cox, D., Koh, E., Wahba, G. and Yandell, B. (1988). Testing the (parametric) null model hypothesis in (semiparametric) partial and generalized spline models. Ann. Stat., 16, 113–119.), the generalized maximum likelihood (GML) ratio test and the generalized cross validation (GCV) test by Wahba (Wahba, G. (1990). Spline models for observational data. CBMS-NSF Regional Conference Series in Applied Mathematics, SIAM.) were developed from the corresponding Bayesian models. Their frequentist properties have not been studied. We conduct simulations to evaluate and compare finite sample performances. Simulation results show that the performances of these tests depend on the shape of the true function. The LMP and GML tests are more powerful for low frequency functions while the GCV test is more powerful for high frequency functions. For all test statistics, distributions under the null hypothesis are complicated. Computationally intensive Monte Carlo methods can be used to calculate null distributions. We also propose approximations to these null distributions and evaluate their performances by simulations.  相似文献   

13.
In this article, we propose a testing technique for multivariate heteroscedasticity, which is expressed as a test of linear restrictions in a multivariate regression model. Four test statistics with known asymptotical null distributions are suggested, namely the Wald, Lagrange multiplier (LM), likelihood ratio (LR) and the multivariate Rao F-test. The critical values for the statistics are determined by their asymptotic null distributions, but bootstrapped critical values are also used. The size, power and robustness of the tests are examined in a Monte Carlo experiment. Our main finding is that all the tests limit their nominal sizes asymptotically, but some of them have superior small sample properties. These are the F, LM and bootstrapped versions of Wald and LR tests.  相似文献   

14.
The size and power of various generalization tests for the Granger-causality in integrated-cointegrated VAR systems are considered. By using Monte Carlo methods, properties of eight versions of the test are studied in two different forms, the standard form and the modified form by Dolado & Lütkepohl (1996) in a study confined to properties of the Wald test only. In their study as well as in ours, both the standard and the modified Wald tests are shown to perform badly especially in small samples. We find, however, that the corrected LR tests exhibit correct size even in small samples. The power of the test is higher when the true VAR(2) model is estimated, and the modified test loses information by estimating the extra coefficients. The same is true when considering the power results in the VAR(3) model, and the power of the tests is somewhat lower than those in the VAR(2).  相似文献   

15.
The Durbin–Watson (DW) test for lag 1 autocorrelation has been generalized (DWG) to test for autocorrelations at higher lags. This includes the Wallis test for lag 4 autocorrelation. These tests are also applicable to test for the important hypothesis of randomness. It is found that for small sample sizes a normal distribution or a scaled beta distribution by matching the first two moments approximates well the null distribution of the DW and DWG statistics. The approximations seem to be adequate even when the samples are from nonnormal distributions. These approximations require the first two moments of these statistics. The expressions of these moments are derived.  相似文献   

16.

Sign test using median ranked set samples (MRSS) is introduced and investigated. We show that, this test is more powerful than the sign tests based on simple random sample (SRS) and ranked set sample (RSS) for finite sample size. It is found that, when the set size of MRSS is odd, the null distribution of the MRSS sign test is the same as the sign test obtained by using SRS. The exact null distributions and the power functions, in case of finite sample sizes, of these tests are derived. Also, the asymptotic distribution of the MRSS sign tests are derived. Numerical comparison of the MRSS sign test power with the power of the SRS sign test and the RSS sign test is given. Illustration of the procedure, using real data set of bilirubin level in Jaundice babies who stay in neonatal intensive care is introduced.  相似文献   

17.
In this paper, we show that the widely used stationarity tests such as the Kwiatkowski Phillips, Schmidt, and Shin (KPSS) test have power close to size in the presence of time-varying unconditional variance. We propose a new test as a complement of the existing tests. Monte Carlo experiments show that the proposed test possesses the following characteristics: (i) in the presence of unit root or a structural change in the mean, the proposed test is as powerful as the KPSS and other tests; (ii) in the presence of a changing variance, the traditional tests perform badly whereas the proposed test has high power comparing to the existing tests; (iii) the proposed test has the same size as traditional stationarity tests under the null hypothesis of stationarity. An application to daily observations of return on U.S. Dollar/Euro exchange rate reveals the existence of instability in the unconditional variance when the entire sample is considered, but stability is found in subsamples.  相似文献   

18.
In this paper, we show that the widely used stationarity tests such as the Kwiatkowski Phillips, Schmidt, and Shin (KPSS) test have power close to size in the presence of time-varying unconditional variance. We propose a new test as a complement of the existing tests. Monte Carlo experiments show that the proposed test possesses the following characteristics: (i) in the presence of unit root or a structural change in the mean, the proposed test is as powerful as the KPSS and other tests; (ii) in the presence of a changing variance, the traditional tests perform badly whereas the proposed test has high power comparing to the existing tests; (iii) the proposed test has the same size as traditional stationarity tests under the null hypothesis of stationarity. An application to daily observations of return on U.S. Dollar/Euro exchange rate reveals the existence of instability in the unconditional variance when the entire sample is considered, but stability is found in subsamples.  相似文献   

19.
Developing new medical tests and identifying single biomarkers or panels of biomarkers with superior accuracy over existing classifiers promotes lifelong health of individuals and populations. Before a medical test can be routinely used in clinical practice, its accuracy within diseased and non-diseased populations must be rigorously evaluated. We introduce a method for sample size determination for studies designed to test hypotheses about medical test or biomarker sensitivity and specificity. We show how a sample size can be determined to guard against making type I and/or type II errors by calculating Bayes factors from multiple data sets simulated under null and/or alternative models. The approach can be implemented across a variety of study designs, including investigations into one test or two conditionally independent or dependent tests. We focus on a general setting that involves non-identifiable models for data when true disease status is unavailable due to the nonexistence of or undesirable side effects from a perfectly accurate (i.e. ‘gold standard’) test; special cases of the general method apply to identifiable models with or without gold-standard data. Calculation of Bayes factors is performed by incorporating prior information for model parameters (e.g. sensitivity, specificity, and disease prevalence) and augmenting the observed test-outcome data with unobserved latent data on disease status to facilitate Gibbs sampling from posterior distributions. We illustrate our methods using a thorough simulation study and an application to toxoplasmosis.  相似文献   

20.
We study the finite-sample performance of test statistics in linear regression models where the error dependence is of unknown form. With an unknown dependence structure, there is traditionally a trade-off between the maximum lag over which the correlation is estimated (the bandwidth) and the amount of heterogeneity in the process. When allowing for heterogeneity, through conditional heteroskedasticity, the correlation at far lags is generally omitted and the resultant inflation of the empirical size of test statistics has long been recognized. To allow for correlation at far lags, we study the test statistics constructed under the possibly misspecified assumption of conditional homoskedasticity. To improve the accuracy of the test statistics, we employ the second-order asymptotic refinement in Rothenberg [Approximate power functions for some robust tests of regression coefficients, Econometrica 56 (1988), pp. 997–1019] to determine the critical values. The simulation results of this paper suggest that when sample sizes are small, modelling the heterogeneity of a process is secondary to accounting for dependence. We find that a conditionally homoskedastic covariance matrix estimator (when used in conjunction with Rothenberg's second-order critical value adjustment) improves test size with only a minimal loss in test power, even when the data manifest significant amounts of heteroskedasticity. In some specifications, the size inflation was cut by nearly 40% over the traditional heteroskedasticity and autocorrelation consistent (HAC) test. Finally, we note that the proposed test statistics do not require that the researcher specify the bandwidth or the kernel.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号