首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Several methods for comparing k populations have been proposed in the literature. These methods assess the same null hypothesis of equal distributions but differ in the alternative hypothesis they consider. We focus on two important alternative hypotheses: monotone and umbrella ordering. Two new families of test statistics are proposed, including two known tests, as well as two new powerful tests under monotone ordering. Furthermore, these families are adapted for testing umbrella ordering. We compare some members of the families with respect to power and Type I errors under different simulation scenarios. Finally, the methods are illustrated in several applications to real data.  相似文献   

2.
We propose new tests of the martingale hypothesis based on generalized versions of the Kolmogorov–Smirnov and Cramér–von Mises tests. The tests are distribution-free and allow for a weak drift in the null model. The methods do not require either smoothing parameters or bootstrap resampling for their implementation and so are well suited to practical work. The article develops limit theory for the tests under the null and shows that the tests are consistent against a wide class of nonlinear, nonmartingale processes. Simulations show that the tests have good finite sample properties in comparison with other tests particularly under conditional heteroscedasticity and mildly explosive alternatives. An empirical application to major exchange rate data finds strong evidence in favor of the martingale hypothesis, confirming much earlier research.  相似文献   

3.
Nonparametric regression models are often used to check or suggest a parametric model. Several methods have been proposed to test the hypothesis of a parametric regression function against an alternative smoothing spline model. Some tests such as the locally most powerful (LMP) test by Cox et al. (Cox, D., Koh, E., Wahba, G. and Yandell, B. (1988). Testing the (parametric) null model hypothesis in (semiparametric) partial and generalized spline models. Ann. Stat., 16, 113–119.), the generalized maximum likelihood (GML) ratio test and the generalized cross validation (GCV) test by Wahba (Wahba, G. (1990). Spline models for observational data. CBMS-NSF Regional Conference Series in Applied Mathematics, SIAM.) were developed from the corresponding Bayesian models. Their frequentist properties have not been studied. We conduct simulations to evaluate and compare finite sample performances. Simulation results show that the performances of these tests depend on the shape of the true function. The LMP and GML tests are more powerful for low frequency functions while the GCV test is more powerful for high frequency functions. For all test statistics, distributions under the null hypothesis are complicated. Computationally intensive Monte Carlo methods can be used to calculate null distributions. We also propose approximations to these null distributions and evaluate their performances by simulations.  相似文献   

4.
In many engineering problems it is necessary to draw statistical inferences on the mean of a lognormal distribution based on a complete sample of observations. Statistical demonstration of mean time to repair (MTTR) is one example. Although optimum confidence intervals and hypothesis tests for the lognormal mean have been developed, they are difficult to use, requiring extensive tables and/or a computer. In this paper, simplified conservative methods for calculating confidence intervals or hypothesis tests for the lognormal mean are presented. In this paper, “conservative” refers to confidence intervals (hypothesis tests) whose infimum coverage probability (supremum probability of rejecting the null hypothesis taken over parameter values under the null hypothesis) equals the nominal level. The term “conservative” has obvious implications to confidence intervals (they are “wider” in some sense than their optimum or exact counterparts). Applying the term “conservative” to hypothesis tests should not be confusing if it is remembered that this implies that their equivalent confidence intervals are conservative. No implication of optimality is intended for these conservative procedures. It is emphasized that these are direct statistical inference methods for the lognormal mean, as opposed to the already well-known methods for the parameters of the underlying normal distribution. The method currently employed in MIL-STD-471A for statistical demonstration of MTTR is analyzed and compared to the new method in terms of asymptotic relative efficiency. The new methods are also compared to the optimum methods derived by Land (1971, 1973).  相似文献   

5.
In this paper, we consider a model checking problem for general linear models with randomly missing covariates. Two types of score type tests with inverse probability weight, which is estimated by parameter and nonparameter methods respectively, are proposed to this goodness of fit problem. The asymptotic properties of the test statistics are developed under the null and local alternative hypothesis. Simulation study is carried out to present the performance of the sizes and powers of the tests. We illustrate the proposed method with a data set on monozygotic twins.  相似文献   

6.
We consider the problem of accounting for multiplicity for two correlated endpoints in the comparison of two treatments using weighted hypothesis tests. Various weighted testing procedures are reviewed, and a more powerful method (a variant of the weighted Simes test) is evaluated for the general bivariate normal case and for a particular clinical trial example. Results from these evaluations are summarized and indicate that the weighted methods perform in a manner similar to unweighted methods. Copyright © 2005 John Wiley & Sons, Ltd.  相似文献   

7.
We develop two tests sensitive to various departures from composite goodness-of-fit hypothesis of normality. The tests are based on the sums of squares of some components naturally arising in decomposition of the Shapiro–Wilk-type statistic. Each component itself has diagnostic properties. The numbers of squared components in sums are determined via some novel selection rules based on the data. The new solutions prove to be effective tools in detecting a broad spectrum of sources of non-Gaussianity. We also discuss two variants of the new tests adjusted to verification of simple goodness-of-fit hypothesis of normality. These variants also compare well to popular competitors.  相似文献   

8.
A comparative study is made of three tests, developed by James (1951), Welch (1951) and Brown & Forsythe (1974). James presented two methods of which only one is considered in this paper. It is shown that this method gives better control over the size than the other two tests. None of these methods is uniformly more powerful than the other two. In some cases the tests of James and Welch reject a false null hypothesis more often than the test of Brown & Forsythe, but there are also situations in which it is the other way around.

We conclude that for implementation in a statistical software package the very complicated test of James is the most attractive. A practical disadvantage of this method can be overcome by a minor modification.  相似文献   

9.
Detecting dependence between marks and locations of marked point processes   总被引:1,自引:0,他引:1  
Summary.  We introduce two characteristics for stationary and isotropic marked point proces- ses, E ( h ) and V ( h ), and describe their use in investigating mark–point interactions. These quantities are functions of the interpoint distance h and denote the conditional expectation and the conditional variance of a mark respectively, given that there is a further point of the process a distance h away. We present tests based on E and V for the hypothesis that the values of the marks can be modelled by a random field which is independent of the unmarked point process. We apply the methods to two data sets in forestry.  相似文献   

10.
We present a test of the fit to a Poisson model based on the empirical probability generating function (epgf). We derive the limiting distribution of the test under the Poisson hypothesis and show that a rescaling of it is approximately independent of the mean parameter in the Poisson distribution. When inspected under a simulation study over a range of alternative distributions, we find that this test shows reasonable behaviour compared to other goodness-of-fit tests like the Poisson index of dispersion and smooth test applied to the Poisson model. These results illustrate that epgf-based methods for anlyzing count data are promising.  相似文献   

11.
In this article, we investigate the efficiency of score tests for testing a censored Poisson regression model against censored negative binomial regression alternatives. Based on the results of a simulation study, score tests using the normal approximation, underestimate the nominal significance level. To remedy this problem, bootstrap methods are proposed. We find that bootstrap methods keep the significance level close to the nominal one and have greater power uniformly than does the normal approximation for testing the hypothesis.  相似文献   

12.
ABSTRACT

The Mack–Wolfe test is the most frequently used non parametric procedure for the umbrella alternative problem. In this paper, modifications of the Mack–Wolfe test are proposed for both known peak and unknown peak umbrellas. The exact mean and variance of the proposed tests in the null hypothesis are also derived. We compare these tests with some of the existing tests in terms of the type I error rate and power. In addition, a real data example is presented.  相似文献   

13.
This article considers inference on correlation coefficients of bivariate log-normal distributions. We developed generalized confidence intervals and hypothesis tests for the correlation coefficients, and extended the results to compare two independent correlations. Simulation studies show that the suggested methods work well. Two practical examples are used to illustrate the application of the proposed methods.  相似文献   

14.
Many applications of nonparametric tests based on curve estimation involve selecting a smoothing parameter. The author proposes an adaptive test that combines several generalized likelihood ratio tests in order to get power performance nearly equal to whichever of the component tests is best. She derives the asymptotic joint distribution of the component tests and that of the proposed test under the null hypothesis. She also develops a simple method of selecting the smoothing parameters for the proposed test and presents two approximate methods for obtaining its P‐value. Finally, she evaluates the proposed test through simulations and illustrates its application to a set of real data.  相似文献   

15.
Some distribution-free methods are suggested in the paper for testing the hypothesis about the slope parameter in a one-sample linear regression model with multiple observations at each level of independent variable. Asymptotic relative efficiencies of these tests are discussed, and the tests are compared with their nonparametric competitors.  相似文献   

16.
ABSTRACT

This article examines the evidence contained in t statistics that are marginally significant in 5% tests. The bases for evaluating evidence are likelihood ratios and integrated likelihood ratios, computed under a variety of assumptions regarding the alternative hypotheses in null hypothesis significance tests. Likelihood ratios and integrated likelihood ratios provide a useful measure of the evidence in favor of competing hypotheses because they can be interpreted as representing the ratio of the probabilities that each hypothesis assigns to observed data. When they are either very large or very small, they suggest that one hypothesis is much better than the other in predicting observed data. If they are close to 1.0, then both hypotheses provide approximately equally valid explanations for observed data. I find that p-values that are close to 0.05 (i.e., that are “marginally significant”) correspond to integrated likelihood ratios that are bounded by approximately 7 in two-sided tests, and by approximately 4 in one-sided tests.

The modest magnitude of integrated likelihood ratios corresponding to p-values close to 0.05 clearly suggests that higher standards of evidence are needed to support claims of novel discoveries and new effects.  相似文献   

17.
We propose a class of goodness-of-fit tests for the gamma distribution that utilizes the empirical Laplace transform. The consistency of the tests as well as their asymptotic distribution under the null hypothesis are investigated. As the decay of the weight function tends to infinity, the test statistics approach limit values related to the first non zero component of Neyman's smooth test for the gamma law. The new tests are compared with other omnibus tests for the gamma distribution.  相似文献   

18.
The objective of this article is to propose and study frequentist tests that have maximum average power, averaging with respect to some specified weight function. First, some relationships between these tests, called maximum average-power (MAP) tests, and most powerful or uniformly most powerful tests are presented. Second, the existence of a maximum average-power test for any hypothesis testing problem is shown. Third, an MAP test for any hypothesis testing problem with a simple null hypothesis is constructed, including some interesting classical examples. Fourth, an MAP test for a hypothesis testing problem with a composite null hypothesis is discussed. From any one-parameter exponential family, a commonly used UMPU test is shown to be also an MAP test with respect to a rich class of weight functions. Finally, some remarks are given to conclude the article.  相似文献   

19.
Peto and Peto (1972) have studied rank invariant tests to compare two survival curves for right censored data. We apply their tests, including the logrank test and the generalized Wilcoxon test, to left truncated and interval censored data. The significance levels of the tests are approximated by Monte Carlo permutation tests. Simulation studies are conducted to show their size and power under different distributional differences. In particular, the logrank test works well under the Cox proportional hazards alternatives, as for the usual right censored data. The methods are illustrated by the analysis of the Massachusetts Health Care Panel Study dataset.  相似文献   

20.
A number of statistical tests have been recommended over the last twenty years for assessing the randomness of long binary strings used in cryptographic algorithms. Several of these tests include methods of examining subblock patterns. These tests are the uniformity test, the universal test and the repetition test. The effectiveness of these tests are compared based on the subblock length, the limitations on data requirements, and on their power in detecting deviations from randomness. Due to the complexity of the test statistics, the power functions are estimated by simulation methods. The results show that for small subblocks the uniformity test is more powerful than the universal test, and that there is some doubt about the parameters of the hypothesised distribution for the universal test statistic. For larger subblocks the results show that the repetition test is the most effective test, since it requires far less data than either of the other two tests and is an efficient test in detecting deviations from randomness in binary strings.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号