首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
The exact and asymptotic upper tail probabilities ( α= .l0, .05, .01, .001) of the three chi-squared goodness-of-fit statistics Pearson's X 2, likelihood ratioG 2, and power-divergence statisticD 2 (λ ) , with λ = 2/3, are compared numerically for simple null hypotheses not involving parameter estimation. Three types of such hypotheses were investigated (equal cell probabilities, proportional cell probabilities, some fixed small expectations together with some increasing large expectations) for the number of cells being between 3 and 15, and for sample sizes from 10 to 40, increasing by steps of one. Rating the relative accuracy of the chi-squared approximation in terms of ±10% and ±20% intervals around α led to the following conclusions: 1. Using G 2 is not recommended. 2 . At the more relevant significance levels α = .10 and α = .05X 2 should be preferred over D 2. Solely in case of unequal cell probabilitiesD 2 is the better choice at α = .O1 and α = .001. 3 . Yarnold's (1970; Journal of the Amerin Statistical Association, 65, 864-886) rule for the minimum expectation when using X 2 ("If the number of cells k is 3 or more, and if r denotes the number of expectations less than 5, then the minimum expectation may be as small as 5r/k.") generalizes to D 2; it gives a good lower limit for the expected cell frequencies, however, when the number of cells is greater than 3. For k = 3 , even sample sizes over 15 may be insufficient.  相似文献   

2.
I am concerned with the admissibility under quadratic loss of certain estimators of binomial probabilities. The minimum variance unbiased estimator is shown to be admissible for Pr(X = 0) and Pr(X = n), but it is inadmissible for Pr(X = k), where 0 < k < n. An example is given of an admissible maximum likelihood estimator (MLE). It is conjectured that the MLE is always admissible.  相似文献   

3.
Exact unconditional tests for comparing two binomial probabilities are generally more powerful than conditional tests like Fisher's exact test. Their power can be further increased by the Berger and Boos confidence interval method, where a p-value is found by restricting the common binomial probability under H 0 to a 1?γ confidence interval. We studied the average test power for the exact unconditional z-pooled test for a wide range of cases with balanced and unbalanced sample sizes, and significance levels 0.05 and 0.01. The detailed results are available online on the web. Among the values 10?3, 10?4, …, 10?10, the value γ=10?4 gave the highest power, or close to the highest power, in all the cases we looked at, and can be given as a general recommendation as an optimal γ.  相似文献   

4.
The small-sample accuracy of seven members of the family of power-divergence statistics for testing independence or homogeneity in contingency tables was studied via simulation. The likelihood ratio statistic G 2 and Pearson's X 2 statistic are among these seven members, whose behavior was studied at nominal test sizes of.01 and.05 with marginal distributions that could be uniform or skewed and with a set of sample sizes that included sparseness conditions as measured through table density (i.e., the ratio of sample size to number of cells). The likelihood ratio statistic G 2 rejected the null hypothesis too often even with large table density, whereas Pearson's X 2 was sufficiently accurate and only presented a minor misbehavior when table density was less than two observations/cell. None of the other five statistics outperformed Pearson's X 2. A nonasymptotic variant of X 2 solved the minor inaccuracies of Pearson's X 2 and turned out to be the most accurate statistic for testing independence or homogeneity, even with table densities of one observation/cell. These results clearly advise against the use of the likelihood ratio statistic G 2.  相似文献   

5.
Consider a family of square-integrable Rd-valued statistics Sk = Sk(X1,k1; X2,k2;…; Xm,km), where the independent samples Xi,kj respectively have ki i.i.d. components valued in some separable metric space Xi. We prove a strong law of large numbers, a central limit theorem and a law of the iterated logarithm for the sequence {Sk}, including both the situations where the sample sizes tend to infinity while m is fixed and those where the sample sizes remain small while m tends to infinity. We also obtain two almost sure convergence results in both these contexts, under the additional assumption that Sk is symmetric in the coordinates of each sample Xi,kj. Some extensions to row-exchangeable and conditionally independent observations are provided. Applications to an estimator of the dimension of a data set and to the Henze-Schilling test statistic for equality of two densities are also presented.  相似文献   

6.
Abstract

Through simulation and regression, we study the alternative distribution of the likelihood ratio test in which the null hypothesis postulates that the data are from a normal distribution after a restricted Box–Cox transformation and the alternative hypothesis postulates that they are from a mixture of two normals after a restricted (possibly different) Box–Cox transformation. The number of observations in the sample is called N. The standardized distance between components (after transformation) is D = (μ2 ? μ1)/σ, where μ1 and μ2 are the component means and σ2 is their common variance. One component contains the fraction π of observed, and the other 1 ? π. The simulation results demonstrate a dependence of power on the mixing proportion, with power decreasing as the mixing proportion differs from 0.5. The alternative distribution appears to be a non-central chi-squared with approximately 2.48 + 10N ?0.75 degrees of freedom and non-centrality parameter 0.174N(D ? 1.4)2 × [π(1 ? π)]. At least 900 observations are needed to have power 95% for a 5% test when D = 2. For fixed values of D, power, and significance level, substantially more observations are necessary when π ≥ 0.90 or π ≤ 0.10. We give the estimated powers for the alternatives studied and a table of sample sizes needed for 50%, 80%, 90%, and 95% power.  相似文献   

7.
Let {X, Xn; n ≥ 1} be a sequence of real-valued iid random variables, 0 < r < 2 and p > 0. Let D = { A = (ank; 1 ≤ kn, n ≥ 1); ank, ? R and supn, k |an,k| < ∞}. Set Sn( A ) = ∑nk=1an, kXk for A ? D and n ≥ 1. This paper is devoted to determining conditions whereby E{supn ≥ 1, |Sn( A )|/n1/r}p < ∞ or E{supn ≥ 2 |Sn( A )|/2n log n)1/2}p < ∞ for every A ? D. This generalizes some earlier results, including those of Burkholder (1962), Choi and Sung (1987), Davis (1971), Gut (1979), Klass (1974), Siegmund (1969) and Teicher (1971).  相似文献   

8.
In this article, we propose two test statistics for testing the underlying serial correlation in a partially linear single-index model Y = η(Z τα) + X τβ + ? when X is measured with additive error. The proposed test statistics are shown to have asymptotic normal or chi-squared distributions under the null hypothesis of no serial correlation. Monte Carlo experiments are also conducted to illustrate the finite sample performance of the proposed test statistics. The simulation results confirm that these statistics perform satisfactorily in both estimated sizes and powers.  相似文献   

9.
Power-divergence goodness-of-fit statistics have asymptotically a chi-squared distribution. Asymptotic results may not apply in small-sample situations, and the exact significance of a goodness-of-fit statistic may potentially be over- or under-stated by the asymptotic distribution. Several correction terms have been proposed to improve the accuracy of the asymptotic distribution, but their performance has only been studied for the equiprobable case. We extend that research to skewed hypotheses. Results are presented for one-way multinomials involving k = 2 to 6 cells with sample sizes N = 20, 40, 60, 80 and 100 and nominal test sizes f = 0.1, 0.05, 0.01 and 0.001. Six power-divergence goodness-of-fit statistics were investigated, and five correction terms were included in the study. Our results show that skewness itself does not affect the accuracy of the asymptotic approximation, which depends only on the magnitude of the smallest expected frequency (whether this comes from a small sample with the equiprobable hypothesis or a large sample with a skewed hypothesis). Throughout the conditions of the study, the accuracy of the asymptotic distribution seems to be optimal for Pearson's X2 statistic (the power-divergence statistic of index u = 1) when k > 3 and the smallest expected frequency is as low as between 0.1 and 1.5 (depending on the particular k, N and nominal test size), but a computationally inexpensive improvement can be obtained in these cases by using a moment-corrected h2 distribution. If the smallest expected frequency is even smaller, a normal correction yields accurate tests through the log-likelihood-ratio statistic G2 (the power-divergence statistic of index u = 0).  相似文献   

10.
For X with binomial (n, p) distribution the usual measure of the error of X/n as an estimator of p is its standard error Sn(p) = √{E(X/n – p)2} = √{p(1 – p)/n}. A somewhat more natural measure is the average absolute error Dn(p) = E‖X/n – p‖. This article considers use of Dn(p) instead of Sn(p) in a student's first introduction to statistical estimation. Exact and asymptotic values of Dn(p), and the appearance of its graph, are described in detail. The same is done for the Poisson distribution.  相似文献   

11.
Small sample properties of seven confidence intervals for the binomial parameterp (based on various normal approximations) and of the Clopper-Pearson interval are compared. Coverage probabilities and expected lower and upper limits of the intervals are graphically displayed as functions of the binomial parameterp for various sample sizes.  相似文献   

12.
This paper rejects the preference expressed by Berkson for the heuristic test statistic TN with standard normal distribution testing equality of two binomial probabilities in favour of Fisher's conditional exact test statistic TE. Conditioning upon k1 + k2 = k shows that TN admits too large first kind error probabilities. But also unconditionally TN is systematically too large compared to TE, giving too small critical levels at given experimental outcomes and power which is misleadingly too large. This is mainly due to the fact that TN is not suitably corrected for continuity (at small sample sizes).  相似文献   

13.
In this paper, we obtain some results for the asymptotic behavior of the tail probability of a random sum Sτ = ∑τk = 1Xk, where the summands Xk, k = 1, 2, …, are conditionally dependent random variables with a common subexponential distribution F, and the random number τ is a non negative integer-valued random variable, independent of {Xk: k ? 1}.  相似文献   

14.
ABSTRACT

In this article, we consider a (k + 1)n-dimensional elliptically contoured random vector (XT1, X2T, …, XTk, ZT)T = (X11, …, X1n, …, Xk1, …, Xkn, Z1, …, Zn)T and derive the distribution of concomitant of multivariate order statistics arising from X1, X2, …, Xk. Specially, we derive a mixture representation for concomitant of bivariate order statistics. The joint distribution of the concomitant of bivariate order statistics is also obtained. Finally, the usefulness of our result is illustrated by a real-life data.  相似文献   

15.
The problem of inference in Bayesian Normal mixture models is known to be difficult. In particular, direct Bayesian inference (via quadrature) suffers from a combinatorial explosion in having to consider every possible partition of n observations into k mixture components, resulting in a computation time which is O(k n). This paper explores the use of discretised parameters and shows that for equal-variance mixture models, direct computation time can be reduced to O(D k n k), where relevant continuous parameters are each divided into D regions. As a consequence, direct inference is now possible on genuine data sets for small k, where the quality of approximation is determined by the level of discretisation. For large problems, where the computational complexity is still too great in O(D k n k) time, discretisation can provide a convergence diagnostic for a Markov chain Monte Carlo analysis.  相似文献   

16.
G = F k (k > 1); G = 1 − (1−F) k (k < 1); G = F k (k < 1); and G = 1 − (1−F) k (k > 1), where F and G are two continuous cumulative distribution functions. If an optimal precedence test (one with the maximal power) is determined for one of these four classes, the optimal tests for the other classes of alternatives can be derived. Application of this is given using the results of Lin and Sukhatme (1992) who derived the best precedence test for testing the null hypothesis that the lifetimes of two types of items on test have the same distibution. The test has maximum power for fixed κ in the class of alternatives G = 1 − (1−F) k , with k < 1. Best precedence tests for the other three classes of Lehmann-type alternatives are derived using their results. Finally, a comparison of precedence tests with Wilcoxon's two-sample test is presented. Received: February 22, 1999; revised version: June 7, 2000  相似文献   

17.
The mixture of Rayleigh random variables X 1and X 2 are identified in terms of relations between the conditional expectation of ( X2:22 -X1:22)r{\left( {X_{2:2}^2 -X_{1:2}^2}\right)^{r}} given X 1:2 (or X2:22k{X_{2:2}^{2k}} given X1:2,"kr){X_{1:2},\forall k\leq r)} and hazard rate function of the distribution, where X 1:2 and X 2:2 denote the corresponding order statistics, r is a positive integer. In addition, we also mention some related theorems to characterize the mixtures of Rayleigh distributions. Finally, we also give an application to Multi-Hit models of carcinogenesis (Parallel Systems) and a simulated example is used to illustrate our results.  相似文献   

18.
When a sample discriminant function is computed, it is desired to estimate the error rate using this function. This is often done by computing G(-D/2), where G is the cumulative normal distribution and D2 is the estimated Mahalanobis' distance. In this paper an asymptotic expansion of the expectation of G(-D/2) is derived and is compared with existing Monte Carlo estimates. The asymptotic bias of G(-D/2) is derived also and the well-known practical result that G(-D/2) gives too favourable an estimate of the true error rate  相似文献   

19.
A simulation study was done to compare seven confidence interval methods, based on the normal approximation, for the difference of two binomial probabilities. Cases considered included minimum expected cell sizes ranging from 2 to 15 and smallest group sizes (NMIN) ranging from 6 to 100. Our recommendation is to use a continuity correction of 1/(2 NMIN) combined with the use of (N ? 1) rather than N in the estimate of the standard error. For all of the cases considered with minimum expected cell size of at least 3, this method gave coverage probabilities close to or greater than the nominal 90% and 95%. The Yates method is also acceptable, but it is slightly more conservative. At the other extreme, the usual method (with no continuity correction) does not provide adequate coverage even at the larger sample sizes. For the 99% intervals, our recommended method and the Yates correction performed equally well and are reasonable for minimum expected cell sizes of at least 5. None of the methods performed consistently well for a minimum expected cell size of 2.  相似文献   

20.
ABSTRACT

Least squares estimator of the stability parameter ? ? |α| + |β| for a spatial unilateral autoregressive process Xk, ? = αXk ? 1, ? + βXk, ? ? 1 + ?k, ? is investigated and asymptotic normality with a scaling factor n5/4 is shown in the unstable case ? = 1. The result is in contrast to the unit root case of the AR(p) model Xk = α1Xk ? 1 + ??? + αpXk ? p + ?k, where the limiting distribution of the least squares estimator of the unit root parameter ? ? α1 + ??? + αp is not normal.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号