首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
Generally, confidence regions for the probabilities of a multinomial population are constructed based on the Pearson χ2 statistic. Morales et al. (Bootstrap confidence regions in multinomial sampling. Appl Math Comput. 2004;155:295–315) considered the bootstrap and asymptotic confidence regions based on a broader family of test statistics known as power-divergence test statistics. In this study, we extend their work and propose penalized power-divergence test statistics-based confidence regions. We only consider small sample sizes where asymptotic properties fail and alternative methods are needed. Both bootstrap and asymptotic confidence regions are constructed. We consider the percentile and the bias corrected and accelerated bootstrap confidence regions. The latter confidence region has not been studied previously for the power-divergence statistics much less for the penalized ones. Designed simulation studies are carried out to calculate average coverage probabilities. Mean absolute deviation between actual and nominal coverage probabilities is used to compare the proposed confidence regions.  相似文献   

2.
Power-divergence test statistics have been considered to test linear by linear association for two-way contingency tables. These test statistics have been compared based on designed simulation study and asymptotic results for 2 × 2, 2 × 3, and 3 × 3 tables. According to the results, there are test statistics with better properties than the well-known likelihood ratio test statistic for small and moderate samples.  相似文献   

3.
Two-different types of adjustments to the power-divergence test statistics have been introduced for the problem of testing goodness-of-fit under clustered sampling. Penalization has also been introduced to handle the cells with zero frequencies. The asymptotic distribution of the proposed power-divergence test statistics has been investigated under clustered sampling and the performances of the proposed statistics for finite samples have been studied through a designed simulation study.  相似文献   

4.
Outlier detection is fundamental to statistical modelling. When there are multiple outliers, many traditional approaches in use are stepwise detection procedures, which can be computationally expensive and ignore stochastic error in the outlier detection process. Outlier detection can be performed by a heteroskedasticity test. In this article, a rapid outlier detection method via multiple heteroskedasticity test based on penalized likelihood approaches is proposed to handle these kinds of problems. The proposed method detects the heteroskedasticity of all data only by one step and estimate coefficients simultaneously. The proposed approach is distinguished from others in that a rapid modelling approach uses a weighted least squares formulation coupled with nonconvex sparsity-including penalization. Furthermore, the proposed approach does not need to construct test statistics and calculate their distributions. A new algorithm is proposed for optimizing penalized likelihood functions. Favourable theoretical properties of the proposed approach are obtained. Our simulation studies and real data analysis show that the newly proposed methods compare favourably with other traditional outlier detection techniques.  相似文献   

5.
We consider the problem of detecting a ‘bump’ in the intensity of a Poisson process or in a density. We analyze two types of likelihood ratio‐based statistics, which allow for exact finite sample inference and asymptotically optimal detection: The maximum of the penalized square root of log likelihood ratios (‘penalized scan’) evaluated over a certain sparse set of intervals and a certain average of log likelihood ratios (‘condensed average likelihood ratio’). We show that penalizing the square root of the log likelihood ratio — rather than the log likelihood ratio itself — leads to a simple penalty term that yields optimal power. The thus derived penalty may prove useful for other problems that involve a Brownian bridge in the limit. The second key tool is an approximating set of intervals that is rich enough to allow for optimal detection, but which is also sparse enough to allow justifying the validity of the penalization scheme simply via the union bound. This results in a considerable simplification in the theoretical treatment compared with the usual approach for this type of penalization technique, which requires establishing an exponential inequality for the variation of the test statistic. Another advantage of using the sparse approximating set is that it allows fast computation in nearly linear time. We present a simulation study that illustrates the superior performance of the penalized scan and of the condensed average likelihood ratio compared with the standard scan statistic.  相似文献   

6.
Power-divergence goodness-of-fit statistics have asymptotically a chi-squared distribution. Asymptotic results may not apply in small-sample situations, and the exact significance of a goodness-of-fit statistic may potentially be over- or under-stated by the asymptotic distribution. Several correction terms have been proposed to improve the accuracy of the asymptotic distribution, but their performance has only been studied for the equiprobable case. We extend that research to skewed hypotheses. Results are presented for one-way multinomials involving k = 2 to 6 cells with sample sizes N = 20, 40, 60, 80 and 100 and nominal test sizes f = 0.1, 0.05, 0.01 and 0.001. Six power-divergence goodness-of-fit statistics were investigated, and five correction terms were included in the study. Our results show that skewness itself does not affect the accuracy of the asymptotic approximation, which depends only on the magnitude of the smallest expected frequency (whether this comes from a small sample with the equiprobable hypothesis or a large sample with a skewed hypothesis). Throughout the conditions of the study, the accuracy of the asymptotic distribution seems to be optimal for Pearson's X2 statistic (the power-divergence statistic of index u = 1) when k > 3 and the smallest expected frequency is as low as between 0.1 and 1.5 (depending on the particular k, N and nominal test size), but a computationally inexpensive improvement can be obtained in these cases by using a moment-corrected h2 distribution. If the smallest expected frequency is even smaller, a normal correction yields accurate tests through the log-likelihood-ratio statistic G2 (the power-divergence statistic of index u = 0).  相似文献   

7.
For square contingency tables with ordered categories, there may be some cases that one wants to analyze them by considering collapsed tables with some adjacent categories combined in the original table. This paper considers the symmetry model for collapsed square contingency tables and proposes a measure to represent the degree of departure from symmetry. The proposed measure is defined as the arithmetic mean of submeasures each of which represents the degree of departure from symmetry for each collapsed 3×3 table. Each submeasure also represents the mean of power-divergence or diversity index for each collapsed table. Examples are given.  相似文献   

8.
In this article, we propose two test statistics for testing the underlying serial correlation in a partially linear single-index model Y = η(Z τα) + X τβ + ? when X is measured with additive error. The proposed test statistics are shown to have asymptotic normal or chi-squared distributions under the null hypothesis of no serial correlation. Monte Carlo experiments are also conducted to illustrate the finite sample performance of the proposed test statistics. The simulation results confirm that these statistics perform satisfactorily in both estimated sizes and powers.  相似文献   

9.
The small-sample behaviour of power-divergence goodness-of-fit statistics with composite hypotheses was evaluated with multinomial models of up to five cells and up to three parameters. Their performance was assessed by comparing asymptotic test sizes with exact test sizes obtained by enumeration in the near right tail, where 1-?∈?(0.90,?0.95], and in the far right tail, where 1-?∈?(0.95,?0.99]. The study addressed all combinations of power-diparse JAS312HH01.sgmvergence estimates of indices ν?∈?{-1/2,?0,?1/3,?1/2,?2/3,?1,?3/ 2} and power-divergence statistics of indices λ?∈?{-1/2,?0,?1/3,?1/2,?2/3,?1,?3/2}. The results indicate that the asymptotic approximation is sufficiently accurate (by the criterion that the average exact size is no larger than ±10% of the nominal asymptotic test size) in the near right tail when ν=0 and λ=1/2, and in the far right tail when ν=0 and λ=1/3, in both cases providing that the smallest expectation in the composite hypothesis exceeds 5. The only exception to this rule is the case of models that render a near-equiprobable composite hypothesis on the boundaries of the parameter space, where average exact sizes are usually quite different from nominal sizes despite the fact that the smallest expectation in these conditions is usually well above 5.  相似文献   

10.
This article considers K pairs of incomplete correlated 2 × 2 tables in which the interesting measurement is the risk difference between marginal and conditional probabilities. A Wald-type statistic and a score-type statistic are presented to test the homogeneity hypothesis about risk differences across strata. Powers and sample size formulae based on the above two statistics are deduced. Figures about sample size against risk difference (or marginal probability) are given. A real example is used to illustrate the proposed methods.  相似文献   

11.

Cressie et al. (2000; 2003) introduced and studied a new family of statistics, based on the φ-divergence measure, for solving the problem of testing a nested sequence of loglinear models. In that family of test statistics the parameters are estimated using the minimum φ-divergence estimator which is a generalization of the maximum likelihood estimator. In this paper we study the minimum power-divergence estimator (the most important family of minimum φ-divergence estimator) for a nested sequence of loglinear models in three-way contingency tables under assumptions of multinomial sampling. A simulation study illustrates that the minimum chi-squared estimator is simultaneously the most robust and efficient estimator among the family of the minimum power-divergence estimator.  相似文献   

12.
In this paper we first show that the k-sample Anderson–Darling test is basically an average of Pearson statistics in 2?×?k contingency tables that are induced by observation-based partitions of the sample space. As an extension, we construct a family of rank test statistics, indexed by c?∈??, which is based on similarly constructed c?×?k partitions. An extensive simulation study, in which we compare the new test with others, suggests that generally very high powers are obtained with the new tests. Finally we propose a decomposition of the test statistic in interpretable components.  相似文献   

13.
The efficiency of the penalized methods (Fan and Li, 2001 Fan , J. , Li , R. ( 2001 ). Variable selection via nonconcave penalized likelihood and its oracle properties . Journal of the American Statistical Association 96 : 13481360 .[Taylor & Francis Online], [Web of Science ®] [Google Scholar]) depends strongly on a tuning parameter due to the fact that it controls the extent of penalization. Therefore, it is important to select it appropriately. In general, tuning parameters are chosen by data-driven approaches, such as the commonly used generalized cross validation. In this article, we propose an alternative method for the derivation of the tuning parameter selector in penalized least squares framework, which can lead to an ameliorated estimate. Simulation studies are presented to support theoretical findings and a comparison of the Type I and Type II error rates, considering the L 1, the hard thresholding and the Smoothly Clipped Absolute Deviation penalty functions, is performed. The results are given in tables and discussion follows.  相似文献   

14.
In this paper we introduce a new measure for the analysis of association in cross-classifications having ordered categories. Association is measured in terms of the odd-ratios in 2 × 2 subtables formed from adjacent rows and adjacent columns. We focus our attention in the uniform association model. Our measure is based in the family of divergences introduced by Burbea and Rao [1] Burbea, J. and Rao, C. R. 1982a. On the convexity of some divergence measures based on entropy functions. IEEE Transactions on Information Theory, 28: 489495. [Crossref], [Web of Science ®] [Google Scholar]. Some well-known sets of data are reanalyzed and a simulation study is presented to analyze the behavior of the new families of test statistics introduced in this paper.  相似文献   

15.
Pearson’s chi-square (Pe), likelihood ratio (LR), and Fisher (Fi)–Freeman–Halton test statistics are commonly used to test the association of an unordered r×c contingency table. Asymptotically, these test statistics follow a chi-square distribution. For small sample cases, the asymptotic chi-square approximations are unreliable. Therefore, the exact p-value is frequently computed conditional on the row- and column-sums. One drawback of the exact p-value is that it is conservative. Different adjustments have been suggested, such as Lancaster’s mid-p version and randomized tests. In this paper, we have considered 3×2, 2×3, and 3×3 tables and compared the exact power and significance level of these test’s standard, mid-p, and randomized versions. The mid-p and randomized test versions have approximately the same power and higher power than that of the standard test versions. The mid-p type-I error probability seldom exceeds the nominal level. For a given set of parameters, the power of Pe, LR, and Fi differs approximately the same way for standard, mid-p, and randomized test versions. Although there is no general ranking of these tests, in some situations, especially when averaged over the parameter space, Pe and Fi have the same power and slightly higher power than LR. When the sample sizes (i.e., the row sums) are equal, the differences are small, otherwise the observed differences can be 10% or more. In some cases, perhaps characterized by poorly balanced designs, LR has the highest power.  相似文献   

16.
We consider the optimal configuration of a square array group testing algorithm (denoted A2) to minimize the expected number of tests per specimen. For prevalence greater than 0.2498, individual testing is shown to be more efficient than A2. For prevalence less than 0.2498, closed form lower and upper bounds on the optimal group sizes for A2 are given. Arrays of dimension 2 × 2, 3 × 3, and 4 × 4 are shown to never be optimal. The results are illustrated by considering the design of a specimen pooling algorithm for detection of recent HIV infections in Malawi.  相似文献   

17.
The incidence of most diseases is low enough that in. large populations the number of new cases may be considered a Poisson variate. This paper explores models and methods for analyzing such data Specific cases are the estimation and testing of ratios and the cross-product ratios, both simple and stratified* We assume the Poisson means are exponential functions of the relevant parameters. The resulting sets of sufficient statistics are partitioned into a test statistic and a vector of statistics related to the nuisance parameters . The methods derived are based on the conditional distribution of the test statistic given the other sufficient statistics. The analyses of stratified cross-product ratios are seen to be analogues of the noncentral distribution associated with theanalysis of the common odds ratio in several 2×2 tables. The various methods are illustrated in numerical examples involving incidence rates of cancer in two metropolitan areas adjusting for both age and sex.  相似文献   

18.
M-estimation is a widely used method for robust statistical inference. In this article, using a B-spline series approximation with a double smoothly clipped absolute deviation penalization, we solve the problem of simultaneous variable selection and parametric component identification in a non parametric additive model. The theoretical properties of the double non concave penalized M-estimation are established. The proposed approach is resistant to heavy-tailed errors or outliers in the responses. Simulation studies for finite-sample cases are conducted and a real dataset is also analyzed for illustration of this new approach.  相似文献   

19.
Summary.  We consider the problem of testing null hypotheses that include restrictions on the variance component in a linear mixed model with one variance component and we derive the finite sample and asymptotic distribution of the likelihood ratio test and the restricted likelihood ratio test. The spectral representations of the likelihood ratio test and the restricted likelihood ratio test statistics are used as the basis of efficient simulation algorithms of their null distributions. The large sample χ 2 mixture approximations using the usual asymptotic theory for a null hypothesis on the boundary of the parameter space have been shown to be poor in simulation studies. Our asymptotic calculations explain these empirical results. The theory of Self and Liang applies only to linear mixed models for which the data vector can be partitioned into a large number of independent and identically distributed subvectors. One-way analysis of variance and penalized splines models illustrate the results.  相似文献   

20.
Several tests for testing independence of 2 by 2 contingency tables have been proposed over the years. Cressie and Read (1984) identified several of these tests as members of a power-divergence family, and much of the characteristics for these tests are unified. However, the question of which test is best is still not fully understood. This paper provides algorithms for chi-square estimates and investigates the convergence rates and powers of these chi-square tests.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号