首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 78 毫秒
1.
Kappa and B assess agreement between two observers independently classifying N units into k categories. We study their behavior under zero cells in the contingency table and unbalanced asymmetric marginal distributions. Zero cells arise when a cross-classification is never endorsed by both observers; biased marginal distributions occur when some categories are preferred differently between the observers. Simulations studied the distributions of the unweighted and weighted statistics for k=4, under fixed proportions of diagonal agreement and different patterns off-diagonal, with various sample sizes, and under various zero cell count scenarios. Marginal distributions were first uniform and homogeneous, and then unbalanced asymmetric distributions. Results for unweighted kappa and B statistics were comparable to work of Muñoz and Bangdiwala, even with zero cells. A slight increased variation was observed as the sample size decreased. Weighted statistics did show greater variation as the number of zero cells increased, with weighted kappa increasing substantially more than weighted B. Under biased marginal distributions, weighted kappa with Cicchetti weights were higher than with squared weights. Both statistics for observer agreement behaved well under zero cells. The weighted B was less variable than the weighted kappa under similar circumstances and different weights. In general, B's performance and graphical interpretation make it preferable to kappa under the studied scenarios.  相似文献   

2.
The kappa coefficient is a widely used measure for assessing agreement on a nominal scale. Weighted kappa is an extension of Cohen's kappa that is commonly used for measuring agreement on an ordinal scale. In this article, it is shown that weighted kappa can be computed as a function of unweighted kappas. The latter coefficients are kappa coefficients that correspond to smaller contingency tables that are obtained by merging categories.  相似文献   

3.
We consider a 2×2 contingency table, with dichotomized qualitative characters (A,A) and (B,B), as a sample of size n drawn from a bivariate binomial (0,1) distribution. Maximum likelihood estimates p?1p?2 and p? are derived for the parameters of the two marginals p1p2 and the coefficient of correlation. It is found that p? is identical to Pearson's (1904)?=(χ2/n)½, where ?2 is Pearson's usual chi-square for the 2×2 table. The asymptotic variance-covariance matrix of p?lp?2and p is also derived.  相似文献   

4.
We consider the problem of estimating and testing a general linear hypothesis in a general multivariate linear model, the so-called Growth Curve model, when the p × N observation matrix is normally distributed.

The maximum likelihood estimator (MLE) for the mean is a weighted estimator with the inverse of the sample covariance matrix which is unstable for large p close to N and singular for p larger than N. We modify the MLE to an unweighted estimator and propose new tests which we compare with the previous likelihood ratio test (LRT) based on the weighted estimator, i.e., the MLE. We show that the performance of these new tests based on the unweighted estimator is better than the LRT based on the MLE.  相似文献   


5.
In this paper we first show that the k-sample Anderson–Darling test is basically an average of Pearson statistics in 2?×?k contingency tables that are induced by observation-based partitions of the sample space. As an extension, we construct a family of rank test statistics, indexed by c?∈??, which is based on similarly constructed c?×?k partitions. An extensive simulation study, in which we compare the new test with others, suggests that generally very high powers are obtained with the new tests. Finally we propose a decomposition of the test statistic in interpretable components.  相似文献   

6.
When an r×c contingency table has many cells having very small expectations, the usual χ2 approximation to the upper tail of the Pearson χ2 goodness-of-fit statistic becomes very conservative. The alternatives considered in this paper are to use either a lognormal approximation, or to scale the usual χ2 approximation. The study involves thousands of tables with various sample sizes, and with tables whose sizes range from 2×2 through 2×10×10. Subject to certain restrictions the new scaled χ2 approximations are recommended for use with tables having an average cell expectation as small as 0·5.  相似文献   

7.
Cohen’s kappa is a weighted average   总被引:1,自引:0,他引:1  
  相似文献   

8.
For an R×R square contingency table with nominal categories, the present paper proposes a model which indicates that the absolute values of log odds of the odds ratio for rows i and j and columns j and R to the corresponding symmetric odds ratio for rows j and R and columns i and j are constant for every i<j<R. The model is an extension of the quasi-symmetry model and states a structure of asymmetry of odds ratios. An example is given.  相似文献   

9.
J. Anděl  I. Netuka 《Statistics》2013,47(4):279-287
The article deals with methods for computing the stationary marginal distribution in linear models of time series. Two approaches are described. First, an algorithm based on approximation of solution of the corresponding integral equation is briefly reviewed. Then, we study the limit behaviour of the partial sums c 1 η1+c 2 η2+···+c n η n where η i are i.i.d. random variables and c i real constants. We generalize procedure of Haiman (1998) [Haiman, G., 1998, Upper and lower bounds for the tail of the invariant distribution of some AR(1) processes. Asymptotic Methods in Probability and Statistics, 45, 723–730.] to an arbitrary causal linear process and relax the assumptions of his result significantly. This is achieved by investigating the properties of convolution of densities.  相似文献   

10.
Monte Carlo simulations were done to estimate the means and standard deviations of the characteristic roots of a Wishart matrix which can be used in computing tests of hypotheses concerning multiplicative terms in balanced linear-bilinear (multiplicative) models for an m × n table of data. In this report we extend the previous results (Mandel, 1971; Cornelius, 1980) to r ≤ 199, c ≤ 149 or r ≤ 149, c ≤ 199, where r and c are row and column degrees of freedom, respectively, of the two-way array of residuals (with total degrees of freedom rc) after fitting the linear effects. For 187 combinations of r and c at intervals over this domain, we used 5000 simulations to estimate expectations and standard deviations of the Wishart roots. Using weighted linear regression variable selection techniques, symmetric functions of r and c were obtained for approximating the simulated means and standard deviations. Use of these approximating functions will avoid the need for reference to tables for input to computer programs which require these values for tests of significance of sequentially fitted terms in the analyses of balanced linear-bilinear models.  相似文献   

11.
Pearson’s chi-square (Pe), likelihood ratio (LR), and Fisher (Fi)–Freeman–Halton test statistics are commonly used to test the association of an unordered r×c contingency table. Asymptotically, these test statistics follow a chi-square distribution. For small sample cases, the asymptotic chi-square approximations are unreliable. Therefore, the exact p-value is frequently computed conditional on the row- and column-sums. One drawback of the exact p-value is that it is conservative. Different adjustments have been suggested, such as Lancaster’s mid-p version and randomized tests. In this paper, we have considered 3×2, 2×3, and 3×3 tables and compared the exact power and significance level of these test’s standard, mid-p, and randomized versions. The mid-p and randomized test versions have approximately the same power and higher power than that of the standard test versions. The mid-p type-I error probability seldom exceeds the nominal level. For a given set of parameters, the power of Pe, LR, and Fi differs approximately the same way for standard, mid-p, and randomized test versions. Although there is no general ranking of these tests, in some situations, especially when averaged over the parameter space, Pe and Fi have the same power and slightly higher power than LR. When the sample sizes (i.e., the row sums) are equal, the differences are small, otherwise the observed differences can be 10% or more. In some cases, perhaps characterized by poorly balanced designs, LR has the highest power.  相似文献   

12.
Let {X t , t ∈ ?} be a sequence of iid random variables with an absolutely continuous distribution. Let a > 0 and c ∈ ? be some constants. We consider a sequence of 0-1 valued variables {ξ t , t ∈ ?} obtained by clipping an MA(1) process X t  ? aX t?1 at the level c, i.e., ξ t  = I[X t  ? aX t?1 < c] for all t ∈ ?. We deal with the estimation problem in this model. Properties of the estimators of the parameters a and c, the success probability p, and the 1-lag autocorrelation r 1 are investigated. A numerical study is provided as an illustration of the theoretical results.  相似文献   

13.
Trend tests in dose-response have been central problems in medicine. The likelihood ratio test is often used to test hypotheses involving a stochastic order. Stratified contingency tables are common in practice. The distribution theory of likelihood ratio test has not been full developed for stratified tables and more than two stochastically ordered distributions. Under c strata of m × r tables, for testing the conditional independence against simple stochastic order alternative, this article introduces a model-free test method and gives the asymptotic distribution of the test statistic, which is a chi-bar-squared distribution. A real data set concerning an ordered stratified table will be used to show the validity of this test method.  相似文献   

14.
In this second part of this paper, reproducibility of discrete ordinal and nominal outcomes is addressed. The first part deals with continuous outcomes, concentrating on intraclass correlation (ρ) in the context of one‐way analysis of variance. For categorical data, the focus has generally not been on a meaningful population parameter such as ρ. However, intraclass correlation has been defined for discrete ordinal data, ρc, and for nominal data, κI. Therefore, a unified approach to reproducibility is proposed. The relevance of these parameters is outlined. Estimation and inferential procedures for ρc and κI are reviewed, together with worked examples. Topics related to reproducibility that are not addressed in either this or the previous paper are highlighted. Considerations for designing reproducibility studies and for interpreting their results are provided. Copyright © 2004 John Wiley & Sons, Ltd.  相似文献   

15.
For many diseases, logistic constraints render large incidence studies difficult to carry out. This becomes a drawback, particularly when a new study is needed each time the incidence rate is investigated in a new population. By carrying out a prevalent cohort study with follow‐up it is possible to estimate the incidence rate if it is constant. The authors derive the maximum likelihood estimator (MLE) of the overall incidence rate, λ, as well as age‐specific incidence rates, by exploiting the epidemiologic relationship, (prevalence odds) = (incidence rate) × (mean duration) (P/[1 ? P] = λ × µ). The authors establish the asymptotic distributions of the MLEs and provide approximate confidence intervals for the parameters. Moreover, the MLE of λ is asymptotically most efficient and is the natural estimator obtained by substituting the marginal maximum likelihood estimators for P and µ into P/[1 ? P] = λ × µ. Following‐up the subjects allows the authors to develop these widely applicable procedures. The authors apply their methods to data collected as part of the Canadian Study of Health and Ageing to estimate the incidence rate of dementia amongst elderly Canadians. The Canadian Journal of Statistics © 2009 Statistical Society of Canada  相似文献   

16.
The odds ratio (OR) is a measure of association used for analysing an I × J contingency table. The total number of ORs to check grows with I and J. Several statistical methods have been developed for summarising them. These methods begin from two different starting points, the I × J contingency table and the two‐way table composed by the ORs. In this paper we focus our attention on the relationship between these methods and point out that, for an exhaustive analysis of association through log ORs, it is necessary to consider all the outcomes of these methods. We also introduce some new methodological and graphical features. In order to illustrate previously used methodologies, we consider a data table of the cross‐classification of the colour of eyes and hair of 5387 children from Scotland. We point out how, through the log OR analysis, it is possible to extract useful information about the association between variables.  相似文献   

17.
《随机性模型》2013,29(1):55-69
Abstract

This paper presents an improved method to calculate the delay distribution of a type k customer in a first-come-first-serve (FCFS) discrete-time queueing system with multiple types of customers, where each type has different service requirements, and c servers, with c = 1, 2 (the MMAP[K]/PH[K]/c queue). The first algorithms to compute this delay distribution, using the GI/M/1 paradigm, were presented by Van Houdt and Blondia [Van Houdt, B.; Blondia, C. The delay distribution of a type k customer in a first come first served MMAP[K]/PH[K]/1 queue. J. Appl. Probab. 2002, 39 (1), 213–222; The waiting time distribution of a type k customer in a FCFS MMAP[K]/PH[K]/2 queue. Technical Report; 2002]. The two most limiting properties of these algorithms are: (i) the computation of the rate matrix R related to the GI/M/1 type Markov chain, (ii) the amount of memory needed to store the transition matrices A l and B l . In this paper we demonstrate that each of the three GI/M/1 type Markov chains used to develop the algorithms in the above articles can be reduced to a QBD with a block size which is only marginally larger than that of its corresponding GI/M/1 type Markov chain. As a result, the two major limiting factors of each of these algorithms are drastically reduced to computing the G matrix of the QBD and storing the 6 matrices that characterize the QBD. Moreover, these algorithms are easier to implement, especially for the system with c = 2 servers. We also include some numerical examples that further demonstrate the reduction in computational resources.  相似文献   

18.
In a 1965 Decision Theory course at Stanford University, Charles Stein began a digression with “an amusing problem”: is there a proper confidence interval for the mean based on a single observation from a normal distribution with both mean and variance unknown? Stein introduced the interval with endpoints ± c|X| and showed indeed that for c large enough, the minimum coverage probability (over all values for the mean and variance) could be made arbitrarily near one. While the problem and coverage calculation were in the author’s hand-written notes from the course, there was no development of any optimality result for the interval. Here, the Hunt–Stein construction plus analysis based on special features of the problem provides a “minimax” rule in the sense that it minimizes the maximum expected length among all procedures with fixed coverage (or, equivalently, maximizes the minimal coverage among all procedures with a fixed expected length). The minimax rule is a mixture of two confidence procedures that are equivariant under scale and sign changes, and are uniformly better than the classroom example or the natural interval X ± c|X|?.  相似文献   

19.
Let Xl,…,Xn (Yl,…,Ym) be a random sample from an absolutely continuous distribution with distribution function F(G).A class of distribution-free tests based on U-statistics is proposed for testing the equality of F and G against the alternative that X's are more dispersed then Y's. Let 2 ? C ? n and 2 ? d ? m be two fixed integers. Let ?c,d(Xil,…,Xic ; Yjl,…,Xjd)=1(-1)when max as well as min of {Xil,…,Xic ; Yjl,…,Yjd } are some Xi's (Yj's)and zero oterwise. Let Sc,d be the U-statistic corresponding to ?c,d.In case of equal sample sizes, S22 is equivalent to Mood's Statistic.Large values of Sc,d are significant and these tests are quite efficient  相似文献   

20.
Let S (p×p) have a Wishart distribution -with v degrees of freedom and non-centrality matrix θ= [θjK] (p×p). Define θ0= min {| θjk |}, let θ0→∞, and suppose that | θjK | = 0(θo). Then the limiting form of the standardized non-central distribution, as θ while n? remains fixed, is a multivariate Gaussian distribution. This result in turn is used to obtain known asymptotic properties of multivariate chi-square and Rayleigh distributions under somewhat weaker conditions.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号