首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
As new diagnostic tests are developed and marketed, it is very important to be able to compare the accuracy of a given two continuous‐scale diagnostic tests. An effective method to evaluate the difference between the diagnostic accuracy of two tests is to compare partial areas under the receiver operating characteristic curves (AUCs). In this paper, we review existing parametric methods. Then, we propose a new semiparametric method and a new nonparametric method to investigate the difference between two partial AUCs. For the difference between two partial AUCs under each method, we derive a normal approximation, define an empirical log‐likelihood ratio, and show that the empirical log‐likelihood ratio follows a scaled chi‐square distribution. We construct five confidence intervals for the difference based on normal approximation, bootstrap, and empirical likelihood methods. Finally, extensive simulation studies are conducted to compare the finite‐sample performances of these intervals, and a real example is used as an application of our recommended intervals. The simulation results indicate that the proposed hybrid bootstrap and empirical likelihood intervals outperform other existing intervals in most cases.  相似文献   

2.
The accuracy of a binary diagnostic test is usually measured in terms of its sensitivity and its specificity. Other measures of the performance of a diagnostic test are the positive and negative likelihood ratios, which quantify the increase in knowledge about the presence of the disease through the application of a diagnostic test, and which depend on the sensitivity and specificity of the diagnostic test. In this article, we construct an asymptotic hypothesis test to simultaneously compare the positive and negative likelihood ratios of two or more diagnostic tests in unpaired designs. The hypothesis test is based on the logarithmic transformation of the likelihood ratios and on the chi-square distribution. Simulation experiments have been carried out to study the type I error and the power of the constructed hypothesis test when comparing two and three binary diagnostic tests. The method has been extended to the case of multiple multi-level diagnostic tests.  相似文献   

3.
The receiver operating characteristic (ROC) curve is one of the most commonly used methods to compare the diagnostic performance of two or more laboratory or diagnostic tests. In this paper, we propose semi-empirical likelihood based confidence intervals for ROC curves of two populations, where one population is parametric and the other one is non-parametric and both have missing data. After imputing missing values, we derive the semi-empirical likelihood ratio statistic and the corresponding likelihood equations. It is shown that the log-semi-empirical likelihood ratio statistic is asymptotically scaled chi-squared. The estimating equations are solved simultaneously to obtain the estimated lower and upper bounds of semi-empirical likelihood confidence intervals. We conduct extensive simulation studies to evaluate the finite sample performance of the proposed empirical likelihood confidence intervals with various sample sizes and different missing probabilities.  相似文献   

4.
We consider likelihood ratio, score and Wald tests for a three-way random effects ANOVA model. Competitor tests are compared using criteria such as small sample power, asymptotic relative efficiency, and convenient null distribution. The final choice is between a new test and two tests long used in practice.  相似文献   

5.
In this paper, we derive sequential conditional probability ratio tests to compare diagnostic tests without distributional assumptions on test results. The test statistics in our method are nonparametric weighted areas under the receiver-operating characteristic curves. By using the new method, the decision of stopping the diagnostic trial early is unlikely to be reversed should the trials continue to the planned end. The conservatism reflected in this approach to have more conservative stopping boundaries during the course of the trial is especially appealing for diagnostic trials since the end point is not death. In addition, the maximum sample size of our method is not greater than a fixed sample test with similar power functions. Simulation studies are performed to evaluate the properties of the proposed sequential procedure. We illustrate the method using data from a thoracic aorta imaging study.  相似文献   

6.
Abstract. We introduce fully non‐parametric two‐sample tests for testing the null hypothesis that the samples come from the same distribution if the values are only indirectly given via current status censoring. The tests are based on the likelihood ratio principle and allow the observation distributions to be different for the two samples, in contrast with earlier proposals for this situation. A bootstrap method is given for determining critical values and asymptotic theory is developed. A simulation study, using Weibull distributions, is presented to compare the power behaviour of the tests with the power of other non‐parametric tests in this situation.  相似文献   

7.
For the problem of testing the homogeneity of the variances in a covariance matrix with a block compound symmetric structure, the likelihood ratio test is derived in this paper, A modification of the test that allows its distribution to be better approximated by the chi-square distribution is also considered, Formulae for calculating approximate sample size and power are derived, Small sample performances of these tests in the case of two dependent bivariate or trivariate normals are compared to each other and to the competing tests by simulating levels of significance and powers, and recommendation is made of the ones that have good performance, The recommended tests are then demonstrated in an illustrative example.  相似文献   

8.
For a specified decision rule, a general class of likelihood ratio based repeated significance tests is considered. An invariance principle for the likelihood ratio statistics is established and incorporated in the study of the asymptotic theory of the proposed tests. For comparing these tests with the conventional likelihood ratio tests, based solely on the target sample size, some Bahadur efficiency results are presented. The theory is then adapted in the study of some multiple comparison procedures  相似文献   

9.
This article considers the different methods for determining sample sizes for Wald, likelihood ratio, and score tests for logistic regression. We review some recent methods, report the results of a simulation study comparing each of the methods for each of the three types of test, and provide Mathematica code for calculating sample size. We consider a variety of covariate distributions, and find that a calculation method based on a first order expansion of the likelihood ratio test statistic performs consistently well in achieving a target level of power for each of the three types of test.  相似文献   

10.
Compared to tests for localized clusters, the tests for global clustering only collect evidence for clustering throughout the study region without evaluating the statistical significance of the individual clusters. The weighted likelihood ratio (WLR) test based on the weighted sum of likelihood ratios represents an important class of tests for global clustering. Song and Kulldorff (Likelihood based tests for spatial randomness. Stat Med. 2006;25(5):825–839) developed a wide variety of weight functions with the WLR test for global clustering. However, these weight functions are often defined based on the cell population size or the geographic information such as area size and distance between cells. They do not make use of the information from the observed count, although the likelihood ratio of a potential cluster depends on both the observed count and its population size. In this paper, we develop a self-adjusted weight function to directly allocate weights onto the likelihood ratios according to their values. The power of the test was evaluated and compared with existing methods based on a benchmark data set. The comparison results favour the suggested test especially under global chain clustering models.  相似文献   

11.
We study a hypothesis testing problem involving the location model suggested by Olkin and Tate (1961). Specifically, we derive a likelihood ratio lest of the associated location hypothesis as an alternative to the conventional method of carrying out separate tests for each of the parameters. A small sample Monte Carlo comparison indicates the general superiority of the former in terms of statistical power. We also comment briefly on the properties of the test.  相似文献   

12.
We consider the issue of performing accurate small-sample testing inference in beta regression models, which are useful for modeling continuous variates that assume values in (0,1), such as rates and proportions. We derive the Bartlett correction to the likelihood ratio test statistic and also consider a bootstrap Bartlett correction. Using Monte Carlo simulations we compare the finite sample performances of the two corrected tests to that of the standard likelihood ratio test and also to its variant that employs Skovgaard's adjustment; the latter is already available in the literature. The numerical evidence favors the corrected tests we propose. We also present an empirical application.  相似文献   

13.
This paper is concerned with testing the equality of scale parameters of K(> 2) two-parameter exponential distributions in presence of unspecified location parameters based on complete and type II censored samples. We develop a marginal likelihood ratio statistic, a quadratic statistic (Qu) (Nelson, 1982) based on maximum marginal likelihood estimates of the scale parameters under the null and the alternative hypotheses, a C(a) statistic (CPL) (Neyman, 1959) based on the profile likelihood estimate of the scale parameter under the null hypothesis and an extremal scale parameter ratio statistic (ESP) (McCool, 1979). We show that the marginal likelihood ratio statistic is equivalent to the modified Bartlett test statistic. We use Bartlett's small sample correction to the marginal likelihood ratio statistic and call it the modified marginal likelihood ratio statistic (MLB). We then compare the four statistics, MLBi Qut CPL and ESP in terms of size and power by using Monte Carlo simulation experiments. For the variety of sample sizes and censoring combinations and nominal levels considered the statistic MLB holds nominal level most accurately and based on empirically calculated critical values, this statistic performs best or as good as others in most situations. Two examples are given.  相似文献   

14.
空间面板数据模型由于考虑了经济变量间的空间相关性,其优势日益凸显,已成为计量经济学的热点研究领域。将空间相关性与动态模式同时扩展到面板模型中的空间动态面板模型,不仅考虑了经济变量之间的空间相关性,还考虑了时间上的滞后性,是空间面板模型的发展,增强了模型的解释力。考虑一种带固定个体效应、因变量的时间滞后项、因变量与随机误差项均存在空间自相关性的空间动态面板回归模型,提出了在个体数n和时间数T都很大,且T相对地大于n的条件下空间动态面板模型中时间滞后效应存在性的LM和LR检验方法,其检验方法包括联合检验、一维及二维的边际和条件检验;推导出这些检验在零假设下的极限分布;其极限分布均服从卡方分布。通过模拟试验研究检验统计量的小样本性质,结果显示其具有优良的统计性质。  相似文献   

15.
In the presence of partial disease verification, the comparison of the accuracy of binary diagnostic tests cannot be carried out through the paired comparison of the diagnostic tests applying McNemar's test, since for a subsample of patients the disease status is unknown. In this study, we have deduced the maximum likelihood estimators for the sensitivities and specificities of multiple binary diagnostic tests and we have studied various joint hypothesis tests based on the chi-square distribution to compare simultaneously the accuracy of these binary diagnostic tests when for some patients in the sample the disease status is unknown. Simulation experiments were carried out to study the type I error and the power of each hypothesis test deduced. The results obtained were applied to the diagnosis of coronary stenosis.  相似文献   

16.
ABSTRACT

In clustered survival data, the dependence among individual survival times within a cluster has usually been described using copula models and frailty models. In this paper we propose a profile likelihood approach for semiparametric copula models with different cluster sizes. We also propose a likelihood ratio method based on profile likelihood for testing the absence of association parameter (i.e. test of independence) under the copula models, leading to the boundary problem of the parameter space. For this purpose, we show via simulation study that the proposed likelihood ratio method using an asymptotic chi-square mixture distribution performs well as sample size increases. We compare the behaviors of the two models using the profile likelihood approach under a semiparametric setting. The proposed method is demonstrated using two well-known data sets.  相似文献   

17.
The statistical methodology under order restriction is very mathematical and complex. Thus, we provide a brief methodological background of order-restricted likelihood ratio tests for the normal theoretical case for the basic understanding of its applications, and relegate more technical details to the appendices. For data analysis, algorithms for computing the order-restricted estimates and computation of p-values are described. A two-step procedure is presented for obtaining the sample size in clinical trials when the minimum power, say 0.80 or 0.90 is specified, and the normal means satisfy an order restriction. Using this approach will result in reduction of 14-24% in the sample size required when one-sided ordered alternatives are used, as illustrated by several examples.  相似文献   

18.
A disease prevalence can be estimated by classifying subjects according to whether they have the disease. When gold-standard tests are too expensive to be applied to all subjects, partially validated data can be obtained by double-sampling in which all individuals are classified by a fallible classifier, and some of individuals are validated by the gold-standard classifier. However, it could happen in practice that such infallible classifier does not available. In this article, we consider two models in which both classifiers are fallible and propose four asymptotic test procedures for comparing disease prevalence in two groups. Corresponding sample size formulae and validated ratio given the total sample sizes are also derived and evaluated. Simulation results show that (i) Score test performs well and the corresponding sample size formula is also accurate in terms of the empirical power and size in two models; (ii) the Wald test based on the variance estimator with parameters estimated under the null hypothesis outperforms the others even under small sample sizes in Model II, and the sample size estimated by this test is also accurate; (iii) the estimated validated ratios based on all tests are accurate. The malarial data are used to illustrate the proposed methodologies.  相似文献   

19.
The importance of the normal distribution for fitting continuous data is well known. However, in many practical situations data distribution departs from normality. For example, the sample skewness and the sample kurtosis are far away from 0 and 3, respectively, which are nice properties of normal distributions. So, it is important to have formal tests of normality against any alternative. D'Agostino et al. [A suggestion for using powerful and informative tests of normality, Am. Statist. 44 (1990), pp. 316–321] review four procedures Z 2(g 1), Z 2(g 2), D and K 2 for testing departure from normality. The first two of these procedures are tests of normality against departure due to skewness and kurtosis, respectively. The other two tests are omnibus tests. An alternative to the normal distribution is a class of skew-normal distributions (see [A. Azzalini, A class of distributions which includes the normal ones, Scand. J. Statist. 12 (1985), pp. 171–178]). In this paper, we obtain a score test (W) and a likelihood ratio test (LR) of goodness of fit of the normal regression model against the skew-normal family of regression models. It turns out that the score test is based on the sample skewness and is of very simple form. The performance of these six procedures, in terms of size and power, are compared using simulations. The level properties of the three statistics LR, W and Z 2(g 1) are similar and close to the nominal level for moderate to large sample sizes. Also, their power properties are similar for small departure from normality due to skewness (γ1≤0.4). Of these, the score test statistic has a very simple form and computationally much simpler than the other two statistics. The LR statistic, in general, has highest power, although it is computationally much complex as it requires estimates of the parameters under the normal model as well as those under the skew-normal model. So, the score test may be used to test for normality against small departure from normality due to skewness. Otherwise, the likelihood ratio statistic LR should be used as it detects general departure from normality (due to both skewness and kurtosis) with, in general, largest power.  相似文献   

20.
ABSTRACT

A frequently encountered statistical problem is to determine if the variability among k populations is heterogeneous. If the populations are measured using different scales, comparing variances may not be appropriate. In this case, comparing coefficient of variation (CV) can be used because CV is unitless. In this paper, a non-parametric test is introduced to test whether the CVs from k populations are different. With the assumption that the populations are independent normally distributed, the Miller test, Feltz and Miller test, saddlepoint-based test, log likelihood ratio test and the proposed simulated Bartlett-corrected log likelihood ratio test are derived. Simulation results show the extreme accuracy of the simulated Bartlett-corrected log likelihood ratio test if the model is correctly specified. If the model is mis-specified and the sample size is small, the proposed test still gives good results. However, with a mis-specified model and large sample size, the non-parametric test is recommended.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号