首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
A test for empty sets is described. This new test is suitable for testing for mutual exclusivity, sets being nested, and many other configurations. It might be hypothesized that two medical conditions do not both occur in individuals. This can be expressed as an empty intersection of the two sets of people with each condition. An important feature of the test is the incorporation of misclassification rates into the analysis. The test utilizes the potential misclassifications for false designations in the sets that are hypothesized to be empty. The test is quite powerful. When the null hypothesis is rejected, follow-up tests on one or more sets that are included in the null hypothesis can be performed using the same new test. MatLab code is supplied.  相似文献   

2.
Statistical hypotheses and test statistics are Boolean functions that can be manipulated using the tools of Boolean algebra. These tools are particularly useful for exploring multiple comparisons or simultaneous inference theory, in which multiparameter hypotheses or multiparameter test statistics may be decomposed into combinations of uniparameter hypotheses or uniparameter tests. These concepts are illustrated with both finite and infinite decompositions of familiar multiparameter hypotheses and tests. The corresponding decompositions of acceptance regions and rejection regions are also shown. Finally, the close relationship between hypothesis and test decompositions and Roy's union—intersection principle is demonstrated by a derivation of the union—intersection test of the univariate general linear hypothesis.  相似文献   

3.
In this paper, the application of the intersection–union test method in fixed‐dose combination drug studies is discussed. An approximate sample size formula for the problem of testing the efficacy of a combination drug using intersection–union tests is proposed. The sample sizes obtained from the formula are found to be reasonably accurate in terms of attaining the target power 1?β for a specified β. Copyright © 2003 John Wiley & Sons, Ltd.  相似文献   

4.
A semiparametric logistic regression model is proposed in which its nonparametric component is approximated with fixed-knot cubic B-splines. To assess the linearity of the nonparametric component, we construct a penalized likelihood ratio test statistic. When the number of knots is fixed, the null distribution of the test statistic is shown to be asymptotically the distribution of a linear combination of independent chi-squared random variables, each with one degree of freedom. We set the asymptotic null expectation of this test statistic equal to a value to determine the smoothing parameter value. Monte Carlo experiments are conducted to investigate the performance of the proposed test. Its practical use is illustrated with a real-life example.  相似文献   

5.
A sequential method for approximating a general permutation test (SAPT) is proposed and evaluated. Permutations are randomly generated from some set G, and a sequential probability ratio test (SPRT) is used to determine whether an observed test statistic falls sufficiently far in the tail of the permutation distribution to warrant rejecting some hypothesis. An estimate and bounds on the power function of the SPRT are used to find bounds on the effective significance level of the SAPT. Guidelines are developed for choosing parameters in order to obtain a desired significance level and minimize the number of permutations needed to reach a decision. A theoretical estimate of the average number of permutations under the null hypothesis is given along with simulation results demonstrating the power and average number of permutations for various alternatives. The sequential approximation retains the generality of the permutation test,- while avoiding the computational complexities that arise in attempting to computer the full permutation distribution exactly  相似文献   

6.
Gronnesby and Borgan (1996) propose an overall goodness-of-fit test for the Cox proportional hazards model. The basis of their test is a grouping of subjects by their estimated risk score. We show that the Gronnesby and Borgan test is algebraically identical to one obtained from adding group indicator variables to the model and testing the hypothesis the coefficients of the group indicator variables are zero via the score test. Thus showing that the test can be calculated using existing software. We demonstrate that the table of observed and estimated expected number of events within each group of the risk score is a useful adjunct to the test to help identify potential problems in fit.  相似文献   

7.
The nonparametric component in a partially linear model is estimated by a linear combination of fixed-knot cubic B-splines with a second-order difference penalty on the adjacent B-spline coefficients. The resulting penalized least-squares estimator is used to construct two Wald-type spline-based test statistics for the null hypothesis of the linearity of the nonparametric function. When the number of knots is fixed, the first test statistic asymptotically has the distribution of a linear combination of independent chi-squared random variables, each with one degree of freedom, under the null hypothesis. The smoothing parameter is determined by specifying a value for the asymptotically expected value of the test statistic under the null hypothesis. When the number of knots is fixed and under the null hypothesis, the second test statistic asymptotically has a chi-squared distribution with K=q+2 degrees of freedom, where q is the number of knots used for estimation. The power performances of the two proposed tests are investigated via simulation experiments, and the practicality of the proposed methodology is illustrated using a real-life data set.  相似文献   

8.
Using a minimum p-value principle, a new two-sample test MIN3 is proposed in the paper. The cumulative distribution function of the MIN3 test statistic is studied and approximated by the Beta distribution of the third kind. Lower percentage points of the distribution of the new test statistic under the null hypothesis are computed. Also the test power for a lot of types of alternative hypotheses (with 0, 1 and 2 point(-s) of the intersection(-s) of survival functions) is studied and we found that the usage of the MIN3 test is a preferred strategy by the Wald and Savage decision-making criteria under risk and uncertainty. The results of application of the MIN3 test are shown for two examples from lifetime data analysis.  相似文献   

9.
The practice for testing homogeneity of several rival models is of interest. In this article, we consider a non parametric multiple test for non nested distributions in the context of the model selection. Based on the linear sign rank test, and the known union–intersection principle, we let the magnitude of the data to give a better performance to the test statistic. We consider the sample and the non nested rival models as blocks and treatments, respectively, and introduce the extended Friedman test version to compare with the results of the test based on the linear sign rank test. A real dataset based on the waiting time to earthquake is considered to illustrate the results.  相似文献   

10.
Grønnesby and Borgan (1996, Lifetime Data Analysis 2, 315–328) propose an omnibus goodness-of-fit test for the Cox proportional hazards model. The test is based on grouping the subjects by their estimated risk score and comparing the number of observed and a model based estimated number of expected events within each group. We show, using extensive simulations, that even for moderate sample sizes the choice of number of groups is critical for the test to attain the specified size. In light of these results we suggest a grouping strategy under which the test attains the correct size even for small samples. The power of the test statistic seems to be acceptable when compared to other goodness-of-fit tests.  相似文献   

11.
A generalization of step-up and step-down multiple test procedures is proposed. This step-up-down procedure is useful when the objective is to reject a specified minimum number, q, out of a family of k hypotheses. If this basic objective is met at the first step, then it proceeds in a step-down manner to see if more than q hypotheses can be rejected. Otherwise it proceeds in a step-up manner to see if some number less than q hypotheses can be rejected. The usual step-down procedure is the special case where q = 1, and the usual step-up procedure is the special case where q = k. Analytical and numerical comparisons between the powers of the step-up-down procedures with different choices of q are made to see how these powers depend on the actual number of false hypotheses. Examples of application include comparing the efficacy of a treatment to a control for multiple endpoints and testing the sensitivity of a clinical trial for comparing the efficacy of a new treatment with a set of standard treatments.  相似文献   

12.
In this article, two different types of precedence tests, each with two different test statistics, based on ranked set samples for testing the equality of two distributions are discussed. The exact null distributions of proposed test statistics are derived, critical values are tabulated for both set size and number of cycles up to 8, and the exact power functions of these two types of precedence tests under the Lehmann alternative are derived. Then, the power values of these two test procedures and their competitors based on simple random samples and based on ranked set samples are compared under the Lehmann alternative exactly and also under a location-shift alternative by means of Monte Carlo simulations. Finally, the impact of imperfect ranking is discussed and some concluding remarks are presented.  相似文献   

13.
An adaptive test is proposed for the one-way layout. This test procedure uses the order statistics of the combined data to obtain estimates of percentiles, which are used to select an appropriate set of rank scores for the one-way test statistic. This test is designed to have reasonably high power over a range of distributions. The adaptive procedure proposed for a one-way layout is a generalization of an existing two-sample adaptive test procedure. In this Monte Carlo study, the power and significance level of the F-test, the Kruskal-Wallis test, the normal scores test, and the adaptive test were evaluated for the one-way layout. All tests maintained their significance level for data sets having at least 24 observations. The simulation results show that the adaptive test is more powerful than the other tests for skewed distributions if the total number of observations equals or exceeds 24. For data sets having at least 60 observations the adaptive test is also more powerful than the F-test for some symmetric distributions.  相似文献   

14.
In some exceptional circumstances, as in very rare diseases, nonrandomized one‐arm trials are the sole source of evidence to demonstrate efficacy and safety of a new treatment. The design of such studies needs a sound methodological approach in order to provide reliable information, and the determination of the appropriate sample size still represents a critical step of this planning process. As, to our knowledge, no method exists for sample size calculation in one‐arm trials with a recurrent event endpoint, we propose here a closed sample size formula. It is derived assuming a mixed Poisson process, and it is based on the asymptotic distribution of the one‐sample robust nonparametric test recently developed for the analysis of recurrent events data. The validity of this formula in managing a situation with heterogeneity of event rates, both in time and between patients, and time‐varying treatment effect was demonstrated with exhaustive simulation studies. Moreover, although the method requires the specification of a process for events generation, it seems to be robust under erroneous definition of this process, provided that the number of events at the end of the study is similar to the one assumed in the planning phase. The motivating clinical context is represented by a nonrandomized one‐arm study on gene therapy in a very rare immunodeficiency in children (ADA‐SCID), where a major endpoint is the recurrence of severe infections. Copyright © 2012 John Wiley & Sons, Ltd.  相似文献   

15.
Tests that combine p-values, such as Fisher's product test, are popular to test the global null hypothesis H0 that each of n component null hypotheses, H1,…,Hn, is true versus the alternative that at least one of H1,…,Hn is false, since they are more powerful than classical multiple tests such as the Bonferroni test and the Simes tests. Recent modifications of Fisher's product test, popular in the analysis of large scale genetic studies include the truncated product method (TPM) of Zaykin et al. (2002), the rank truncated product (RTP) test of Dudbridge and Koeleman (2003) and more recently, a permutation based test—the adaptive rank truncated product (ARTP) method of Yu et al. (2009). The TPM and RTP methods require users' specification of a truncation point. The ARTP method improves the performance of the RTP method by optimizing selection of the truncation point over a set of pre-specified candidate points. In this paper we extend the ARTP by proposing to use all the possible truncation points {1,…,n} as the candidate truncation points. Furthermore, we derive the theoretical probability distribution of the test statistic under the global null hypothesis H0. Simulations are conducted to compare the performance of the proposed test with the Bonferroni test, the Simes test, the RTP test, and Fisher's product test. The simulation results show that the proposed test has higher power than the Bonferroni test and the Simes test, as well as the RTP method. It is also significantly more powerful than Fisher's product test when the number of truly false hypotheses is small relative to the total number of hypotheses, and has comparable power to Fisher's product test otherwise.  相似文献   

16.
In this paper we investigate the asymptotic critical value behaviour of certain multiple decision procedures as e.g. simultaneous confidence intervals and simultaneous as well as stepwise multiple test procedures. Supposing that n hypotheses or parameters of interest are under consideration we investigate the critical value behaviour when n increases. More specifically, we answer e.g. the question by which amount the lengths of confidence intervals increase when an additional parameter is added to the statistical analysis. Furthermore, critical values of different multiple decision procedures as for instance step-down and step-up procedures will be compared. Some general theoretic results are derived and applied for various distributions.  相似文献   

17.
This article introduces a kernel-based nonparametric inferential procedure to test for Granger causality in distribution. This test is a multivariate extension of the kernel-based Granger causality test in tail event. The main advantage of this test is its ability to examine a large number of lags, with higher-order lags discounted. In addition, our test is highly flexible because it can be used to identify Granger causality in specific regions on the distribution supports, such as the center or tails. We prove that the test converges asymptotically to a standard Gaussian distribution under the null hypothesis and thus is free of parameter estimation uncertainty. Monte Carlo simulations illustrate the excellent small sample size and power properties of the test. This new test is applied to a set of European stock markets to analyze spillovers during the recent European crisis and to distinguish contagion from interdependence effects.  相似文献   

18.
Graphs are presented on which the empirical distribution function can be plotted to test the assumption of normality by the Lilliefors test. A second set of graphs is presented for using the Lilliefors test on exponential distributions. The graphs allow for tests at the 10 percent, 5 percent, and 1 percent levels of significance. Use of these graphs makes it easy for students in a first course in statistics to test normal and exponential distributions without having to unravel the mystery associated with putting together a chi-squared goodness-of-fit test.  相似文献   

19.
The problem of testing whether one regression function is larger than another on a specified set R is considered The regression functions must be linear functions of the parameters but need not be linear functions of the independent variables The proposed test has an exactly specified size in typical situations The test's critical value is a standard t percentile. The power function of the test is investigated.  相似文献   

20.
A diagnostic key defines a hierarchial sequence of tests used to identify a specimen from a set of known taxa. The usual measure of the efficiency of a key, the expected number of tests per identification, may not be appropriate when the responses to tests are not known for all taxon/test pairs. An alternative measure is derived and it is shown that the test selected for use at each point in the sequence should depend on which measure is used. Two suggestions of Gower and Payne (1975), regarding test selection, are shown to be appropriate only to the new measure. Tests are usually selected by calculating the value of some selection criterion function. Functions are reviewed for use in each of the two situations, and new functions are derived. The functions discussed are shown to be interpretable in terms of the number of tests required to complete the key from the current point in the sequence, given that a particular test is selected. This interpretation enables the functions to be extended to select tests with different costs and allows recommendations to be made as to which function to use, depending on how many ‘good’ tests are available.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号