首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 218 毫秒
1.
In this paper a test for outliers based on externally studentized residuals is shown to be related to a test for predictive failure. The relationships between a test for outliers, a test for a correlated mean shift and a test for an intercept shift are developed. A sequential testing procedure for outliers and structural change is shown to be independent, so that the overall size of the joint test can be determined exactly. It is established that a joint test for outliers and constancy of variances cannot be performed.  相似文献   

2.
A data base that provides a multivariate statistical history for each of a number of individual entities is called a pooled cross-sectional and time series data base in the econometrics literature. In marketing and survey literature the terms panel data or longitudinal data are often used. In management science a convenient term might be management data base. Such a data base provides a particularly rich environment for statistical analysis. This article reviews methods for estimating multivariate relationships particular to each individual entity and for summarizing these relationships for a number of individuals. Inference to a larger population when the data base is viewed as a sample is also considered.  相似文献   

3.
We present a simple framework for studying empirical-distribution-function goodness-of-fit tests for discrete models. A key tool is a weak-convergence result for an estimated discrete empirical process, regarded as a random element in some suitable sequence space. Special emphasis is given to the problem of testing for a Poisson model and for the geometric distribution. Simulations show that parametric bootstrap versions of the tests maintain a nominal level of significance very closely even for small samples where reliance upon asymptotic critical values is doubtful.  相似文献   

4.
This paper describes how a multistage analysis strategy for a clinical trial can assess a sequence of hypotheses that pertain to successively more stringent criteria for excess risk exclusion or superiority for a primary endpoint with a low event rate. The criteria for assessment can correspond to excess risk of an adverse event or to a guideline for sufficient efficacy as in the case of vaccine trials. The proposed strategy is implemented through a set of interim analyses, and success for one or more of the less stringent criteria at an interim analysis can be the basis for a regulatory submission, whereas the clinical trial continues to accumulate information to address the more stringent, but not futile, criteria. Simulations show that the proposed strategy is satisfactory for control of type I error, sufficient power, and potential success at interim analyses when the true relative risk is more favorable than assumed for the planned sample size. Copyright © 2013 John Wiley & Sons, Ltd.  相似文献   

5.
In this article, we consider several statistical models for censored exponential data. We prove a large deviation result for the maximum likelihood estimators (MLEs) of each model, and a unique result for the posterior distributions which works well for all the cases. Finally, comparing the large deviation rate functions for MLEs and posterior distributions, we show that a typical feature fails for one model; moreover, we illustrate the relation between this fact and a well-known result for curved exponential models.  相似文献   

6.
In this work, we propose a method for estimating the Hurst index, or memory parameter, of a stationary process with long memory in a Bayesian fashion. Such approach provides an approximation for the posterior distribution for the memory parameter and it is based on a simple application of the so-called approximate Bayesian computation (ABC), also known as likelihood-free method. Some popular existing estimators are reviewed and compared to this method for the fractional Brownian motion, for a long-range binary process and for the Rosenblatt process. The performance of our proposal is remarkably efficient.  相似文献   

7.
An asymptotic series for sums of powers of binomial coefficients is derived, the general term being defined and usable with a computer symbolic language. Sums of squares of coefficients in the symmetric case are shown to have a link with classical moment problems, but this property breaks down for cubes and higher powers. Problems of remainders for the asymptotic series are mentioned. Using the reflection formula for I'(.), a continuous form for a binomial function is set up, and this becomes oscillatory outstde the usual range. A new contmued fraction emerges for the logarithm of an adjusted sum of binomial squares. The note is a contribution to the problem of the interpretation of asymptotic series and processes for their convergence acceleration.  相似文献   

8.
Some comments are made concerning the possible forms of a correlation coefficient type goodness-of-fit statistic, and their relationship with other goodness-of-fit statistics, Critical values for a correlation goodness-of-fit statistic and for the Cramer-von Mises statistic are provided for testing a completely-specified null hypothesis for both complete and censored sampling, Critical values for a correlation test statistic are provided for complete and censored sampling for testing the hypothesis of normality, two parameter exponentiality, Weibull (or, extreme value) and an exponential-power distribution, respectively. Critical values are also provided for a test of one-parameter exponentiality based on the Cramer-von Mises statistic  相似文献   

9.
Confidence intervals are developed for the mode of a discrete unimodal distribution in the case where only a single observation is available. These intervals are centered on either the observation, X, or a weighted average of X with a constant, b, chosen by the investigator. Intervals are derived for nonrestricted unimodal distributions, for unimodal distributions with a symmetry property, and for a family of two-sided truncated geometric distributions.  相似文献   

10.
We have developed a new approach to determine the threshold of a biomarker that maximizes the classification accuracy of a disease. We consider a Bayesian estimation procedure for this purpose and illustrate the method using a real data set. In particular, we determine the threshold for Apolipoprotein B (ApoB), Apolipoprotein A1 (ApoA1) and the ratio for the classification of myocardial infarction (MI). We first conduct a literature review and construct prior distributions. We then develop classification rules based on the posterior distribution of the location and scale parameters for these biomarkers. We identify the threshold for ApoB and ApoA1, and the ratio as 0.908 (gram/liter), 1.138 (gram/liter) and 0.808, respectively. We also observe that the threshold for disease classification varies substantially across different age and ethnic groups. Next, we identify the most informative predictor for MI among the three biomarkers. Based on this analysis, ApoA1 appeared to be a stronger predictor than ApoB for MI classification. Given that we have used this data set for illustration only, the results will require further investigation for use in clinical applications. However, the approach developed in this article can be used to determine the threshold of any continuous biomarker for a binary disease classification.  相似文献   

11.
Several methods exist for testing interaction in unreplicated two-way layouts. Some are based on specifying a functional form for the interaction term and perform well provided that the functional form is appropriate. Other methods do not require such a functional form to be specified but only test for the presence of non-additivity and do not provide a suitable estimate of error variance for a non-additive model. This paper presents a method for testing for interaction in unreplicated two-way tables that is based on testing all pairwise interaction contrasts. This method (i) is easy to implement, (ii) does not assume a functional form for the interaction term, (iii) can find a sub-table of data which may be free from interaction and to base the estimate of unknown error variance, and (iv) can be used for incomplete two-way layouts. The proposed method is illustrated using examples and its power is investigated via simulation studies. Simulation results show that the proposed method is competitive with existing methods for testing for interaction in unreplicated two-way layouts.  相似文献   

12.
The QTc interval of the electrocardiogram is a pharmacodynamic biomarker for drug-induced cardiac toxicity. The ICH E14 guideline Questions and Answers offer a solution for evaluating a concentration-QTc relationship in early clinical studies as an alternative to conducting a thorough QT/QTc study. We focused on covariance structures of QTc intervals on the baseline day and dosing day (two-day covariance structure,) and proposed a two-day QTc model to analyze a concentration-QTc relationship for placebo-controlled parallel phase 1 single ascending dose studies. The proposed two-day QTc model is based on a constrained longitudinal data analysis model and a mixed effects model, thus allowing various variance components to capture the two-day covariance structure. We also propose a one-day QTc model for the situation where no baseline day or only a pre-dose baseline is available and models for multiple ascending dose studies where concentration and QTc intervals are available over multiple days. A simulation study shows that the proposed models control the false negative rate for positive drugs and have both higher accuracy and power for negative drugs than existing models in a variety of settings for the two-day covariance structure. The proposed models will promote early and accurate evaluation of the cardiac safety of new drugs.  相似文献   

13.
This article considers the different methods for determining sample sizes for Wald, likelihood ratio, and score tests for logistic regression. We review some recent methods, report the results of a simulation study comparing each of the methods for each of the three types of test, and provide Mathematica code for calculating sample size. We consider a variety of covariate distributions, and find that a calculation method based on a first order expansion of the likelihood ratio test statistic performs consistently well in achieving a target level of power for each of the three types of test.  相似文献   

14.
The authors propose methods for Bayesian inference for generalized linear models with missing covariate data. They specify a parametric distribution for the covariates that is written as a sequence of one‐dimensional conditional distributions. They propose an informative class of joint prior distributions for the regression coefficients and the parameters arising from the covariate distributions. They examine the properties of the proposed prior and resulting posterior distributions. They also present a Bayesian criterion for comparing various models, and a calibration is derived for it. A detailed simulation is conducted and two real data sets are examined to demonstrate the methodology.  相似文献   

15.
We consider the problem of setting up a confidence region for the mean of amultivariate timeseries ont he basis of a part-realisation of that series.A procedure for setting up a confidence interval for the mean of a univariate time series Is implicitin Jones(1976).We present an analogous procedure for setting up a confidence region for the mean of a multivariatet ime series.This procedure is base donastatistic which is an analogue of Hotelling'sT'.Our results are applied to a comparison of climate means obtained from experiments with a General Circulation Model of the earth's atmosphere.  相似文献   

16.
This article proposes a method for estimating principal points for a multivariate binary distribution, assuming a log-linear model for the distribution. Through numerical simulation studies, the proposed parametric estimation method using a log-linear model is compared with a nonparametric estimation method.  相似文献   

17.
Some multiple comparison procedures are described for multiple armed studies. The procedures are appropriate for testing all hypotheses for comparing two endpoints and multiple test arms to a single control group, for example three different fixed doses compared to a placebo. The procedure assumes that among the two endpoints, one is designated as a primary endpoint such that for a given treatment arm, no hypothesis for the secondary endpoint can be rejected unless the hypothesis for the primary endpoint was rejected. The procedures described control the family-wise error rate in the strong sense at a specified level α.  相似文献   

18.
This article describes a method for partitioning with respect to a control for the situation in which the treatment sample sizes are unequal and also for the situation where the treatment sample sizes are equal except for a few missing values. Calculation of the critical values required for finding confidence limits is discussed and tables are presented for the “almost equal” sample size case. An application of this method to length of stay data for congestive heart failure patients is also provided.  相似文献   

19.
Bayesian semiparametric inference is considered for a loglinear model. This model consists of a parametric component for the regression coefficients and a nonparametric component for the unknown error distribution. Bayesian analysis is studied for the case of a parametric prior on the regression coefficients and a mixture-of-Dirichlet-processes prior on the unknown error distribution. A Markov-chain Monte Carlo (MCMC) method is developed to compute the features of the posterior distribution. A model selection method for obtaining a more parsimonious set of predictors is studied. The method adds indicator variables to the regression equation. The set of indicator variables represents all the possible subsets to be considered. A MCMC method is developed to search stochastically for the best subset. These procedures are applied to two examples, one with censored data.  相似文献   

20.
In 1991 Marsh and co-workers made the case for a sample of anonymized records (SAR) from the 1991 census of population. The case was accepted by the Office for National Statistics (then the Office of Population Censuses and Surveys) and a request was made by the Economic and Social Research Council to purchase the SARs. Two files were released for Great Britain—a 2% sample of individuals and a 1% sample of households. Subsequently similar samples were released for Northern Ireland. Since their release, the files have been heavily used for research and there has been no known breach of confidentiality. There is a considerable demand for similar files from the 2001 census, with specific requests for a larger sample size and lower population threshold for the individual SAR. This paper reassesses the analysis of Marsh and co-workers of the risk of identification of an individual or household in a sample of microdata from the 1991 census and also uses alternative ways of assessing risks with the 1991 SARs. The results of both the reassessment and the new analyses are reassuring and allow us to take the 1991 SARs as a base-line against which to assess proposals for changes to the size and structure of samples from the 2001 census.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号