首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
This article provides some views on the statistical design and analysis of weather modification experiments. Perspectives were developed from experience with analyses of the Santa Barbara Phase I experiment summarized in Section 2, Randomization analvses are reported and compared with previously published parametric analyses. The parametric significance levels of tests for a cloud seeding effect agree well with the significance levels of the new corresponding randomization tests, These results, along with similar results of others, suggest that parametric analyses may be used as approximations to randomization analyses in exploratory analyses or reanalyses of weather modification experimental data.  相似文献   

2.
Missing data pose a serious challenge to the integrity of randomized clinical trials, especially of treatments for prolonged illnesses such as schizophrenia, in which long‐term impact assessment is of great importance, but the follow‐up rates are often no more than 50%. Sensitivity analysis using Bayesian modeling for missing data offers a systematic approach to assessing the sensitivity of the inferences made on the basis of observed data. This paper uses data from an 18‐month study of veterans with schizophrenia to demonstrate this approach. Data were obtained from a randomized clinical trial involving 369 patients diagnosed with schizophrenia that compared long‐acting injectable risperidone with a psychiatrist's choice of oral treatment. Bayesian analysis utilizing a pattern‐mixture modeling approach was used to validate the reported results by detecting bias due to non‐random patterns of missing data. The analysis was applied to several outcomes including standard measures of schizophrenia symptoms, quality of life, alcohol use, and global mental status. The original study results for several measures were confirmed against a wide range of patterns of non‐random missingness. Robustness of the conclusions was assessed using sensitivity parameters. The missing data in the trial did not likely threaten the validity of previously reported results. Copyright © 2014 John Wiley & Sons, Ltd.  相似文献   

3.
One way to analyze the AB-BA crossover trial with multivariate response is proposed. The multivariate model is given and the assumptions discussed. Two possibilities for the treatment eff ects hypothesis are considered. The statistical tests include the use of Hotelling's T 2 statistic, and a transformation equivalent to that of Jones and Kenward for the univariate case. Data from a nutrition experiment in Mexico illustrate the method. The multiple comparisons are carried out using Bonferroni intervals and the validity of the assumptions is explored. The main conclusions include the finding that some of the assumptions are not a requirement for the multivariate analysis; however, the sample sizes are important.  相似文献   

4.
We describe inferactive data analysis, so-named to denote an interactive approach to data analysis with an emphasis on inference after data analysis. Our approach is a compromise between Tukey's exploratory and confirmatory data analysis allowing also for Bayesian data analysis. We see this as a useful step in concrete providing tools (with statistical guarantees) for current data scientists. The basis of inference we use is (a conditional approach to) selective inference, in particular its randomized form. The relevant reference distributions are constructed from what we call a DAG-DAG—a Data Analysis Generative DAG, and a selective change of variables formula is crucial to any practical implementation of inferactive data analysis via sampling these distributions. We discuss a canonical example of an incomplete cross-validation test statistic to discriminate between black box models, and a real HIV dataset example to illustrate inference after making multiple queries on data.  相似文献   

5.
This article introduces BestClass, a set of SAS macros, available in the mainframe and workstation environment, designed for solving two-group classification problems using a class of recently developed nonparametric classification methods. The criteria used to estimate the classification function are based on either minimizing a function of the absolute deviations from the surface which separates the groups, or directly minimizing a function of the number of misclassified entities in the training sample. The solution techniques used by BestClass to estimate the classification rule use the mathematical programming routines of the SAS/OR software. Recently, a number of research studies have reported that under certain data conditions this class of classification methods can provide more accurate classification results than existing methods, such as Fisher's linear discriminant function and logistic regression. However, these robust classification methods have not yet been implemented in the major statistical packages, and hence are beyond the reach of those statistical analysts who are unfamiliar with mathematical programming techniques. We use a limited simulation experiment and an example to compare and contrast properties of the methods included in Best-Class with existing parametric and nonparametric methods. We believe that BestClass contributes significantly to the field of nonparametric classification analysis, in that it provides the statistical community with convenient access to this recently developed class of methods. BestClass is available from the authors.  相似文献   

6.
The Fels growth data record at half-yearly intervals the heights of children from birth to adulthood, and are the basis for pediatricians' growth charts used throughout North America. Aspects of human growth are the subject of a large medical and statistical literature. This paper uses smoothing splines to study the variation in height acceleration. By use of a functional version of principal-components analysis, we find that variation in the acceleration curve is essentially three-dimensional in nature. Evidence for a small growth spurt between the ages of six and eight, reported for data collected in Switzerland, is examined, and little support is found for the existence of this phenomenon in the Fels data.  相似文献   

7.
8.
This article describes a statistical computing system, the Computer-Assisted Data Analysis (CADA) Monitor, for use in performing interactive statistical data analysis. Especially easy to use because of its conversational nature, CADA includes facilities for data management, evaluation of probability distributions, Bayesian parametric models, Bayesian simultaneous estimation, Bayesian full-rank analysis of variance, and exploratory data analysis. CADA is written in a transportable subset of BASIC, and versions are currently available for a variety of computers.  相似文献   

9.
Datasets are sometimes divided into distinct subsets, e.g. due to multi-center sampling, or to variations in instruments, questionnaire item ordering or mode of administration, and the data analyst then needs to assess whether a joint analysis is meaningful. The Principal Component Analysis-based Data Structure Comparisons (PCADSC) tools are three new non-parametric, visual diagnostic tools for investigating differences in structure for two subsets of a dataset through covariance matrix comparisons by use of principal component analysis. The PCADCS tools are demonstrated in a data example using European Social Survey data on psychological well-being in three countries, Denmark, Sweden, and Bulgaria. The data structures are found to be different in Denmark and Bulgaria, and thus a comparison of for example mean psychological well-being scores is not meaningful. However, when comparing Denmark and Sweden, very similar data structures, and thus comparable concepts of well-being, are found. Therefore, inter-country comparisons are warranted for these countries.  相似文献   

10.
属性数据的多变量对应分析模型及应用   总被引:1,自引:0,他引:1  
属性数据的建模和分析是市场研究活动中最重要的一类应用问题,学术界和市场研究行业一般使用列联表及统计检验的方法进行分析。故将法国统计学家J.P.Beozecri提出的对应分析法导入市场研究实证过程,使问卷调查中常见的多变量属性数据分析建模过程呈现出稳健的统计结论和直观的二维图示,并推广到广义多变量对应分析场合,其结论可靠,有应用创新。  相似文献   

11.
It is important to educational planners to estimate the likelihood and time-scale of graduation of students enrolled on a curriculum. The particular case we are concerned with, emerges when studies are not completed in the prescribed interval of time. Under these circumstances we use a framework of survival analysis applied to lifetime-type educational data to examine the distribution of duration of undergraduate studies for 10,313 students, enrolled in a Greek university during ten consecutive academic years. Non-parametric and parametric survival models have been developed for handling this distribution as well as a modified procedure for testing goodness-of-fit of the models. Data censoring was taken into account in the statistical analysis and the problems of thresholding of graduation and of perpetual students are also addressed. We found that the proposed parametric model adequately describes the empirical distribution provided by non-parametric estimation. We also found significant difference between duration of studies of men and women students. The proposed methodology could be useful to analyse data from any other type and level of education or general lifetime data with similar characteristics.  相似文献   

12.
Missing data in clinical trials is a well‐known problem, and the classical statistical methods used can be overly simple. This case study shows how well‐established missing data theory can be applied to efficacy data collected in a long‐term open‐label trial with a discontinuation rate of almost 50%. Satisfaction with treatment in chronically constipated patients was the efficacy measure assessed at baseline and every 3 months postbaseline. The improvement in treatment satisfaction from baseline was originally analyzed with a paired t‐test ignoring missing data and discarding the correlation structure of the longitudinal data. As the original analysis started from missing completely at random assumptions regarding the missing data process, the satisfaction data were re‐examined, and several missing at random (MAR) and missing not at random (MNAR) techniques resulted in adjusted estimate for the improvement in satisfaction over 12 months. Throughout the different sensitivity analyses, the effect sizes remained significant and clinically relevant. Thus, even for an open‐label trial design, sensitivity analysis, with different assumptions for the nature of dropouts (MAR or MNAR) and with different classes of models (selection, pattern‐mixture, or multiple imputation models), has been found useful and provides evidence towards the robustness of the original analyses; additional sensitivity analyses could be undertaken to further qualify robustness. Copyright © 2012 John Wiley & Sons, Ltd.  相似文献   

13.
Tests of sharp null hypotheses, although frequently computed, are.rarely appropriate as the major end product of statistical analyses, except possibly in some, areas of the natural sciences. But procedures closely akin to tests, although often less formal, are needed in almost every investigation in which exploratory data analysis is used to help to decide upon the statistical model appropriate for the final analysis. The term “diagnostic check”has been suggested by Box and Jenkins for these procedures. Traditional statistical tests often suggest useful diagnostic checks -and this, in my view, is what tests are mainly good for-but visual examination and interpretation of data plots are often equally important. Biere is also much to be gained by the development of new diagnostic checks, and testing theory may be useful as one guide to this development.  相似文献   

14.
A class of computing devices known as desktop computers has emerged over the last several years. The International Data Corporation (McGovern 1980) estimates that the number of desktop computers is increasing by approximately 53,000 each month. Because of the projected widespread use of desktop computers and anticipated improvements in hardware, the potential for impressive statistical computing on these devices is exciting. Two features of desktop computers will be particularly important for those doing statistical analyses: (a) the ease-of-use of the computers, and (b) their extensive graphics capabilities. The author suggests that sophisticated statistical software will be available in the near future on many different models of desktop computers. Indeed, several of the manufacturers provide high-quality software at the present time. The implications for statisticians of a rapid growth rate for desktop computers are discussed for data analysis, software development, graphics, and instructional usage.  相似文献   

15.
In modern statistical practice, it is increasingly common to observe a set of curves or images, often measured with noise, and to use these as the basis of analysis (functional data analysis). We consider a functional data model consisting of measurement error and functional random effects motivated by data from a study of human vision. By transforming the data into the wavelet domain we are able to exploit the expected sparse representation of the underlying function and the mechanism generating the random effects. We propose simple fitting procedures and illustrate the methods on the vision data.  相似文献   

16.
This paper deals with the analysis of datasets, where the subjects are described by the estimated means of a p-dimensional variable. Classical statistical methods of data analysis do not treat measurements affected by intrinsic variability, as in the case of estimates, so that the heterogeneity induced among subjects by this condition is not taken into account. In this paper a way to solve the problem is suggested in the context of symbolic data analysis, whose specific aim is to handle data tables where single valued measurements are substituted by complex data structures like frequency distributions, intervals, and sets of values. A principal component analysis is carried out according to this proposal, with a significant improvement in the treatment of information.  相似文献   

17.
The need to use rigorous, transparent, clearly interpretable, and scientifically justified methodology for preventing and dealing with missing data in clinical trials has been a focus of much attention from regulators, practitioners, and academicians over the past years. New guidelines and recommendations emphasize the importance of minimizing the amount of missing data and carefully selecting primary analysis methods on the basis of assumptions regarding the missingness mechanism suitable for the study at hand, as well as the need to stress‐test the results of the primary analysis under different sets of assumptions through a range of sensitivity analyses. Some methods that could be effectively used for dealing with missing data have not yet gained widespread usage, partly because of their underlying complexity and partly because of lack of relatively easy approaches to their implementation. In this paper, we explore several strategies for missing data on the basis of pattern mixture models that embody clear and realistic clinical assumptions. Pattern mixture models provide a statistically reasonable yet transparent framework for translating clinical assumptions into statistical analyses. Implementation details for some specific strategies are provided in an Appendix (available online as Supporting Information), whereas the general principles of the approach discussed in this paper can be used to implement various other analyses with different sets of assumptions regarding missing data. Copyright © 2013 John Wiley & Sons, Ltd.  相似文献   

18.
面板数据的有序聚类分析是多元统计分析的新兴研究领域。借鉴多元统计学中主成分分析方法对面板数据在时间变量上进行降维处理,把变异信息的损失降低到最小,较为准确地反映了样本在各时间段内的整体变化水平;采用费希尔最优求解算法对主成分得分进行有序聚类,为研究有序面板数据的亲疏关系提供一些思路;对全球气候变化进行聚类分析,分析五十年来全球及区域气候变化特点,与国外研究结论对比,显示出良好的应用性。  相似文献   

19.
Many clinical research studies evaluate a time‐to‐event outcome, illustrate survival functions, and conventionally report estimated hazard ratios to express the magnitude of the treatment effect when comparing between groups. However, it may not be straightforward to interpret the hazard ratio clinically and statistically when the proportional hazards assumption is invalid. In some recent papers published in clinical journals, the use of restricted mean survival time (RMST) or τ ‐year mean survival time is discussed as one of the alternative summary measures for the time‐to‐event outcome. The RMST is defined as the expected value of time to event limited to a specific time point corresponding to the area under the survival curve up to the specific time point. This article summarizes the necessary information to conduct statistical analysis using the RMST, including the definition and statistical properties of the RMST, adjusted analysis methods, sample size calculation, information fraction for the RMST difference, and clinical and statistical meaning and interpretation. Additionally, we discuss how to set the specific time point to define the RMST from two main points of view. We also provide developed SAS codes to determine the sample size required to detect an expected RMST difference with appropriate power and reconstruct individual survival data to estimate an RMST reference value from a reported survival curve.  相似文献   

20.
Measuring the efficiency of public services: the limits of analysis   总被引:2,自引:0,他引:2  
Summary.  Policy makers are increasingly seeking to develop overall measures of the effi-ciency of public service organizations. For that, the use of 'off-the-shelf' statistical tools such as data envelopment analysis and stochastic frontier analysis have been advocated as tools to measure organizational efficiency. The analytical sophistication of such methods has reached an advanced stage of development. We discuss the context within which such models are deployed, their underlying assumptions and their usefulness for a regulator of public services. Four specific model building issues are discussed: the weights that are attached to public service outputs; the specification of the statistical model; the treatment of environmental influences on performance; the treatment of dynamic effects. The paper concludes with recommendations for policy makers and researchers on the development and use of efficiency measurement techniques.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号