首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
We derive a test in order to examine the need of modeling survival data using frailty models based on the likelihood ratio (LR) test for homogeneity. Test is developed for both complete and censored samples from a family of baseline distributions that satisfy a closure property. Approach motivated by I-divergence distance is used in order to determine “credible” regions for all parameters of baseline distribution for which homogeneity hypothesis is not rejected. Proposed test outperforms the usual asymptotic LR test both in very small samples with known frailty and for all small sample sizes under misspecified frailty.  相似文献   

2.
Abstract

The Poisson distribution is here used to illustrate Bayesian inference concepts with the ultimate goal to construct credible intervals for a mean. The evaluation of the resulting intervals is in terms of “mismatched” priors and posteriors. The discussion is in the form of an imaginary dialog between a teacher and a student, who have met earlier, discussing and evaluating the Wald and score confidence intervals, as well as confidence intervals based on transformation and bootstrap techniques. From the perspective of the student the learning process is akin to a real research situation. The student is supposed to have studied mathematical statistics for at least two semesters.  相似文献   

3.
Correlated survival data arise frequently in biomedical and epidemiologic research, because each patient may experience multiple events or because there exists clustering of patients or subjects, such that failure times within the cluster are correlated. In this paper, we investigate the appropriateness of the semi-parametric Cox regression and of the generalized estimating equations as models for clustered failure time data that arise from an epidemiologic study in veterinary medicine. The semi-parametric approach is compared with a proposed fully parametric frailty model. The frailty component is assumed to follow a gamma distribution. Estimates of the fixed covariates effects were obtained by maximizing the likelihood function, while an estimate of the variance component ( frailty parameter) was obtained from a profile likelihood construction.  相似文献   

4.
The frailty model in survival analysis accounts for unobserved heterogeneity between individuals by assuming that the hazard rate of an individual is the product of an individual specific quantity, called “frailty” and a baseline hazard rate. It is well known that the choice of the frailty distribution strongly affects the nonparametric estimate of the baseline hazard as well as that of the conditional probabilities. This paper reviews the basic concepts of a frailty model, presents various probability inequalities and other monotonicity results which may prove useful in choosing among alternative specifications. More specifically, our main result lies in seeing how well known stochastic orderings between distributions of two frailities translate into orderings between the corresponding survival functions. Some probabilistic aspects and implications of the models resulting from competing choices of the distributions of frailty or the baseline are compared.  相似文献   

5.
We extend a diagnostic plot for the frailty distribution in proportional hazards models to the case of shared frailty. The plot is based on a closure property of exponential family failure distributions with canonical statistics z and g(z), namely that the frailty distribution among survivors at time t has the same form, with the same values of the parameters associated with g(z). We extend this property to shared frailty, considering various definitions of a “surviving” cluster at time t. We illustrate the effectiveness of the method in the case where the “death” of the cluster is defined by the first death among its members.  相似文献   

6.
In reliability analysis, it is common to consider several causes, either mechanical or electrical, those are competing to fail a unit. These causes are called “competing risks.” In this paper, we consider the simple step-stress model with competing risks for failure from Weibull distribution under progressive Type-II censoring. Based on the proportional hazard model, we obtain the maximum likelihood estimates (MLEs) of the unknown parameters. The confidence intervals are derived by using the asymptotic distributions of the MLEs and bootstrap method. For comparison, we obtain the Bayesian estimates and the highest posterior density (HPD) credible intervals based on different prior distributions. Finally, their performance is discussed through simulations.  相似文献   

7.
Survival models involving frailties are commonly applied in studies where correlated event time data arise due to natural or artificial clustering. In this paper we present an application of such models in the animal breeding field. Specifically, a mixed survival model with a multivariate correlated frailty term is proposed for the analysis of data from over 3611 Brazilian Nellore cattle. The primary aim is to evaluate parental genetic effects on the trait length in days that their progeny need to gain a commercially specified standard weight gain. This trait is not measured directly but can be estimated from growth data. Results point to the importance of genetic effects and suggest that these models constitute a valuable data analysis tool for beef cattle breeding.  相似文献   

8.
王星  马璇 《统计研究》2015,32(10):74-81
文章旨在研究受航空业动态定价机制影响下的机票价格序列变点估计模型,文中分析了机票价格u8序列数据的结构特点,提出了可用于高噪声数据环境下、阶梯状、带明显多变点的多阶段序列变点估计框架,该框架中级联组合了DBSCAN算法、EM-高斯混合模型聚类、凝聚层次聚类算法和基于乘积划分模型的变点估计方法等多种成熟的数据分析方法,通过对“北京-昆明”航线航班的实证分析,验证了数据分析框架的有效性和普遍适用性。  相似文献   

9.
We study a Bayesian approach to recovering the initial condition for the heat equation from noisy observations of the solution at a later time. We consider a class of prior distributions indexed by a parameter quantifying “smoothness” and show that the corresponding posterior distributions contract around the true parameter at a rate that depends on the smoothness of the true initial condition and the smoothness and scale of the prior. Correct combinations of these characteristics lead to the optimal minimax rate. One type of priors leads to a rate-adaptive Bayesian procedure. The frequentist coverage of credible sets is shown to depend on the combination of the prior and true parameter as well, with smoother priors leading to zero coverage and rougher priors to (extremely) conservative results. In the latter case, credible sets are much larger than frequentist confidence sets, in that the ratio of diameters diverges to infinity. The results are numerically illustrated by a simulated data example.  相似文献   

10.
We propose replacing the usual Student's-t statistic, which tests for equality of means of two distributions and is used to construct a confidence interval for the difference, by a biweight-“t” statistic. The biweight-“t” is a ratio of the difference of the biweight estimates of location from the two samples to an estimate of the standard error of this difference. Three forms of the denominator are evaluated: weighted variance estimates using both pooled and unpooled scale estimates, and unweighted variance estimates using an unpooled scale estimate. Monte Carlo simulations reveal that resulting confidence intervals are highly efficient on moderate sample sizes, and that nominal levels are nearly attained, even when considering extreme percentage points.  相似文献   

11.
Multi-type insurance claim processes have attracted considerable research interest in the literature. The existing statistical inference for such processes, however, may encounter “curse of dimensionality” due to high-dimensional covariates. In this article, a technique of sufficient dimension reduction is applied to multiple-type insurance claim data, which uses a copula to model the dependence between different types of claim processes, and incorporates a one-dimensional frailty to fit the dependence of claims “within” the same claim process. A two-step procedure is proposed to estimate model parameters. The first step develops nonparametric estimators of the baseline, the basis of the central subspace and its dimension, and the regression function. Then the second step estimates the copula parameter. Simulations are performed to evaluate and confirm the theoretical results.  相似文献   

12.
For interval estimation of a proportion, coverage probabilities tend to be too large for “exact” confidence intervals based on inverting the binomial test and too small for the interval based on inverting the Wald large-sample normal test (i.e., sample proportion ± z-score × estimated standard error). Wilson's suggestion of inverting the related score test with null rather than estimated standard error yields coverage probabilities close to nominal confidence levels, even for very small sample sizes. The 95% score interval has similar behavior as the adjusted Wald interval obtained after adding two “successes” and two “failures” to the sample. In elementary courses, with the score and adjusted Wald methods it is unnecessary to provide students with awkward sample size guidelines.  相似文献   

13.
This article presents the results of a simulation study investigating the performance of an approach developed by Miller and Landis (1991) for the analysis of clustered categorical responses. Evaluation of this “two-step” approach, which utilizes the method of moments to estimate the extra-variation pardmeters and subsequently incorporates these parameters into estimating equations for modelling the marginal expectations, is carried out in an experimental setting involving a comparison between two groups of observations. We assume that data for both groups are collected from each cluster and responses are measured on a three-point ordinal scale. The performance of the estimators used in both “steps” of the analysisis investigated and comparisons are made to an alternative analysismethod that ignores the clustering. The results indicate that in the chosen setting the test for a difference between groups generally operatbs at the nominal α=0.05 for 10 or more clusters and hasincreasing power with both an increasing number of clusters and an inrreasing treatment effect. These results provide a striking contrasc to those obtained from an improper analysis that ignores clustering.  相似文献   

14.
We introduce a matrix operator, which we call “vecd” operator. This operator stacks up “diagonals” of a symmetric matrix. This operator is more convenient for some statistical analyses than the commonly used “vech” operator. We show an explicit relationship between the vecd and vech operators. Using this relationship, various properties of the vecd operator are derived. As applications of the vecd operator, we derive concise and explicit expressions of the Wald and score tests for equal variances of a multivariate normal distribution and for the diagonality of variance coefficient matrices in a multivariate generalized autoregressive conditional heteroscedastic (GARCH) model, respectively.  相似文献   

15.
In many engineering problems it is necessary to draw statistical inferences on the mean of a lognormal distribution based on a complete sample of observations. Statistical demonstration of mean time to repair (MTTR) is one example. Although optimum confidence intervals and hypothesis tests for the lognormal mean have been developed, they are difficult to use, requiring extensive tables and/or a computer. In this paper, simplified conservative methods for calculating confidence intervals or hypothesis tests for the lognormal mean are presented. In this paper, “conservative” refers to confidence intervals (hypothesis tests) whose infimum coverage probability (supremum probability of rejecting the null hypothesis taken over parameter values under the null hypothesis) equals the nominal level. The term “conservative” has obvious implications to confidence intervals (they are “wider” in some sense than their optimum or exact counterparts). Applying the term “conservative” to hypothesis tests should not be confusing if it is remembered that this implies that their equivalent confidence intervals are conservative. No implication of optimality is intended for these conservative procedures. It is emphasized that these are direct statistical inference methods for the lognormal mean, as opposed to the already well-known methods for the parameters of the underlying normal distribution. The method currently employed in MIL-STD-471A for statistical demonstration of MTTR is analyzed and compared to the new method in terms of asymptotic relative efficiency. The new methods are also compared to the optimum methods derived by Land (1971, 1973).  相似文献   

16.
In the prospective study of a finely stratified population, one individual from each stratum is chosen at random for the “treatment” group and one for the “non-treatment” group. For each individual the probability of failure is a logistic function of parameters designating the stratum, the treatment and a covariate. Uniformly most powerful unbiased tests for the treatment effect are given. These tests are generally cumbersome but, if the covariate is dichotomous, the tests and confidence intervals are simple. Readily usable (but non-optimal) tests are also proposed for poly-tomous covariates and factorial designs. These are then adapted to retrospective studies (in which one “success” and one “failure” per stratum are sampled). Tests for retrospective studies with a continuous “treatment” score are also proposed.  相似文献   

17.
The well-known Wilson and Agresti–Coull confidence intervals for a binomial proportion p are centered around a Bayesian estimator. Using this as a starting point, similarities between frequentist confidence intervals for proportions and Bayesian credible intervals based on low-informative priors are studied using asymptotic expansions. A Bayesian motivation for a large class of frequentist confidence intervals is provided. It is shown that the likelihood ratio interval for p approximates a Bayesian credible interval based on Kerman’s neutral noninformative conjugate prior up to O(n? 1) in the confidence bounds. For the significance level α ? 0.317, the Bayesian interval based on the Jeffreys’ prior is then shown to be a compromise between the likelihood ratio and Wilson intervals. Supplementary materials for this article are available online.  相似文献   

18.
面板数据聚类方法及应用   总被引:7,自引:0,他引:7  
 基于面板数据的时序特征和截面特征,综合考虑面板数据“绝对指标”,“增量指标”及其“时序波动”特征,在重构面板数据相似性测度的距离函数和Ward聚类算法的基础上,提出了面板数据聚类方法。并以2003-2007年财政金融面板数据为例,对中国14个沿海开放城市进行了聚类分析,显示了良好的应用性。  相似文献   

19.
In the analysis of semi‐competing risks data interest lies in estimation and inference with respect to a so‐called non‐terminal event, the observation of which is subject to a terminal event. Multi‐state models are commonly used to analyse such data, with covariate effects on the transition/intensity functions typically specified via the Cox model and dependence between the non‐terminal and terminal events specified, in part, by a unit‐specific shared frailty term. To ensure identifiability, the frailties are typically assumed to arise from a parametric distribution, specifically a Gamma distribution with mean 1.0 and variance, say, σ2. When the frailty distribution is misspecified, however, the resulting estimator is not guaranteed to be consistent, with the extent of asymptotic bias depending on the discrepancy between the assumed and true frailty distributions. In this paper, we propose a novel class of transformation models for semi‐competing risks analysis that permit the non‐parametric specification of the frailty distribution. To ensure identifiability, the class restricts to parametric specifications of the transformation and the error distribution; the latter are flexible, however, and cover a broad range of possible specifications. We also derive the semi‐parametric efficient score under the complete data setting and propose a non‐parametric score imputation method to handle right censoring; consistency and asymptotic normality of the resulting estimators is derived and small‐sample operating characteristics evaluated via simulation. Although the proposed semi‐parametric transformation model and non‐parametric score imputation method are motivated by the analysis of semi‐competing risks data, they are broadly applicable to any analysis of multivariate time‐to‐event outcomes in which a unit‐specific shared frailty is used to account for correlation. Finally, the proposed model and estimation procedures are applied to a study of hospital readmission among patients diagnosed with pancreatic cancer.  相似文献   

20.
The role of Wikipedia for learning has been debated because it does not conform to the usual standards. Despite this, people use it, due to the ubiquity of Wikipedia entries in the outcomes from popular search engines. It is important for academic disciplines, including statistics, to ensure they are correctly represented in a medium where anyone can assume the role of discipline expert. In this context, we first develop a tool for evaluating Wikipedia articles for topics with a procedural component. Then, using this tool, five Wikipedia articles on basic statistical concepts are critiqued from the point of view of a self-learner: “arithmetic mean,” “standard deviation,” “standard error,” “confidence interval,” and “histogram.” We find that the articles, in general, are poor, and some articles contain inaccuracies. We propose that Wikipedia be actively discouraged for self-learning (using, for example, a classroom activity) except to give a brief overview; that in more formal learning environments, teachers be explicit about not using Wikipedia as a learning resource for course content; and, because Wikipedia is used regardless of considered advice or the organizational protocols in place, teachers move away from minimal contact with Wikipedia towards more constructive engagement.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号