首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
In studies that involve censored time-to-event data, stratification is frequently encountered due to different reasons, such as stratified sampling or model adjustment due to violation of model assumptions. Often, the main interest is not in the clustering variables, and the cluster-related parameters are treated as nuisance. When inference is about a parameter of interest in presence of many nuisance parameters, standard likelihood methods often perform very poorly and may lead to severe bias. This problem is particularly evident in models for clustered data with cluster-specific nuisance parameters, when the number of clusters is relatively high with respect to the within-cluster size. However, it is still unclear how the presence of censoring would affect this issue. We consider clustered failure time data with independent censoring, and propose frequentist inference based on an integrated likelihood. We then apply the proposed approach to a stratified Weibull model. Simulation studies show that appropriately defined integrated likelihoods provide very accurate inferential results in all circumstances, such as for highly clustered data or heavy censoring, even in extreme settings where standard likelihood procedures lead to strongly misleading results. We show that the proposed method performs generally as well as the frailty model, but it is superior when the frailty distribution is seriously misspecified. An application, which concerns treatments for a frequent disease in late-stage HIV-infected people, illustrates the proposed inferential method in Weibull regression models, and compares different inferential conclusions from alternative methods.  相似文献   

2.
This article deals with model comparison as an essential part of generalized linear modelling in the presence of covariates missing not at random (MNAR). We provide an evaluation of the performances of some of the popular model selection criteria, particularly of deviance information criterion (DIC) and weighted L (WL) measure, for comparison among a set of candidate MNAR models. In addition, we seek to provide deviance and quadratic loss-based model selection criteria with alternative penalty terms targeting directly the MNAR models. This work is motivated by the need in the literature to understand the performances of these important model selection criteria for comparison among a set of MNAR models. A Monte Carlo simulation experiment is designed to assess the finite sample performances of these model selection criteria in the context of interest under different scenarios for missingness amounts. Some naturally driven DIC and WL extensions are also discussed and evaluated.  相似文献   

3.
Classical inferential procedures induce conclusions from a set of data to a population of interest, accounting for the imprecision resulting from the stochastic component of the model. Less attention is devoted to the uncertainty arising from (unplanned) incompleteness in the data. Through the choice of an identifiable model for non-ignorable non-response, one narrows the possible data-generating mechanisms to the point where inference only suffers from imprecision. Some proposals have been made for assessing the sensitivity to these modelling assumptions; many are based on fitting several plausible but competing models. For example, we could assume that the missing data are missing at random in one model, and then fit an additional model where non-random missingness is assumed. On the basis of data from a Slovenian plebiscite, conducted in 1991, to prepare for independence, it is shown that such an ad hoc procedure may be misleading. We propose an approach which identifies and incorporates both sources of uncertainty in inference: imprecision due to finite sampling and ignorance due to incompleteness. A simple sensitivity analysis considers a finite set of plausible models. We take this idea one step further by considering more degrees of freedom than the data support. This produces sets of estimates (regions of ignorance) and sets of confidence regions (combined into regions of uncertainty).  相似文献   

4.
Summary. We examine three pattern–mixture models for making inference about parameters of the distribution of an outcome of interest Y that is to be measured at the end of a longitudinal study when this outcome is missing in some subjects. We show that these pattern–mixture models also have an interpretation as selection models. Because these models make unverifiable assumptions, we recommend that inference about the distribution of Y be repeated under a range of plausible assumptions. We argue that, of the three models considered, only one admits a parameterization that facilitates the examination of departures from the assumption of sequential ignorability. The three models are nonparametric in the sense that they do not impose restrictions on the class of observed data distributions. Owing to the curse of dimensionality, the assumptions that are encoded in these models are sufficient for identification but not for inference. We describe additional flexible and easily interpretable assumptions under which it is possible to construct estimators that are well behaved with moderate sample sizes. These assumptions define semiparametric models for the distribution of the observed data. We describe a class of estimators which, up to asymptotic equivalence, comprise all the consistent and asymptotically normal estimators of the parameters of interest under the postulated semiparametric models. We illustrate our methods with the analysis of data from a randomized clinical trial of contracepting women.  相似文献   

5.
Cross-over trials with correlated Bernoulli outcomes are common designs. In condom functionality studies, for example, an indicator of condom failure is reported for each sex act using standard or experimental condoms. Two popular analysis methods for such data are Generalized Estimating Equations and logit-normal random effects models. An alternative random effects model, the beta-binomial, is commonly used in contexts involving only between-cluster effects. The flexibility of the beta distribution and the interpretation of random effects as cluster-specific failure probabilities make it appealing, and we consider an extension of the model to account for within-cluster treatment effects using proportional odds assumptions.  相似文献   

6.
A well-known problem in multiple regression is that it is possible to reject the hypothesis that all slope parameters are equal to zero, yet when applying the usual Student's T-test to the individual parameters, no significant differences are found. An alternative strategy is to estimate prediction error via the 0.632 bootstrap method for all models of interest and declare the parameters associated with the model that yields the smallest prediction error to differ from zero. The main results in this paper are that this latter strategy can have practical value versus Student's T; replacing squared error with absolute error can be beneficial in some situations and replacing least squares with an extension of the Theil-Sen estimator can substantially increase the probability of identifying the correct model under circumstances that are described.  相似文献   

7.
Recursive partitioning algorithms separate a feature space into a set of disjoint rectangles. Then, usually, a constant in every partition is fitted. While this is a simple and intuitive approach, it may still lack interpretability as to how a specific relationship between dependent and independent variables may look. Or it may be that a certain model is assumed or of interest and there is a number of candidate variables that may non-linearly give rise to different model parameter values. We present an approach that combines generalized linear models (GLM) with recursive partitioning that offers enhanced interpretability of classical trees as well as providing an explorative way to assess a candidate variable's influence on a parametric model. This method conducts recursive partitioning of a GLM by (1) fitting the model to the data set, (2) testing for parameter instability over a set of partitioning variables, (3) splitting the data set with respect to the variable associated with the highest instability. The outcome is a tree where each terminal node is associated with a GLM. We will show the method's versatility and suitability to gain additional insight into the relationship of dependent and independent variables by two examples, modelling voting behaviour and a failure model for debt amortization, and compare it to alternative approaches.  相似文献   

8.
A randomized trial allows estimation of the causal effect of an intervention compared to a control in the overall population and in subpopulations defined by baseline characteristics. Often, however, clinical questions also arise regarding the treatment effect in subpopulations of patients, which would experience clinical or disease related events post-randomization. Events that occur after treatment initiation and potentially affect the interpretation or the existence of the measurements are called intercurrent events in the ICH E9(R1) guideline. If the intercurrent event is a consequence of treatment, randomization alone is no longer sufficient to meaningfully estimate the treatment effect. Analyses comparing the subgroups of patients without the intercurrent events for intervention and control will not estimate a causal effect. This is well known, but post-hoc analyses of this kind are commonly performed in drug development. An alternative approach is the principal stratum strategy, which classifies subjects according to their potential occurrence of an intercurrent event on both study arms. We illustrate with examples that questions formulated through principal strata occur naturally in drug development and argue that approaching these questions with the ICH E9(R1) estimand framework has the potential to lead to more transparent assumptions as well as more adequate analyses and conclusions. In addition, we provide an overview of assumptions required for estimation of effects in principal strata. Most of these assumptions are unverifiable and should hence be based on solid scientific understanding. Sensitivity analyses are needed to assess robustness of conclusions.  相似文献   

9.
Likelihood-based, mixed-effects models for repeated measures (MMRMs) are occasionally used in primary analyses for group comparisons of incomplete continuous longitudinal data. Although MMRM analysis is generally valid under missing-at-random assumptions, it is invalid under not-missing-at-random (NMAR) assumptions. We consider the possibility of bias of estimated treatment effect using standard MMRM analysis in a motivational case, and propose simple and easily implementable pattern mixture models within the framework of mixed-effects modeling, to handle the NMAR data with differential missingness between treatment groups. The proposed models are a new form of pattern mixture model that employ a categorical time variable when modeling the outcome and a continuous time variable when modeling the missingness-data patterns. The models can directly provide an overall estimate of the treatment effect of interest using the average of the distribution of the missingness indicator and a categorical time variable in the same manner as MMRM analysis. Our simulation results indicate that the bias of the treatment effect for MMRM analysis was considerably larger than that for the pattern mixture model analysis under NMAR assumptions. In the case study, it would be dangerous to interpret only the results of the MMRM analysis, and the proposed pattern mixture model would be useful as a sensitivity analysis for treatment effect evaluation.  相似文献   

10.
Recurrent events involve the occurrences of the same type of event repeatedly over time and are commonly encountered in longitudinal studies. Examples include seizures in epileptic studies or occurrence of cancer tumors. In such studies, interest lies in the number of events that occur over a fixed period of time. One considerable challenge in analyzing such data arises when a large proportion of patients discontinues before the end of the study, for example, because of adverse events, leading to partially observed data. In this situation, data are often modeled using a negative binomial distribution with time‐in‐study as offset. Such an analysis assumes that data are missing at random (MAR). As we cannot test the adequacy of MAR, sensitivity analyses that assess the robustness of conclusions across a range of different assumptions need to be performed. Sophisticated sensitivity analyses for continuous data are being frequently performed. However, this is less the case for recurrent event or count data. We will present a flexible approach to perform clinically interpretable sensitivity analyses for recurrent event data. Our approach fits into the framework of reference‐based imputations, where information from reference arms can be borrowed to impute post‐discontinuation data. Different assumptions about the future behavior of dropouts dependent on reasons for dropout and received treatment can be made. The imputation model is based on a flexible model that allows for time‐varying baseline intensities. We assess the performance in a simulation study and provide an illustration with a clinical trial in patients who suffer from bladder cancer. Copyright © 2015 John Wiley & Sons, Ltd.  相似文献   

11.
Biomedical and psychosocial researchers increasingly utilize multiple indicators to assess an outcome of interest. We apply the ordinal estimating equations model for analysing this kind of measurement. We detail the special complexities of using this model to analyse clustered non-identical items and propose a workable model building strategy. Three graphical methods— cumulative log-odds, partial residual and Pearson residual plotting—are developed to diagnose the adequacy of models. The benefit of incorporating interitem associations and the trade-off between simple versus complex models are evaluated. Throughout the paper, an analysis to determine how measured impairments affect visual disability is used for illustration.  相似文献   

12.
Using the Savage–Dickey density ratio and an alternative approach that uses more relaxed assumptions, we develop methods to calculate the probability that a restriction holds at a point in time without assuming that the restriction holds at any other points in time. Both approaches use MCMC output only from the unrestricted model to compute the time-varying posterior probabilities for all models of interest. Using U.S. data, we find the probability that the long-run Phillips curve is vertical to be fairly high, but decreases over time. The probability that the NAIRU is not identified fluctuates over time, but increases after 1990.  相似文献   

13.

This work is motivated by the need to find experimental designs which are robust under different model assumptions. We measure robustness by calculating a measure of design efficiency with respect to a design optimality criterion and say that a design is robust if it is reasonably efficient under different model scenarios. We discuss two design criteria and an algorithm which can be used to obtain robust designs. The first criterion employs a Bayesian-type approach by putting a prior or weight on each candidate model and possibly priors on the corresponding model parameters. We define the first criterion as the expected value of the design efficiency over the priors. The second design criterion we study is the minimax design which minimizes the worst value of a design criterion over all candidate models. We establish conditions when these two criteria are equivalent when there are two candidate models. We apply our findings to the area of accelerated life testing and perform sensitivity analysis of designs with respect to priors and misspecification of planning values.  相似文献   

14.
The fundamental difficulty with inference in nontrivial extrapolation where model selection is involved from a rich space of models is that any model estimated in one regime used for decision making in another is fundamentally confounded with disruptive alternatives. These are alternative models which if true would support a diametrically opposed action from the one the estimated model supports. One strategy to support extrapolation and reduce arbitrary fitting and confounding is to force the model to derive from the same mathematical structure that underlies the substantive science appropriate for the phenomena. Then statistical model fitting follows the form of theory generation in artificial intelligence, with statistical model selection tools and the statistician taking the place of the inference engine.  相似文献   

15.
Recurrent events data are frequently encountered and could be stopped by a terminal event in clinical trials. It is of interest to assess the treatment efficacy simultaneously with respect to both the recurrent events and the terminal event in many applications. In this paper we propose joint covariate-adjusted score test statistics based on joint models of recurrent events and a terminal event. No assumptions on the functional form of the covariates are needed. Simulation results show that the proposed tests can improve the efficiency over tests based on covariate unadjusted model. The proposed tests are applied to the SOLVD data for illustration.  相似文献   

16.
Useful models for time series of counts or simply wrong ones?   总被引:1,自引:0,他引:1  
There has been a considerable and growing interest in low integer-valued time series data leading to a diversification of modelling approaches. In addition to static regression models, both observation-driven and parameter-driven models are considered here. We compare and contrast a variety of time series models for counts using two very different data sets as a testbed. A range of diagnostic devices is employed to help inform model adequacy. Special attention is paid to dynamic structure and underlying distributional assumptions including associated dispersion properties. Competing models show attractive features, but overall no one modelling approach is seen to dominate.  相似文献   

17.
By their very nature, statistical models are constructed on the basis of a number of simplifying assumptions. It is usual to assume that all the individuals in a 'group' or 'cohort' have similar survival, recovery or reporting probabilities. From a number of earlier studies of the Cape Griffon Gyps coprotheres in southern Africa, it is clear that there have been many violations of these assumptions of homogeneity. To get a better understanding of the process whereby a dead ringed bird is found and reported, an analysis was undertaken of 575 recoveries from 7130 individuals ringed as nestlings. From a series of univariate generalized linear models, it was found that the proportion of ringed birds reported dead varied with the following factors: (1) ring prefix (representing different grades and thicknesses of aluminium): there was considerable variation in reporting rate between cohorts fitted with different ring prefix series used; (2) metal type: birds fitted with monel metal rings were reported at a rate twice that of those bearing aluminium rings; (3) colour rings: recoveries of birds with colour rings were much more likely to be reported than birds with only a metal ring; (4) epoch: the reporting rate has increased steadily from the 1950s through to the mid-1980s. All of these factors are confounded and so a number of multivariate generalized linear models were constructed. It was found that the variations in the cohort-specific reporting rate could be described well by a model including factors for metal-ring type and the presence or absence of colour rings. Using the tougher monel metal ring along with a set of colour rings more than doubles the reporting rate and their continued use is strongly recommended for future studies. The year-to-year variations could be accounted for by this model but the colony of ringing did not enter the model. The models used were based on two assumptions: (i) the reporting rate was constant for all individuals within a given cohort and (ii) the recoveries were complete. It is argued that the results are congruent with these assumptions. There is now a clearer model of the manner in which the ring-recovery reporting process proceeds and this has opened the way to building a more realistic statistical model to estimate survival in the Cape Griffon.  相似文献   

18.
We have tested alternative models of the demand for medical care using experimental data. The estimated response of demand to insurance plan is sensitive to the model used. We therefore use a split-sample analysis and find that a model that more closely approximates distributional assumptions and uses a nonparametric retransformation factor performs better in terms of mean squared forecast error. Simpler models are inferior either because they are not robust to outliers (e.g., ANOVA, ANOCOVA), or because they are inconsistent when strong distributional assumptions are violated (e.g., a two-parameter Box-Cox transformation).  相似文献   

19.
By their very nature, statistical models are constructed on the basis of a number of simplifying assumptions. It is usual to assume that all the individuals in a 'group' or 'cohort' have similar survival, recovery or reporting probabilities. From a number of earlier studies of the Cape Griffon Gyps coprotheres in southern Africa, it is clear that there have been many violations of these assumptions of homogeneity. To get a better understanding of the process whereby a dead ringed bird is found and reported, an analysis was undertaken of 575 recoveries from 7130 individuals ringed as nestlings. From a series of univariate generalized linear models, it was found that the proportion of ringed birds reported dead varied with the following factors: (1) ring prefix (representing different grades and thicknesses of aluminium): there was considerable variation in reporting rate between cohorts fitted with different ring prefix series used; (2) metal type: birds fitted with monel metal rings were reported at a rate twice that of those bearing aluminium rings; (3) colour rings: recoveries of birds with colour rings were much more likely to be reported than birds with only a metal ring; (4) epoch: the reporting rate has increased steadily from the 1950s through to the mid-1980s. All of these factors are confounded and so a number of multivariate generalized linear models were constructed. It was found that the variations in the cohort-specific reporting rate could be described well by a model including factors for metal-ring type and the presence or absence of colour rings. Using the tougher monel metal ring along with a set of colour rings more than doubles the reporting rate and their continued use is strongly recommended for future studies. The year-to-year variations could be accounted for by this model but the colony of ringing did not enter the model. The models used were based on two assumptions: (i) the reporting rate was constant for all individuals within a given cohort and (ii) the recoveries were complete. It is argued that the results are congruent with these assumptions. There is now a clearer model of the manner in which the ring-recovery reporting process proceeds and this has opened the way to building a more realistic statistical model to estimate survival in the Cape Griffon.  相似文献   

20.
In this paper we address the problem of estimating a vector of regression parameters in the Weibull censored regression model. Our main objective is to provide natural adaptive estimators that significantly improve upon the classical procedures in the situation where some of the predictors may or may not be associated with the response. In the context of two competing Weibull censored regression models (full model and candidate submodel), we consider an adaptive shrinkage estimation strategy that shrinks the full model maximum likelihood estimate in the direction of the submodel maximum likelihood estimate. We develop the properties of these estimators using the notion of asymptotic distributional risk. The shrinkage estimators are shown to have higher efficiency than the classical estimators for a wide class of models. Further, we consider a LASSO type estimation strategy and compare the relative performance with the shrinkage estimators. Monte Carlo simulations reveal that when the true model is close to the candidate submodel, the shrinkage strategy performs better than the LASSO strategy when, and only when, there are many inactive predictors in the model. Shrinkage and LASSO strategies are applied to a real data set from Veteran's administration (VA) lung cancer study to illustrate the usefulness of the procedures in practice.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号