首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 248 毫秒
1.
The influence of individual points in an ordinal logistic model is considered when the aim is to determine their effects on the predictive probability in a Bayesian predictive approach. Our concern is to study the effects produced when the data are slightly perturbed, in particular by observing how these perturbations will affect the predictive probabilities and consequently the classification of future cases. We consider the extent of the change in the predictive distribution when an individual point is omitted (deleted) from the sample by use of a divergence measure suggested by Johnson (1985) as a measure of discrepancy between the full data and the data with the case deleted. The methodology is illustrated on some data used in Titterington et al. (1981).  相似文献   

2.

Ordinal data are often modeled using a continuous latent response distribution, which is partially observed through windows of adjacent intervals defined by cutpoints. In this paper we propose the beta distribution as a model for the latent response. The beta distribution has several advantages over the other common distributions used, e.g. , normal and logistic. In particular, it enables separate modeling of location and dispersion effects which is essential in the Taguchi method of robust design. First, we study the problem of estimating the location and dispersion parameters of a single beta distribution (representing a single treatment) from ordinal data assuming known equispaced cutpoints. Two methods of estimation are compared: the maximum likelihood method and the method of moments. Two methods of treating the data are considered: in raw discrete form and in smoothed continuousized form. A large scale simulation study is carried out to compare the different methods. The mean square errors of the estimates are obtained under a variety of parameter configurations. Comparisons are made based on the ratios of the mean square errors (called the relative efficiencies). No method is universally the best, but the maximum likelihood method using continuousized data is found to perform generally well, especially for estimating the dispersion parameter. This method is also computationally much faster than the other methods and does not experience convergence difficulties in case of sparse or empty cells. Next, the problem of estimating unknown cutpoints is addressed. Here the multiple treatments setup is considered since in an actual application, cutpoints are common to all treatments, and must be estimated from all the data. A two-step iterative algorithm is proposed for estimating the location and dispersion parameters of the treatments, and the cutpoints. The proposed beta model and McCullagh's (1980) proportional odds model are compared by fitting them to two real data sets.  相似文献   

3.
In many situations, instead of a complete sample, data are available only in grouped form. For example, grouped failure time data occur in studies in which subjects are monitored periodically to determine whether failure has occurred in the predetermined intervals. Here the model under consideration is the log-logistic distribution. This paper demonstrates the existence and uniqueness of the MLEs of the parameters of the logistic distribution under mild conditions with grouped data. The times with the maximum failure rate and the mode of the p.d.f. of the log-logistic distribution are also estimated based on the MLEs. The methodology is further studied with simulations and exemplified with a data set with artificially introduced grouping from a locomotive life test study.  相似文献   

4.
This work develops a new methodology in order to discriminate models for interval- censored data based on bootstrap residual simulation by observing the deviance difference from one model in relation to another, according to Hinde (1992). Generally, this sort of data can generate a large number of tied observations and, in this case, survival time can be regarded as discrete. Therefore, the Cox proportional hazards model for grouped data (Prentice & Gloeckler, 1978) and the logistic model (Lawless, 1982) can be fitted by means of generalized linear models. Whitehead (1989) considered censoring to be an indicative variable with a binomial distribution and fitted the Cox proportional hazards model using complementary log-log as a link function. In addition, a logistic model can be fitted using logit as a link function. The proposed methodology arises as an alternative to the score tests developed by Colosimo et al. (2000), where such models can be obtained for discrete binary data as particular cases from the Aranda-Ordaz distribution asymmetric family. These tests are thus developed with a basis on link functions to generate such a fit. The example that motivates this study was the dataset from an experiment carried out on a flax cultivar planted on four substrata susceptible to the pathogen Fusarium oxysoprum . The response variable, which is the time until blighting, was observed in intervals during 52 days. The results were compared with the model fit and the AIC values.  相似文献   

5.
Cox's discrete logistic model was extended to the study of the life table by Thompson (1977) to handle grouped survival data. Inferences about the effect of grouping are studies byMonte Carlo methods. The results show that the effect of grouping is not substantial. This approach is applied to the grouped data on liver cancer. The computer program developed for grouped censored data with continuous and indicator covariates is of practical importance and is available fromThe Ohio State University  相似文献   

6.
In dose-response models, there are cases where only a portion of the administered dose may have an effect. This results in a stochastic compliance of the administered dose. In a previous paper (Chen-Mok and Sen, 1999), we developed suitable adjustments for compliance in the logistic model under the assumption of nondifferential measurement error. These compliance-adjusted models were categorized into three types: (i) Low (or near zero) dose levels, (ii) moderate dose levels, and (iii) high dose levels. In this paper, we analyze a set of data on the atomic bomb survivors of Japan to illustrate the use of the proposed methods. In addition, we examine the performance of these methods under different conditions based on a simulation study. Among all three cases, the adjustments proposed for the moderate dose case do not seem to work adequately. Both bias and variance are larger when using the adjusted model in comparison with the unadjusted model. The adjustments for the low dose case seem to work in reducing the bias in the estimation of the parameters under all types of compliance distributions. The MSEs, however, are larger under some of the compliance distribution considered. Finally, the results of this simulation study show that the adjustments for the high dose case are successful in achieving both a reduction in bias as well as a reduction in MSE, hence the overall efficiency of the estimation is improved.  相似文献   

7.
8.
In contrast to the common belief that the logit model has no analytical presentation, it is possible to find such a solution in the case of categorical predictors. This paper shows that a binary logistic regression by categorical explanatory variables can be constructed in a closed-form solution. No special software and no iterative procedures of nonlinear estimation are needed to obtain a model with all its parameters and characteristics, including coefficients of regression, their standard errors and t-statistics, as well as the residual and null deviances. The derivation is performed for logistic models with one binary or categorical predictor, and several binary or categorical predictors. The analytical formulae can be used for arithmetical calculation of all the parameters of the logit regression. The explicit expressions for the characteristics of logit regression are convenient for the analysis and interpretation of the results of logistic modeling.  相似文献   

9.
The lognormal and Weibull distributions are the most popular distributions for modeling lifetime data. In practical applications, they usually fit the data at hand well. However, their predictions may lead to large differences. The main purpose of the present article is to investigate the impacts of mis-specification between the lognormal and Weibull distributions on the interval estimation of a pth quantile of the distributions for complete data. The coverage probabilities of the confidence intervals (CIs) with mis-specification are evaluated. The results indicate that for both the lognormal and the Weibull distributions, the coverage probabilities are significantly influenced by mis-specification, especially for a small or a large p on lower or upper tail of the distributions. In addition, based on the coverage probabilities with correct and mis-specification, a maxmin criterion is proposed to make a choice between these two distributions. The numerical results indicate that for p ≤ 0.05 and 0.6 ≤ p ≤ 0.8, Weibull distribution is suggested to evaluate CIs of a pth quantile of the distributions, while, for 0.2 ≤ p ≤ 0.5 and p = 0.99, lognormal distribution is suggested to evaluate CIs of a pth quantile of the distributions. Besides, for p = 0.9 and 0.95, lognormal distribution is suggested if the sample size is large enough, while, for p = 0.1, Weibull distribution is suggested if the sample size is large enough. Finally, a simulation study is conducted to evaluate the efficiency of the proposed method.  相似文献   

10.
Abstract.  Epidemiology research often entails the analysis of failure times subject to grouping. In large cohorts interval grouping also offers a feasible choice of data reduction to actually facilitate an analysis of the data. Based on an underlying Cox proportional hazards model for the exact failure times one may deduce a grouped data version of this model which may then be used to analyse the data. The model bears a lot of resemblance to a generalized linear model, yet due to the nature of data one also needs to incorporate censoring. In the case of non-trivial censoring this precludes model checking procedures based on ordinary residuals as calculation of these requires knowledge of the censoring distribution. In this paper, we represent interval grouped data in a dynamical way using a counting process approach. This enables us to identify martingale residuals which can be computed without knowledge of the censoring distribution. We use these residuals to construct graphical as well as numerical model checking procedures. An example from epidemiology is provided.  相似文献   

11.
This paper deals with the estimation of R=P[X<Y] when X and Y come from two independent generalized logistic distributions with different parameters. The maximum-likelihood estimator (MLE) and its asymptotic distribution are proposed. The asymptotic distribution is used to construct an asymptotic confidence interval of R. Assuming that the common scale parameter is known, the MLE, uniformly minimum variance unbiased estimator, Bayes estimation and confidence interval of R are obtained. The MLE of R, asymptotic distribution of R in the general case, is also discussed. Monte Carlo simulations are performed to compare the different proposed methods. Analysis of a real data set has also been presented for illustrative purposes.  相似文献   

12.
Skew normal distribution is an alternative distribution to the normal distribution to accommodate asymmetry. Since then extensive studies have been done on applying Azzalini’s skewness mechanism to other well-known distributions, such as skew-t distribution, which is more flexible and can better accommodate long tailed data than the skew normal one. The Kumaraswamy generalized distribution (Kw ? F) is another new class of distribution which is capable of fitting skewed data that can not be fitted well by existing distributions. Such a distribution has been widely studied and various versions of generalization of this distribution family have been introduced. In this article, we introduce a new generalization of the skew-t distribution based on the Kumaraswamy generalized distribution. The new class of distribution, which we call the Kumaraswamy skew-t (KwST) has the ability of fitting skewed, long, and heavy-tailed data and is more flexible than the skew-t distribution as it contains the skew-t distribution as a special case. Related properties of this distribution family such as mathematical properties, moments, and order statistics are discussed. The proposed distribution is applied to a real dataset to illustrate the estimation procedure.  相似文献   

13.
This article considers the problem of testing the validity of the assumption that the underlying distribution of life is Pareto. For complete and censored samples, the relationship between the Pareto and the exponential distributions could be of vital importance to test for the validity of this assumption. For grouped uncensored data the classical Pearson χ2 test based on the multinomial model can be used. Attention is confined in this article to handle grouped data with withdrawals within intervals. Graphical as well as analytical procedures will be presented. Maximum likelihood estimators for the parameters of the Pareto distribution based on grouped data will be derived.  相似文献   

14.
Matched case–control designs are commonly used in epidemiological studies for estimating the effect of exposure variables on the risk of a disease by controlling the effect of confounding variables. Due to retrospective nature of the study, information on a covariate could be missing for some subjects. A straightforward application of the conditional logistic likelihood for analyzing matched case–control data with the partially missing covariate may yield inefficient estimators of the parameters. A robust method has been proposed to handle this problem using an estimated conditional score approach when the missingness mechanism does not depend on the disease status. Within the conditional logistic likelihood framework, an empirical procedure is used to estimate the odds of the disease for the subjects with missing covariate values. The asymptotic distribution and the asymptotic variance of the estimator when the matching variables and the completely observed covariates are categorical. The finite sample performance of the proposed estimator is assessed through a simulation study. Finally, the proposed method has been applied to analyze two matched case–control studies. The Canadian Journal of Statistics 38: 680–697; 2010 © 2010 Statistical Society of Canada  相似文献   

15.
ABSTRACT

The Concordance statistic (C-statistic) is commonly used to assess the predictive performance (discriminatory ability) of logistic regression model. Although there are several approaches for the C-statistic, their performance in quantifying the subsequent improvement in predictive accuracy due to inclusion of novel risk factors or biomarkers in the model has been extremely criticized in literature. This paper proposed a model-based concordance-type index, CK, for use with logistic regression model. The CK and its asymptotic sampling distribution is derived following Gonen and Heller's approach for Cox PH model for survival data but taking necessary modifications for use with binary data. Unlike the existing C-statistics for logistic model, it quantifies the concordance probability by taking the difference in the predicted risks between two subjects in a pair rather than ranking them and hence is able to quantify the equivalent incremental value from the new risk factor or marker. The simulation study revealed that the CK performed well when the model parameters are correctly estimated for large sample and showed greater improvement in quantifying the additional predictive value from the new risk factor or marker than the existing C-statistics. Furthermore, the illustration using three datasets supports the findings from simulation study.  相似文献   

16.
In this paper, a discrete counterpart of the general class of continuous beta-G distributions is introduced. A discrete analog of the beta generalized exponential distribution of Barreto-Souza et al. [2], as an important special case of the proposed class, is studied. This new distribution contains some previously known discrete distributions as well as two new models. The hazard rate function of the new model can be increasing, decreasing, bathtub-shaped and upside-down bathtub. Some distributional and moment properties of the new distribution as well as its order statistics are discussed. Estimation of the parameters is illustrated using the maximum likelihood method and, finally, the model with a real data set is examined.  相似文献   

17.
ABSTRACT

The logistic distribution has a prominent role in the theory and practice of statistics. We introduce a new family of continuous distributions generated from a logistic random variable called the logistic-X family. Its density function can be symmetrical, left-skewed, right-skewed, and reversed-J shaped, and can have increasing, decreasing, bathtub, and upside-down bathtub hazard rates shaped. Further, it can be expressed as a linear combination of exponentiated densities based on the same baseline distribution. We derive explicit expressions for the ordinary and incomplete moments, quantile and generating functions, Bonferroni and Lorenz curves, Shannon entropy, and order statistics. The model parameters are estimated by the method of maximum likelihood and the observed information matrix is determined. We also investigate the properties of one special model, the logistic-Fréchet distribution, and illustrate its importance by means of two applications to real data sets.  相似文献   

18.
A location sensitive one—sample test V(n,n) similar to the Wilcoxon two—sample test was proposed by Riedwyl (1967) and stuided by Carnal and Riedwyl (1972). A generalization for grouped data was given by Maag, Streit and Drouilly (1973). In the present paper we discuss the application of the test for grouped data. We present a table of the significance limits and discuss the approximation by means of the normal distribution.  相似文献   

19.
We consider the right truncated exponential distribution where the truncation point is unknown and show that the ML equation has a unique solution over an extended parameter space. In the case of the estimation of the truncation point T we show that the asymptotic distribution of the MLE is not centered at T. A modified MLE is introduced which outperforms all other considered estimators including the minimum variance unbiased estimator. Asymptotic as well as small sample properties of different estimators are investigated and compared. The truncated exponential distribution has an increasing failure rate, ideally suited for use as a survival distribution for biological and industrial data.  相似文献   

20.
Abstract

This paper investigates the statistical analysis of grouped accelerated temperature cycling test data when the product lifetime follows a Weibull distribution. A log-linear acceleration equation is derived from the Coffin-Manson model. The problem is transformed to a constant-stress accelerated life test with grouped data and multiple acceleration variables. The Jeffreys prior and reference priors are derived. Maximum likelihood estimation and Bayesian estimation with objective priors are obtained by applying the technique of data augmentation. A simulation study shows that both of these two methods perform well when sample size is large, and the Bayesian method gives better performance under small sample sizes.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号