首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 640 毫秒
1.
Researchers have been developing various extensions and modified forms of the Weibull distribution to enhance its capability for modeling and fitting different data sets. In this note, we investigate the potential usefulness of the new modification to the standard Weibull distribution called odd Weibull distribution in income economic inequality studies. Some mathematical and statistical properties of this model are proposed. We obtain explicit expressions for the first incomplete moment, quantile function, Lorenz and Zenga curves and related inequality indices. In addition to the well-known stochastic order based on Lorenz curve, the stochastic order based on Zenga curve is considered. Since the new generalized Weibull distribution seems to be suitable to model wealth, financial, actuarial and especially income distributions, these findings are fundamental in the understanding of how parameter values are related to inequality. Also, the estimation of parameters by maximum likelihood and moment methods is discussed. Finally, this distribution has been fitted to United States and Austrian income data sets and has been found to fit remarkably well in compare with the other widely used income models.  相似文献   

2.
Discriminating between the generalized Rayleigh and Weibull distributions   总被引:1,自引:0,他引:1  
Generalized Rayleigh (GR) and Weibull (WE) distributions are used quite effectively for analysing skewed lifetime data. In this paper, we consider the problem of selecting either GR or WE distribution as a more appropriate fitting model for a given data set. We use the ratio of maximized likelihoods (RML) for discriminating between the two distributions. The asymptotic and simulated distributions of the logarithm of the RML are applied to determine the probability of correctly selecting between these two families of distributions. It is examined numerically that the asymptotic results work quite well even for small sample sizes. A real data set involving the annual rainfall recorded at Los Angeles Civic Center during 25 years is analysed to illustrate the procedures developed here.  相似文献   

3.
The generalized lambda distribution, GLD(λ1, λ2 λ3, λ4), is a four-parameter family that has been used for fitting distributions to a wide variety of data sets. The analysis of the λ3 and λ4 values that actually yield valid distributions has (until now) been incomplete. Moreover, because of computational problems and theoretical shortcomings, the moment space over which the GLD can be applied has been limited. This paper completes the analysis of the λ3 and λ4 values that are associated with valid distributions, improves previous computational methods to reduce errors associated with fitting data, expands the parameter space over which the GLD can be used, and uses a four-parameter generalized beta distribution to cover the portion of the parameter space where the GLD is not applicable. In short, the paper extends the GLD to an EGLD system that can be used for fitting distributions to data sets that that are cited in the literature as actually occurring in practice. Examples of use of the proposed system are included  相似文献   

4.
Generalized exponential distributions   总被引:8,自引:0,他引:8  
The three-parameter gamma and three-parameter Weibull distributions are commonly used for analysing any lifetime data or skewed data. Both distributions have several desirable properties, and nice physical interpretations. Because of the scale and shape parameters, both have quite a bit of flexibility for analysing different types of lifetime data. They have increasing as well as decreasing hazard rate depending on the shape parameter. Unfortunately both distributions also have certain drawbacks. This paper considers a three-parameter distribution which is a particular case of the exponentiated Weibull distribution originally proposed by Mudholkar, Srivastava & Freimer (1995) when the location parameter is not present. The study examines different properties of this model and observes that this family has some interesting features which are quite similar to those of the gamma family and the Weibull family, and certain distinct properties also. It appears this model can be used as an alternative to the gamma model or the Weibull model in many situations. One dataset is provided where the three-parameter generalized exponential distribution fits better than the three-parameter Weibull distribution or the three-parameter gamma distribution.  相似文献   

5.
When modeling multilevel data, it is important to accurately represent the interdependence of observations within clusters. Ignoring data clustering may result in parameter misestimation. However, it is not well established to what degree parameter estimates are affected by model misspecification when applying missing data techniques (MDTs) to incomplete multilevel data. We compare the performance of three MDTs with incomplete hierarchical data. We consider the impact of imputation model misspecification on the quality of parameter estimates by employing multiple imputation under assumptions of a normal model (MI/NM) with two-level cross-sectional data when values are missing at random on the dependent variable at rates of 10%, 30%, and 50%. Five criteria are used to compare estimates from MI/NM to estimates from MI assuming a linear mixed model (MI/LMM) and maximum likelihood estimation to the same incomplete data sets. With 10% missing data (MD), techniques performed similarly for fixed-effects estimates, but variance components were biased with MI/NM. Effects of model misspecification worsened at higher rates of MD, with the hierarchical structure of the data markedly underrepresented by biased variance component estimates. MI/LMM and maximum likelihood provided generally accurate and unbiased parameter estimates but performance was negatively affected by increased rates of MD.  相似文献   

6.
The authors describe a method for fitting failure time mixture models that postulate the existence of both susceptibles and long‐term survivors when covariate data are only partially observed. Their method is based on a joint model that combines a Weibull regression model for the susceptibles, a logistic regression model for the probability of being a susceptible, and a general location model for the distribution of the covariates. A Bayesian approach is taken, and Gibbs sampling is used to fit the model to the incomplete data. An application to clinical data on tonsil cancer and a small Monte Carlo study indicate potential large gains in efficiency over standard complete‐case analysis as well as reasonable performance in a variety of situations.  相似文献   

7.
《统计学通讯:理论与方法》2012,41(16-17):3278-3300
Under complex survey sampling, in particular when selection probabilities depend on the response variable (informative sampling), the sample and population distributions are different, possibly resulting in selection bias. This article is concerned with this problem by fitting two statistical models, namely: the variance components model (a two-stage model) and the fixed effects model (a single-stage model) for one-way analysis of variance, under complex survey design, for example, two-stage sampling, stratification, and unequal probability of selection, etc. Classical theory underlying the use of the two-stage model involves simple random sampling for each of the two stages. In such cases the model in the sample, after sample selection, is the same as model for the population; before sample selection. When the selection probabilities are related to the values of the response variable, standard estimates of the population model parameters may be severely biased, leading possibly to false inference. The idea behind the approach is to extract the model holding for the sample data as a function of the model in the population and of the first order inclusion probabilities. And then fit the sample model, using analysis of variance, maximum likelihood, and pseudo maximum likelihood methods of estimation. The main feature of the proposed techniques is related to their behavior in terms of the informativeness parameter. We also show that the use of the population model that ignores the informative sampling design, yields biased model fitting.  相似文献   

8.
Summary.  Factor analysis is a powerful tool to identify the common characteristics among a set of variables that are measured on a continuous scale. In the context of factor analysis for non-continuous-type data, most applications are restricted to item response data only. We extend the factor model to accommodate ranked data. The Monte Carlo expectation–maximization algorithm is used for parameter estimation at which the E-step is implemented via the Gibbs sampler. An analysis based on both complete and incomplete ranked data (e.g. rank the top q out of k items) is considered. Estimation of the factor scores is also discussed. The method proposed is applied to analyse a set of incomplete ranked data that were obtained from a survey that was carried out in GuangZhou, a major city in mainland China, to investigate the factors affecting people's attitude towards choosing jobs.  相似文献   

9.
Bayesian methods are often used to reduce the sample sizes and/or increase the power of clinical trials. The right choice of the prior distribution is a critical step in Bayesian modeling. If the prior not completely specified, historical data may be used to estimate it. In the empirical Bayesian analysis, the resulting prior can be used to produce the posterior distribution. In this paper, we describe a Bayesian Poisson model with a conjugate Gamma prior. The parameters of Gamma distribution are estimated in the empirical Bayesian framework under two estimation schemes. The straightforward numerical search for the maximum likelihood (ML) solution using the marginal negative binomial distribution is unfeasible occasionally. We propose a simplification to the maximization procedure. The Markov Chain Monte Carlo method is used to create a set of Poisson parameters from the historical count data. These Poisson parameters are used to uniquely define the Gamma likelihood function. Easily computable approximation formulae may be used to find the ML estimations for the parameters of gamma distribution. For the sample size calculations, the ML solution is replaced by its upper confidence limit to reflect an incomplete exchangeability of historical trials as opposed to current studies. The exchangeability is measured by the confidence interval for the historical rate of the events. With this prior, the formula for the sample size calculation is completely defined. Published in 2009 by John Wiley & Sons, Ltd.  相似文献   

10.
Data from complex surveys are being used increasingly to build the same sort of explanatory and predictive models as those used in the rest of statistics. Unfortunately the assumptions underlying standard statistical methods are not even approximately valid for most survey data. The problem of parameter estimation has been largely solved, at least for routine data analysis, through the use of weighted estimating equations, and software for most standard analytical procedures is now available in the major statistical packages. One notable omission from standard software is an analogue of the likelihood ratio test. An exception is the Rao–Scott test for loglinear models in contingency tables. In this paper we show how the Rao–Scott test can be extended to handle arbitrary regression models. We illustrate the process of fitting a model to survey data with an example from NHANES.  相似文献   

11.
In this article, a class of conjugate prior for estimating incomplete count data based on a broad class of conjugate prior distributions is presented. The new class of prior distributions arises from a conditional perspective, making use of the conditional specification methodology and can be considered as the generalization of the form of prior distributions that have been used previously in the estimation of incomplete count data well. Finally, some examples of simulated and real data are given.  相似文献   

12.
Abstract

We construct a new bivariate mixture of negative binomial distributions which represents over-dispersed data more efficiently. This is an extension of a univariate mixture of beta and negative binomial distributions. Characteristics of this joint distribution are studied including conditional distributions. Some properties of the correlation coefficient are explored. We demonstrate the applicability of our proposed model by fitting to three real data sets with correlated count data. A comparison is made with some previously used models to show the effectiveness of the new model.  相似文献   

13.
14.
In many occupational safety interventions, the objective is to reduce the injury incidence as well as the mean claims cost once injury has occurred. The claims cost data within a period typically contain a large proportion of zero observations (no claim). The distribution thus comprises a point mass at 0 mixed with a non-degenerate parametric component. Essentially, the likelihood function can be factorized into two orthogonal components. These two components relate respectively to the effect of covariates on the incidence of claims and the magnitude of claims, given that claims are made. Furthermore, the longitudinal nature of the intervention inherently imposes some correlation among the observations. This paper introduces a zero-augmented gamma random effects model for analysing longitudinal data with many zeros. Adopting the generalized linear mixed model (GLMM) approach reduces the original problem to the fitting of two independent GLMMs. The method is applied to evaluate the effectiveness of a workplace risk assessment teams program, trialled within the cleaning services of a Western Australian public hospital.  相似文献   

15.
Experimental design and Taguchi's parameter design are widely employed by industry to optimize the process/product. However, censored data are often observed in product lifetime testing during the experiments. After implementing a repetitious experiment with type II censored data, the censored data are usually estimated by establishing a complex statistical model. However, using the incomplete data to fit a model may not accurately estimates the censored data. Moreover, the model fitting process is complicated for a practitioner who has only limited statistical training. This study proposes a less complex approach to analyze censored data, using the least square estimation method and Torres's analysis of unreplicated factorials with possible abnormalities. This study also presents an effective method to analyze the censored data from Taguchi's parameter design using least square estimation method. Finally, examples are given to illustrate the effectiveness of the proposed methods.  相似文献   

16.
We present a novel model, which is a two-parameter extension of the Poisson distribution. Its normalizing constant is related to the Touchard polynomials, hence the name of this model. It is a flexible distribution that can account for both under- or overdispersion and concentration of zeros that are frequently found in non-Poisson count data. In contrast to some other generalizations, the Hessian matrix for maximum likelihood estimation of the Touchard parameters has a simple form. We exemplify with three data sets, showing that our suggested model is a competitive candidate for fitting non-Poisson counts.  相似文献   

17.
We develop a distribution supported on a bounded interval with a probability density function that is constructed from any finite number of linear segments. With an increasing number of segments, the distribution can approach any continuous density function of arbitrary form. The flexibility of the distribution makes it a useful tool for various modeling purposes. We further demonstrate that it is capable of fitting data with considerable precision—outperforming distributions recommended by previous studies. We suggest that this distribution is particularly effective in fitting data with sufficient observations that are skewed and multimodal.  相似文献   

18.
In a multinomial model, the sample space is partitioned into a disjoint union of cells. The partition is usually immutable during sampling of the cell counts. In this paper, we extend the multinomial model to the incomplete multinomial model by relaxing the constant partition assumption to allow the cells to be variable and the counts collected from non-disjoint cells to be modeled in an integrated manner for inference on the common underlying probability. The incomplete multinomial likelihood is parameterized by the complete-cell probabilities from the most refined partition. Its sufficient statistics include the variable-cell formation observed as an indicator matrix and all cell counts. With externally imposed structures on the cell formation process, it reduces to special models including the Bradley–Terry model, the Plackett–Luce model, etc. Since the conventional method, which solves for the zeros of the score functions, is unfruitful, we develop a new approach to establishing a simpler set of estimating equations to obtain the maximum likelihood estimate (MLE), which seeks the simultaneous maximization of all multiplicative components of the likelihood by fitting each component into an inequality. As a consequence, our estimation amounts to solving a system of the equality attainment conditions to the inequalities. The resultant MLE equations are simple and immediately invite a fixed-point iteration algorithm for solution, which is referred to as the weaver algorithm. The weaver algorithm is short and amenable to parallel implementation. We also derive the asymptotic covariance of the MLE, verify main results with simulations, and compare the weaver algorithm with an MM/EM algorithm based on fitting a Plackett–Luce model to a benchmark data set.  相似文献   

19.
In this paper, we study some mathematical properties of the beta Weibull (BW) distribution, which is a quite flexible model in analysing positive data. It contains the Weibull, exponentiated exponential, exponentiated Weibull and beta exponential distributions as special sub-models. We demonstrate that the BW density can be expressed as a mixture of Weibull densities. We provide their moments and two closed-form expressions for their moment-generating function. We examine the asymptotic distributions of the extreme values. Explicit expressions are derived for the mean deviations, Bonferroni and Lorenz curves, reliability and two entropies. The density of the BW-order statistics is a mixture of Weibull densities and two closed-form expressions are derived for their moments. The estimation of the parameters is approached by two methods: moments and maximum likelihood. We compare the performances of the estimates obtained from both the methods by simulation. The expected information matrix is derived. For the first time, we introduce a log-BW regression model to analyse censored data. The usefulness of the BW distribution is illustrated in the analysis of three real data sets.  相似文献   

20.
We introduce a three-parameter extension of the exponential distribution which contains as sub-models the exponential, logistic-exponential and Marshall-Olkin exponential distributions. The new model is very flexible and its associated density function can be decreasing or unimodal. Further, it can produce all of the four major shapes of the hazard rate, that is, increasing, decreasing, bathtub and upside-down bathtub. Given that closed-form expressions are available for the survival and hazard rate functions, the new distribution is quite tractable. It can be used to analyze various types of observations including censored data. Computable representations of the quantile function, ordinary and incomplete moments, generating function and probability density function of order statistics are obtained. The maximum likelihood method is utilized to estimate the model parameters. A simulation study is carried out to assess the performance of the maximum likelihood estimators. Two actual data sets are used to illustrate the applicability of the proposed model.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号