首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
The generalized Poisson distribution (GPD), studied by many researchers and containing two parameters θ and λ, has been found to fit very well data sets arising in biological, ecological, social and marketing fields. Consul and Shoukri (1985) have shown that for negative values of λ the GPD gets truncated and the model becomes deficient; however, the truncation error becomes less than 0.0005 if the minimum number of non-zero probability classes ≥ 4 for all values of θ and λ and the GPD model can be safely used in all such cases. The problem of admissible maximum likelihood (ML) estimation when the sample mean is larger than the sample variance is considered in this paper which complements the earlier work of Consul and Shoukri (1984) on the existence of unique ML estimators of θ and λ when the sample mean is smaller than or equal to the sample variance.  相似文献   

2.
Generalized Pareto distribution (GPD) has been widely used to model exceedances over thresholds. In this article we propose a new method called weighted nonlinear least squares (WNLS) to estimate the parameters of the GPD. The WNLS estimators always exist and are simple to compute. Some asymptotic results of the proposed method are provided. The simulation results indicate that the proposed method performs well compared to existing methods in terms of mean squared error and bias. Its advantages are further illustrated through the analysis of two real data sets.  相似文献   

3.
A discrete probability model always gets truncated during the sampling process and the point of truncation depends upon the sample size. Also, the generalized Poisson distribution cannot be used with full justification when the second parameter is negative. To avoid these problems a truncated generalized Poisson distribution is defined and studied. Estimation of its parameters by moments method, maximum likelihood method and a mixed method are considered. Some examples are given to illustrate the effect on the parameters’ estimates when a non-truncated GPD is used instead of a truncated GPD.  相似文献   

4.
In this article, we propose semiparametric methods to estimate the cumulative incidence function of two dependent competing risks for left-truncated and right-censored data. The proposed method is based on work by Huang and Wang (1995). We extend previous model by allowing for a general parametric truncation distribution and a third competing risk before recruitment. Based on work by Vardi (1989), several iterative algorithms are proposed to obtain the semiparametric estimates of cumulative incidence functions. The asymptotic properties of the semiparametric estimators are derived. Simulation results show that a semiparametric approach assuming the parametric truncation distribution is correctly specified produces estimates with smaller mean squared error than those obtained in a fully nonparametric model.  相似文献   

5.
ABSTRACT

In this article, we derive exact explicit expressions for the single, double, triple, and quadruple moments of order statistics from the generalized Pareto distribution (GPD). Also, we obtain the best linear unbiased estimates of the location and scale parameters (BLUE's) of the GPD. We then use these results to determine the mean, variance, and coefficients of skewness and kurtosis of certain linear functions of order statistics. These are then utilized to develop approximate confidence intervals for the generalized Pareto parameters using Edgeworth approximation and compare them with those based on Monte Carlo simulations. To show the usefulness of our results, we also present a numerical example. Finally, we give an application to real data.  相似文献   

6.
Parameter estimation of the generalized Pareto distribution—Part II   总被引:1,自引:0,他引:1  
This is the second part of a paper which focuses on reviewing methods for estimating the parameters of the generalized Pareto distribution (GPD). The GPD is a very important distribution in the extreme value context. It is commonly used for modeling the observations that exceed very high thresholds. The ultimate success of the GPD in applications evidently depends on the parameter estimation process. Quite a few methods exist in the literature for estimating the GPD parameters. Estimation procedures, such as the maximum likelihood (ML), the method of moments (MOM) and the probability weighted moments (PWM) method were described in Part I of the paper. We shall continue to review methods for estimating the GPD parameters, in particular methods that are robust and procedures that use the Bayesian methodology. As in Part I, we shall focus on those that are relatively simple and straightforward to be applied to real world data.  相似文献   

7.
The generalized Pareto distribution (GPD) has been widely used in the extreme value framework. The success of the GPD when applied to real data sets depends substantially on the parameter estimation process. Several methods exist in the literature for estimating the GPD parameters. Mostly, the estimation is performed by maximum likelihood (ML). Alternatively, the probability weighted moments (PWM) and the method of moments (MOM) are often used, especially when the sample sizes are small. Although these three approaches are the most common and quite useful in many situations, their extensive use is also due to the lack of knowledge about other estimation methods. Actually, many other methods, besides the ones mentioned above, exist in the extreme value and hydrological literatures and as such are not widely known to practitioners in other areas. This paper is the first one of two papers that aim to fill in this gap. We shall extensively review some of the methods used for estimating the GPD parameters, focusing on those that can be applied in practical situations in a quite simple and straightforward manner.  相似文献   

8.
A common approach to modelling extreme data are to consider the distribution of the exceedance value over a high threshold. This approach is based on the distribution of excess, which follows the generalized Pareto distribution (GPD) and has shown to be adequate for this type of situation. As with all data involving analysis in time, excesses above a threshold may also vary and suffer from the influence of covariates. Thus, the GPD distribution can be modelled by entering the presence of these factors. This paper presents a new model for extreme values, where GPD parameters are written on the basis of a dynamic regression model. The estimation of the model parameters is made under the Bayesian paradigm, with sampling points via MCMC. As with environmental data, behaviour data are related to other factors such as time and covariates such as latitude and distance from the sea. Simulation studies have shown the efficiency and identifiability of the model, and applying real rain data from the state of Piaui, Brazil, shows the advantage in predicting and interpreting the model against other similar models proposed in the literature.  相似文献   

9.
Experiments in which very few units are measured many times sometimes present particular difficulties. Interest often centers on simple location shifts between two treatment groups, but appropriate modeling of the error distribution can be challenging. For example, normality may be difficult to verify, or a single transformation stabilizing variance or improving normality for all units and all measurements may not exist. We propose an analysis of two sample repeated measures data based on the permutation distribution of units. This provides a distribution free alternative to standard analyses. The analysis includes testing, estimation and confidence intervals. By assuming a certain structure in the location shift model, the dimension of the problem is reduced by analyzing linear combinations of the marginal statistics. Recently proposed algorithms for computation of two sample permutation distributions, require only a few seconds for experiments having as many as 100 units and any number of repeated measures. The test has high asymptotic efficiency and good power with respect to tests based on the normal distribution. Since the computational burden is minimal, approximation of the permutation distribution is unnecessary.  相似文献   

10.
In a multivariate mean–variance model, the class of linear score (LS) estimators based on an unbiased linear estimating function is introduced. A special member of this class is the (extended) quasi-score (QS) estimator. It is ‘extended’ in the sense that it comprises the parameters describing the distribution of the regressor variables. It is shown that QS is (asymptotically) most efficient within the class of LS estimators. An application is the multivariate measurement error model, where the parameters describing the regressor distribution are nuisance parameters. A special case is the zero-inflated Poisson model with measurement errors, which can be treated within this framework.  相似文献   

11.
Using some logarithmic and integral transformation we transform a continuous covariate frailty model into a polynomial regression model with a random effect. The responses of this mixed model can be ‘estimated’ via conditional hazard function estimation. The random error in this model does not have zero mean and its variance is not constant along the covariate and, consequently, these two quantities have to be estimated. Since the asymptotic expression for the bias is complicated, the two-large-bandwidth trick is proposed to estimate the bias. The proposed transformation is very useful for clustered incomplete data subject to left truncation and right censoring (and for complex clustered data in general). Indeed, in this case no standard software is available to fit the frailty model, whereas for the transformed model standard software for mixed models can be used for estimating the unknown parameters in the original frailty model. A small simulation study illustrates the good behavior of the proposed method. This method is applied to a bladder cancer data set.  相似文献   

12.
Extreme quantile estimation plays an important role in risk management and environmental statistics among other applications. A popular method is the peaks-over-threshold (POT) model that approximate the distribution of excesses over a high threshold through generalized Pareto distribution (GPD). Motivated by a practical financial risk management problem, we look for an appropriate prior choice for Bayesian estimation of the GPD parameters that results in better quantile estimation. Specifically, we propose a noninformative matching prior for the parameters of a GPD so that a specific quantile of the Bayesian predictive distribution matches the true quantile in the sense of Datta et al. (2000).  相似文献   

13.
Bayesian D‐optimal designs supported on a fixed number of points were found by Dette & Wong (1998) for estimating parameters in a polynomial model when the error variance depends exponentially on the explanatory variable. The present authors provide optimal designs under a broader class of error variance structures and investigate the robustness properties of these designs to model and prior distribution assumptions. A comparison of the performance of the optimal designs relative to the popular uniform designs is also given. The authors' results suggest that Bayesian D‐optimal designs suported on a fixed number of points are more likely to be globaly optimal among all designs if the prior distribution is symmetric and is concentrated around its mean.  相似文献   

14.
Zero inflated Poisson regression is a model commonly used to analyze data with excessive zeros. Although many models have been developed to fit zero-inflated data, most of them strongly depend on the special features of the individual data. For example, there is a need for new models when dealing with truncated and inflated data. In this paper, we propose a new model that is sufficiently flexible to model inflation and truncation simultaneously, and which is a mixture of a multinomial logistic and a truncated Poisson regression, in which the multinomial logistic component models the occurrence of excessive counts. The truncated Poisson regression models the counts that are assumed to follow a truncated Poisson distribution. The performance of our proposed model is evaluated through simulation studies, and our model is found to have the smallest mean absolute error and best model fit. In the empirical example, the data are truncated with inflated values of zero and fourteen, and the results show that our model has a better fit than the other competing models.  相似文献   

15.
ABSTRACT

The Lindley distribution is an important distribution for analysing the stress–strength reliability models and lifetime data. In many ways, the Lindley distribution is a better model than that based on the exponential distribution. Order statistics arise naturally in many of such applications. In this paper, we derive the exact explicit expressions for the single, double (product), triple and quadruple moments of order statistics from the Lindley distribution. Then, we use these moments to obtain the best linear unbiased estimates (BLUEs) of the location and scale parameters based on Type-II right-censored samples. Next, we use these results to determine the mean, variance, and coefficients of skewness and kurtosis of some certain linear functions of order statistics to develop Edgeworth approximate confidence intervals of the location and scale Lindley parameters. In addition, we carry out some numerical illustrations through Monte Carlo simulations to show the usefulness of the findings. Finally, we apply the findings of the paper to some real data set.  相似文献   

16.
Penalized likelihood estimators for truncated data   总被引:1,自引:0,他引:1  
We investigate the performance of linearly penalized likelihood estimators for estimating distributional parameters in the presence of data truncation. Truncation distorts the likelihood surface to create instabilities and high variance in the estimation of these parameters, and the penalty terms help in many cases to decrease estimation error and increase robustness. Approximate methods are provided for choosing a priori good penalty estimators, which are shown to perform well in a series of simulation experiments. The robustness of the methods is explored heuristically using both simulated and real data drawn from an operational risk context.  相似文献   

17.
The independent additive errors linear model consists of a structure for the mean and a separate structure for the error distribution. The error structure may be parametric or it may be semiparametric. Under alternative values of the mean structure, the best fitting additive errors model has an error distribution which can be represented as the convolution of the actual error distribution and the marginal distribution of a misspecification term. The model misspecification term results from the covariates' distribution. Conditions are developed to distinguish when the semiparametric model yields sharper inference than the parametric model and vice versa. The main conditions concern the actual error distribution and the covariates' distribution. The theoretical results explain a paradoxical finding in semiparametric Bayesian modelling, where the posterior distribution under a semiparametric model is found to be more concentrated than is the posterior distribution under a corresponding parametric model. The paradox is illustrated on a set of allometric data. The Canadian Journal of Statistics 39: 165–180; 2011 ©2011 Statistical Society of Canada  相似文献   

18.
Our article presents a general treatment of the linear regression model, in which the error distribution is modelled nonparametrically and the error variances may be heteroscedastic, thus eliminating the need to transform the dependent variable in many data sets. The mean and variance components of the model may be either parametric or nonparametric, with parsimony achieved through variable selection and model averaging. A Bayesian approach is used for inference with priors that are data-based so that estimation can be carried out automatically with minimal input by the user. A Dirichlet process mixture prior is used to model the error distribution nonparametrically; when there are no regressors in the model, the method reduces to Bayesian density estimation, and we show that in this case the estimator compares favourably with a well-regarded plug-in density estimator. We also consider a method for checking the fit of the full model. The methodology is applied to a number of simulated and real examples and is shown to work well.  相似文献   

19.

Conventionally, it was shown that the underlying distribution is normal if and only if the sample mean and sample variance from a random sample are independent. This paper focusses on the normal population characterization theorem by showing that, if the joint distribution of a skew normal sample follows certain multivariate skew normal distribution, the sample mean and sample variance are still independent.  相似文献   

20.

This paper is concerned with properties (bias, standard deviation, mean square error and efficiency) of twenty six estimators of the intraclass correlation in the analysis of binary data. Our main interest is to study these properties when data are generated from different distributions. For data generation we considered three over-dispersed binomial distributions, namely, the beta-binomial distribution, the probit normal binomial distribution and a mixture of two binomial distributions. The findings regarding bias, standard deviation and mean squared error of all these estimators, are that (a) in general, the distributions of biases of most of the estimators are negatively skewed. The biases are smallest when data are generated from the beta-binomial distribution and largest when data are generated from the mixture distribution; (b) the standard deviations are smallest when data are generated from the beta-binomial distribution; and (c) the mean squared errors are smallest when data are generated from the beta-binomial distribution and largest when data are generated from the mixture distribution. Of the 26, nine estimators including the maximum likelihood estimator, an estimator based on the optimal quadratic estimating equations of Crowder (1987), and an analysis of variance type estimator is found to have least amount of bias, standard deviation and mean squared error. Also, the distributions of the bias, standard deviation and mean squared error for each of these estimators are, in general, more symmetric than those of the other estimators. Our findings regarding efficiency are that the estimator based on the optimal quadratic estimating equations has consistently high efficiency and least variability in the efficiency results. In the important range in which the intraclass correlation is small (≤0 5), on the average, this estimator shows best efficiency performance. The analysis of variance type estimator seems to do well for larger values of the intraclass correlation. In general, the estimator based on the optimal quadratic estimating equations seems to show best efficiency performance for data from the beta-binomial distribution and the probit normal binomial distribution, and the analysis of variance type estimator seems to do well for data from the mixture distribution.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号