首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 593 毫秒
1.
The present paper focuses attention on the sensitivity of technical inefficiency to most commonly used one-sided distributions of the inefficiency error term, namely the truncated normal, the half-normal, and the exponential distributions. A generalized version of the half-normal, which does not embody the zero-mean restriction, is also explored. For each distribution, the likelihood function and the counterpart of the estimator of technical efficiency are explicitly stated (Jondrow, J., Lovell, C. A. K., Materov, I. S., Schmidt, P. ([1982]), On estimation of technical inefficiency in the stochastic frontier production function model, J. Econometrics19:233-238). Based on our panel data set, related to Tunisian manufacturing firms over the period 1983-1993, formal tests lead to a strong rejection of the zero-mean restriction embodied in the half normal distribution. Our main conclusion is that the degree of measured inefficiency is very sensitive to the postulated assumptions about the distribution of the one-sided error term. The estimated inefficiency indices are, however, unaffected by the choice of the functional form for the production function.  相似文献   

2.
This article describes a generalization of the binomial distribution. The closed form probability function for the probability of k successes out of n correlated, exchangeable Bernoulli trials depends on the number of trials and its two parameters: the common success probability and the common correlation. The distribution is derived under the assumption that the common correlation between all pairs of Bernoulli trials remains unchanged conditional on successes in all completed trials. The distribution was developed to model bond defaults but may be suited to biostatistical applications involving clusters of binary data encountered in repeated measurements or toxicity studies of families of organisms. Maximum likelihood estimates for the parameters of the distribution are found for a set of binary data from a developmental toxicity study on litters of mice.  相似文献   

3.
Typical panel data models make use of the assumption that the regression parameters are the same for each individual cross-sectional unit. We propose tests for slope heterogeneity in panel data models. Our tests are based on the conditional Gaussian likelihood function in order to avoid the incidental parameters problem induced by the inclusion of individual fixed effects for each cross-sectional unit. We derive the Conditional Lagrange Multiplier test that is valid in cases where N → ∞ and T is fixed. The test applies to both balanced and unbalanced panels. We expand the test to account for general heteroskedasticity where each cross-sectional unit has its own form of heteroskedasticity. The modification is possible if T is large enough to estimate regression coefficients for each cross-sectional unit by using the MINQUE unbiased estimator for regression variances under heteroskedasticity. All versions of the test have a standard Normal distribution under general assumptions on the error distribution as N → ∞. A Monte Carlo experiment shows that the test has very good size properties under all specifications considered, including heteroskedastic errors. In addition, power of our test is very good relative to existing tests, particularly when T is not large.  相似文献   

4.
Parameters of a finite mixture model are often estimated by the expectation–maximization (EM) algorithm where the observed data log-likelihood function is maximized. This paper proposes an alternative approach for fitting finite mixture models. Our method, called the iterative Monte Carlo classification (IMCC), is also an iterative fitting procedure. Within each iteration, it first estimates the membership probabilities for each data point, namely the conditional probability of a data point belonging to a particular mixing component given that the data point value is obtained, it then classifies each data point into a component distribution using the estimated conditional probabilities and the Monte Carlo method. It finally updates the parameters of each component distribution based on the classified data. Simulation studies were conducted to compare IMCC with some other algorithms for fitting mixture normal, and mixture t, densities.  相似文献   

5.
Model-based clustering of Gaussian copulas for mixed data   总被引:1,自引:0,他引:1  
Clustering of mixed data is important yet challenging due to a shortage of conventional distributions for such data. In this article, we propose a mixture model of Gaussian copulas for clustering mixed data. Indeed copulas, and Gaussian copulas in particular, are powerful tools for easily modeling the distribution of multivariate variables. This model clusters data sets with continuous, integer, and ordinal variables (all having a cumulative distribution function) by considering the intra-component dependencies in a similar way to the Gaussian mixture. Indeed, each component of the Gaussian copula mixture produces a correlation coefficient for each pair of variables and its univariate margins follow standard distributions (Gaussian, Poisson, and ordered multinomial) depending on the nature of the variable (continuous, integer, or ordinal). As an interesting by-product, this model generalizes many well-known approaches and provides tools for visualization based on its parameters. The Bayesian inference is achieved with a Metropolis-within-Gibbs sampler. The numerical experiments, on simulated and real data, illustrate the benefits of the proposed model: flexible and meaningful parameterization combined with visualization features.  相似文献   

6.
A large number of models have been derived from the two-parameter Weibull distribution including the inverse Weibull (IW) model which is found suitable for modeling the complex failure data set. In this paper, we present the Bayesian inference for the mixture of two IW models. For this purpose, the Bayes estimates of the parameters of the mixture model along with their posterior risks using informative as well as the non-informative prior are obtained. These estimates have been attained considering two cases: (a) when the shape parameter is known and (b) when all parameters are unknown. For the former case, Bayes estimates are obtained under three loss functions while for the latter case only the squared error loss function is used. Simulation study is carried out in order to explore numerical aspects of the proposed Bayes estimators. A real-life data set is also presented for both cases, and parameters obtained under case when shape parameter is known are tested through testing of hypothesis procedure.  相似文献   

7.
In this paper, a multivariate form of truncated generalized Cauchy distribution (TGCD), which is denoted by (MVTGCD), is introduced. The joint density function, conditional density function, moment generating function and mixed moments of order ${b=\sum_{i=1}^{k}b_{i}}$ are obtained. Making use of the mixed moments formula, skewness and kurtosis in case of the bivariate case are obtained. Also, all parameters of the distribution are estimated using the maximum likelihood and Bayes methods. A real data set is introduced and analyzed using three models. The first model is the bivariate Cauchy distribution, the second is the truncated bivariate Cauchy distribution and the third is the bivariate truncated generalized Cauchy distribution. A comparison is carried out between the mentioned models based on the corresponding Kolmogorov–Smirnov (K–S) test statistic to emphasize that the bivariate truncated generalized Cauchy model fits the data better than the other models.  相似文献   

8.
When available data comprise a number of sampled households in each of a number of income classes, the likelihood function is obtained from a multinomial distribution with the income class population proportions as the unknown parameters. Two methods for going from this likelihood function to a posterior distribution on the Gini coefficient are investigated. In the first method, two alternative assumptions about the underlying income distribution are considered, namely a lognormal distribution and the Singh–Maddala (1976) income distribution. In these cases the likelihood function is reparameterized and the Gini coefficient is a nonlinear function of the income distribution parameters. The Metropolis algorithm is used to find the corresponding posterior distributions of the Gini coefficient from a sample of Bangkok households. The second method does not require an assumption about the nature of the income distribution, but uses (a) triangular prior distributions, and (b) beta prior distributions, on the location of mean income within each income class. By sampling from these distributions, and the Dirichlet posterior distribution of the income class proportions, alternative posterior distributions of the Gini coefficient are calculated.  相似文献   

9.
Abstract.  The likelihood ratio statistic for testing pointwise hypotheses about the survival time distribution in the current status model can be inverted to yield confidence intervals (CIs). One advantage of this procedure is that CIs can be formed without estimating the unknown parameters that figure in the asymptotic distribution of the maximum likelihood estimator (MLE) of the distribution function. We discuss the likelihood ratio-based CIs for the distribution function and the quantile function and compare these intervals to several different intervals based on the MLE. The quantiles of the limiting distribution of the MLE are estimated using various methods including parametric fitting, kernel smoothing and subsampling techniques. Comparisons are carried out both for simulated data and on a data set involving time to immunization against rubella. The comparisons indicate that the likelihood ratio-based intervals are preferable from several perspectives.  相似文献   

10.
In this paper, point and interval estimations for the parameters of the exponentiated exponential (EE) distribution are studied based on progressive first-failure-censored data. The Bayes estimates are computed based on squared error and Linex loss functions and using Markov Chain Monte Carlo (MCMC) algorithm. Also, based on this censoring scheme, approximate confidence intervals for the parameters of EE distribution are developed. Monte Carlo simulation study is carried out to compare the performances of the different methods by computing the estimated risks (ERs), as well as Akaike's information criteria (AIC) and Bayesian information criteria (BIC) of the estimates. Finally, a real data set is introduced and analyzed using EE and Weibull distributions. A comparison is carried out between the mentioned models based on the corresponding Kolmogorov–Smirnov (K–S) test statistic to emphasize that the EE model fits the data with the same efficiency as the other model. Point and interval estimation of all parameters are studied based on this real data set as illustrative example.  相似文献   

11.
Abstract

In this paper we introduce a new two-parameter discrete distribution which may be useful for modeling count data. Additionally, the probability mass function is very simple and it may have a zero vertex. We show that the new discrete distribution is a particular solution of a multiple Poisson process, and that it is infinitely divisible. Additionally, various structural properties of the new discrete distribution are derived. We also discuss two methods (moments and maximum likelihood) to estimate the model parameters. The usefulness of the proposed distribution is illustrated by means of real data sets to prove its versatility in practical applications.  相似文献   

12.
An imputation procedure is a procedure by which each missing value in a data set is replaced (imputed) by an observed value using a predetermined resampling procedure. The distribution of a statistic computed from a data set consisting of observed and imputed values, called a completed data set, is affecwd by the imputation procedure used. In a Monte Carlo experiment, three imputation procedures are compared with respect to the empirical behavior of the goodness-of- fit chi-square statistic computed from a completed data set. The results show that each imputation procedure affects the distribution of the goodness-of-fit chi-square statistic in 3. different manner. However, when the empirical behavior of the goodness-of-fit chi-square statistic is compared u, its appropriate asymptotic distribution, there are no substantial differences between these imputation procedures.  相似文献   

13.
A general nonparametric imputation procedure, based on kernel regression, is proposed to estimate points as well as set- and function-indexed parameters when the data are missing at random (MAR). The proposed method works by imputing a specific function of a missing value (and not the missing value itself), where the form of this specific function is dictated by the parameter of interest. Both single and multiple imputations are considered. The associated empirical processes provide the right tool to study the uniform convergence properties of the resulting estimators. Our estimators include, as special cases, the imputation estimator of the mean, the estimator of the distribution function proposed by Cheng and Chu [1996. Kernel estimation of distribution functions and quantiles with missing data. Statist. Sinica 6, 63–78], imputation estimators of a marginal density, and imputation estimators of regression functions.  相似文献   

14.
Shapes of service-time distributions in queueing network models have a great impact on the distribution of system response-times. It is essential for the analysis of response-time distribution that the modeled service-time distributions have the correct shape. Tradionally modeling of service-time distributions is based on a parametric approach by assuming a specific distribution and estimating its parameters. We introduce an alternative approach based on the principles of exploratory data analysis and nonparametric data modeling. The proposed method applies nonlinear data transformation and resistant curve fitting. The method can be used in cases, where the available data is a complete sample, a histogram, or the mean and a set of 5-10 quantiles. The reported results indicate that the proposed method is able to approximate the distribution of measured service times so that accurate estimates for quantiles of the response-time distribution are obtained  相似文献   

15.
In this paper, the asymptotic distribution of the signal-to-noise ratio (SNR) is derived and a new confidence interval for the SNR is introduced. An evaluation of the performance of the new interval compared to Sharma and Krishna (S–K) (1994) confidence interval for the SNR using Monte Carlo simulations is conducted. Data were randomly generated from normal, log-normal, χ2, Gamma, and Weibull distributions. Simulations revealed that the performance of S–K interval is totally dependent on the amount of noise introduced and that it has a constant width for a given sample size. The S–K interval performs poorly in four of the distributions unless the SNR is around one. It is recommended against using the S–K interval for data from log-normal distribution even with SNR = 1. Unlike the S–K interval which does not account for skewness and kurtosis of the distribution, the new confidence interval for the SNR outperforms S–K for all five distributions discussed, especially when SNR?? 2. The proposed ranked set sampling (RSS) instead of simple random sampling (SRS) has improved the performance of both intervals as measured by coverage probability.  相似文献   

16.
In applications of survival analysis, the failure rate function may frequently present a unimodal shape. In such cases, the log-normal and log-logistic distributions are used. In this paper, we shall be concerned only with parametric forms, so a location-scale regression model based on the odd log-logistic Weibull distribution is proposed for modelling data with a decreasing, increasing, unimodal and bathtub failure rate function as an alternative to the log-Weibull regression model. For censored data, we consider a classic method to estimate the parameters of the proposed model. We derive the appropriate matrices for assessing local influences on the parameter estimates under different perturbation schemes and present some ways to assess global influences. Further, for different parameter settings, sample sizes and censoring percentages, various simulations are performed. In addition, the empirical distribution of some modified residuals is determined and compared with the standard normal distribution. These studies suggest that the residual analysis usually performed in normal linear regression models can be extended to a modified deviance residual in the new regression model applied to censored data. We analyse a real data set using the log-odd log-logistic Weibull regression model.  相似文献   

17.
This article proposes the maximum likelihood estimates based on bare bones particle swarm optimization (BBPSO) algorithm for estimating the parameters of Weibull distribution with censored data, which is widely used in lifetime data analysis. This approach can produce more accuracy of the parameter estimation for the Weibull distribution. Additionally, the confidence intervals for the estimators are obtained. The simulation results show that the BB PSO algorithm outperforms the Newton–Raphson method in most cases in terms of bias, root mean square of errors, and coverage rate. Two examples are used to demonstrate the performance of the proposed approach. The results show that the maximum likelihood estimates via BBPSO algorithm perform well for estimating the Weibull parameters with censored data.  相似文献   

18.
In this article, a new three-parameter extension of the two-parameter log-logistic distribution is introduced. Several distributional properties such as moment-generating function, quantile function, mean residual lifetime, the Renyi and Shanon entropies, and order statistics are considered. The estimation of the model parameters for complete and right-censored cases is investigated competently by maximum likelihood estimation (MLE). A simulation study is conducted to show that these MLEs are consistent in moderate samples. Two real datasets are considered; one is a right-censored data to show that the proposed model has a superior performance over several existing popular models.  相似文献   

19.
Single cohort stage‐frequency data are considered when assessing the stage reached by individuals through destructive sampling. For this type of data, when all hazard rates are assumed constant and equal, Laplace transform methods have been applied in the past to estimate the parameters in each stage‐duration distribution and the overall hazard rates. If hazard rates are not all equal, estimating stage‐duration parameters using Laplace transform methods becomes complex. In this paper, two new models are proposed to estimate stage‐dependent maturation parameters using Laplace transform methods where non‐trivial hazard rates apply. The first model encompasses hazard rates that are constant within each stage but vary between stages. The second model encompasses time‐dependent hazard rates within stages. Moreover, this paper introduces a method for estimating the hazard rate in each stage for the stage‐wise constant hazard rates model. This work presents methods that could be used in specific types of laboratory studies, but the main motivation is to explore the relationships between stage maturation parameters that, in future work, could be exploited in applying Bayesian approaches. The application of the methodology in each model is evaluated using simulated data in order to illustrate the structure of these models.  相似文献   

20.
We propose a novel alternative to case-control sampling for the estimation of individual-level risk in spatial epidemiology. Our approach uses weighted estimating equations to estimate regression parameters in the intensity function of an inhomogeneous spatial point process, when information on risk-factors is available at the individual level for cases, but only at a spatially aggregated level for the population at risk. We develop data-driven methods to select the weights used in the estimating equations and show through simulation that the choice of weights can have a major impact on efficiency of estimation. We develop a formal test to detect non-Poisson behavior in the underlying point process and assess the performance of the test using simulations of Poisson and Poisson cluster point processes. We apply our methods to data on the spatial distribution of childhood meningococcal disease cases in Merseyside, U.K. between 1981 and 2007.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号