首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 0 毫秒
1.
We develop a ‘robust’ statistic T2 R, based on Tiku's (1967, 1980) MML (modified maximum likelihood) estimators of location and scale parameters, for testing an assumed meam vector of a symmetric multivariate distribution. We show that T2 R is one the whole considerably more powerful than the prominenet Hotelling T2 statistics. We also develop a robust statistic T2 D for testing that two multivariate distributions (skew or symmetric) are identical; T2 D seems to be usually more powerful than nonparametric statistics. The only assumption we make is that the marginal distributions are of the type (1/σk)f((x-μk)/σk) and the means and variances of these marginal distributions exist.  相似文献   

2.
Grouped data are frequently used in several fields of study. In this work, we use the expectation-maximization (EM) algorithm for fitting the skew-normal (SN) mixture model to the grouped data. Implementing the EM algorithm requires computing the one-dimensional integrals for each group or class. Our simulation study and real data analyses reveal that the EM algorithm not only always converges but also can be implemented in just a few seconds even when the number of components is large, contrary to the Bayesian paradigm that is computationally expensive. The accuracy of the EM algorithm and superiority of the SN mixture model over the traditional normal mixture model in modelling grouped data are demonstrated through the simulation and three real data illustrations. For implementing the EM algorithm, we use the package called ForestFit developed for R environment available at https://cran.r-project.org/web/packages/ForestFit/index.html.  相似文献   

3.
The effect of skewness on hypothesis tests for the existence of a mixture of univariate and bivariate normal distributions is examined through a Monte Carlo study. A likelihood ratio test based on results of the simultaneous estimation of skewness parameters, derived from power transformations, with mixture parameters is proposed. This procedure detects the difference between inherent distributional skewness and the apparent skewness which is a manifestation of the mixture of several distributions. The properties of this test are explored through a simulation study.  相似文献   

4.
5.
ABSTRACT

In this article, we introduce the Gompertz power series (GPS) class of distributions which is obtained by compounding Gompertz and power series distributions. This distribution contains several lifetime models such as Gompertz-geometric (GG), Gompertz-Poisson (GP), Gompertz-binomial (GB), and Gompertz-logarithmic (GL) distributions as special cases. Sub-models of the GPS distribution are studied in details. The hazard rate function of the GPS distribution can be increasing, decreasing, and bathtub-shaped. We obtain several properties of the GPS distribution such as its probability density function, and failure rate function, Shannon entropy, mean residual life function, quantiles, and moments. The maximum likelihood estimation procedure via a EM-algorithm is presented, and simulation studies are performed for evaluation of this estimation for complete data, and the MLE of parameters for censored data. At the end, a real example is given.  相似文献   

6.
7.
8.
ABSTRACT

Asymmetric models have been discussed quite extensively in recent years, in situations where the normality assumption is suspected due to lack of symmetry in the data. Techniques for assessing the quality of fit and diagnostic analysis are important for model validation. This paper presents a study of the mean-shift method for the detection of outliers in regression models under skew scale-mixtures of normal distributions. Analytical solutions for the estimators of the parameters are obtained through the use of Expectation–Maximization algorithm. The observed information matrix for the calculation of standard errors is obtained for each distribution. Simulation studies and an application to the analysis of a data have been carried out, showing the efficiency of the proposed method in detecting outliers.  相似文献   

9.
We consider the testing problems of the structural parameters for the multivariate linear functional relationship model. We treat the likelihood ratio test statistics and the test statistics based on the asymptotic distributions of the maximum likelihood estimators. We derive their asymptotic distributions under each null hypothesis respectively. A simulation study is made to evaluate how we can trust our asymptotic results when the sample size is rather small.  相似文献   

10.
We present the maximum likelihood estimation (MLE) via particle swarm optimization (PSO) algorithm to estimate the mixture of two Weibull parameters with complete and multiple censored data. A simulation study is conducted to assess the performance of the MLE via PSO algorithm, quasi-Newton method and expectation-maximization (EM) algorithm for different parameter settings and sample sizes in both uncensored and censored cases. The simulation results showed that the PSO algorithm outperforms the quasi-Newton method and the EM algorithm in most cases regarding bias and root mean square errors. Two numerical examples are used to demonstrate the performance of our proposed method.  相似文献   

11.
It is well known that the log-likelihood function for samples coming from normal mixture distributions may present spurious maxima and singularities. For this reason here we reformulate some Hathaways results and we propose two constrained estimation procedures for multivariate normal mixture modelling according to the likelihood approach. Their perfomances are illustrated on the grounds of some numerical simulations based on the EM algorithm. A comparison between multivariate normal mixtures and the hot-deck approach in missing data imputation is also considered.Salvatore Ingrassia: S. Ingrassia carried out the research as part of the project Metodi Statistici e Reti Neuronali per lAnalisi di Dati Complessi (PRIN 2000, resp. G. Lunetta).  相似文献   

12.
A faster alternative to the EM algorithm in finite mixture distributions is described, which alternates EM iterations with Gauss-Newton iterations using the observed information matrix. At the expense of modest additional analytical effort in obtaining the observed information, the hybrid algorithm reduces the computing time required and provides asymptotic standard errors at convergence. The algorithm is illustrated on the two-component normal mixture.  相似文献   

13.
We consider the problem of full information maximum likelihood (FIML) estimation in factor analysis when a majority of the data values are missing. The expectation–maximization (EM) algorithm is often used to find the FIML estimates, in which the missing values on manifest variables are included in complete data. However, the ordinary EM algorithm has an extremely high computational cost. In this paper, we propose a new algorithm that is based on the EM algorithm but that efficiently computes the FIML estimates. A significant improvement in the computational speed is realized by not treating the missing values on manifest variables as a part of complete data. When there are many missing data values, it is not clear if the FIML procedure can achieve good estimation accuracy. In order to investigate this, we conduct Monte Carlo simulations under a wide variety of sample sizes.  相似文献   

14.
A problem of estimating regression coefficients is considered when the distribution of error terms is unknown but symmetric. We propose the use of reference distributions having various kurtosis values. It is assumed that the true error distribution is one of the reference distributions, but the indicator variable for the true distribution is missing. The generalized expectation–maximization algorithm combined with a line search is developed for estimating regression coefficients. Simulation experiments are carried out to compare the performance of the proposed approach with some existing robust regression methods including least absolute deviation, Lp, Huber M regression and an approximation using normal mixtures under various error distributions. As the error distribution is far from a normal distribution, the proposed method is observed to show better performance than other methods.  相似文献   

15.
The power of a clinical trial is partly dependent upon its sample size. With continuous data, the sample size needed to attain a desired power is a function of the within-group standard deviation. An estimate of this standard deviation can be obtained during the trial itself based upon interim data; the estimate is then used to re-estimate the sample size. Gould and Shih proposed a method, based on the EM algorithm, which they claim produces a maximum likelihood estimate of the within-group standard deviation while preserving the blind, and that the estimate is quite satisfactory. However, others have claimed that the method can produce non-unique and/or severe underestimates of the true within-group standard deviation. Here the method is thoroughly examined to resolve the conflicting claims and, via simulation, to assess its validity and the properties of its estimates. The results show that the apparent non-uniqueness of the method's estimate is due to an apparently innocuous alteration that Gould and Shih made to the EM algorithm. When this alteration is removed, the method is valid in that it produces the maximum likelihood estimate of the within-group standard deviation (and also of the within-group means). However, the estimate is negatively biased and has a large standard deviation. The simulations show that with a standardized difference of 1 or less, which is typical in most clinical trials, the standard deviation from the combined samples ignoring the groups is a better estimator, despite its obvious positive bias.  相似文献   

16.
We define the mixture likelihood approach to clustering by discussing the sampling distribution of the likelihood ratio test of the null hypothesis that we have observed a sample of observations of a variable having the bivariate normal distribution versus the alternative that the variable has the bivariate normal mixture with unequal means and common within component covariance matrix. The empirical distribution of the likelihood ratio test indicates that convergence to the chi-squared distribution with 2 df is at best very slow, that the sample size should be 5000 or more for the chi-squared result to hold, and that for correlations between 0.1 and 0.9 there is little, if any, dependence of the null distribution on the correlation. Our simulation study suggests a heuristic function based on the gamma.  相似文献   

17.
We are concerned with three different types of multivariate chi-square distributions. Their members play important roles as limiting distributions of vectors of test statistics in several applications of multiple hypotheses testing. We explain these applications and consider the computation of multiplicity-adjusted p-values under the respective global hypothesis. By means of numerical examples, we demonstrate how much gain in level exhaustion or, equivalently, power can be achieved with corresponding multivariate multiple tests compared with approaches which are only based on univariate marginal distributions and do not take the dependence structure among the test statistics into account. As a further contribution of independent value, we provide an overview of essentially all analytic formulas for computing multivariate chi-square probabilities of the considered types which are available up to present. These formulas were scattered in the previous literature and are presented here in a unified manner.  相似文献   

18.
This work presents a new linear calibration model with replication by assuming that the error of the model follows a skew scale mixture of the normal distributions family, which is a class of asymmetric thick-tailed distributions that includes the skew normal distribution and symmetric distributions. In the literature, most calibration models assume that the errors are normally distributed. However, the normal distribution is not suitable when there are atypical observations and asymmetry. The estimation of the calibration model parameters are done numerically by the EM algorithm. A simulation study is carried out to verify the properties of the maximum likelihood estimators. This new approach is applied to a real dataset from a chemical analysis.  相似文献   

19.
We study the problem of testing: H0 : μ ∈ P against H1 : μ ? P, based on a random sample of N observations from a p-dimensional normal distribution Np(μ, Σ) with Σ > 0 and P a closed convex positively homogeneous set. We develop the likelihood-ratio test (LRT) for this problem. We show that the union-intersection principle leads to a test equivalent to the LRT. It also gives a large class of tests which are shown to be admissible by Stein's theorem (1956). Finally, we give the α-level cutoff points for the LRT.  相似文献   

20.
It is well-known that the nonparametric maximum likelihood estimator (NPMLE) of a survival function may severely underestimate the survival probabilities at very early times for left truncated data. This problem might be overcome by instead computing a smoothed nonparametric estimator (SNE) via the EMS algorithm. The close connection between the SNE and the maximum penalized likelihood estimator is also established. Extensive Monte Carlo simulations demonstrate the superior performance of the SNE over that of the NPMLE, in terms of either bias or variance, even for moderately large Samples. The methodology is illustrated with an application to the Massachusetts Health Care Panel Study dataset to estimate the probability of being functionally independent for non-poor male and female groups rcspectively.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号