首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
In this paper, we investigate robust parameter estimation and variable selection for binary regression models with grouped data. We investigate estimation procedures based on the minimum-distance approach. In particular, we employ minimum Hellinger and minimum symmetric chi-squared distances criteria and propose regularized minimum-distance estimators. These estimators appear to possess a certain degree of automatic robustness against model misspecification and/or for potential outliers. We show that the proposed non-penalized and penalized minimum-distance estimators are efficient under the model and simultaneously have excellent robustness properties. We study their asymptotic properties such as consistency, asymptotic normality and oracle properties. Using Monte Carlo studies, we examine the small-sample and robustness properties of the proposed estimators and compare them with traditional likelihood estimators. We also study two real-data applications to illustrate our methods. The numerical studies indicate the satisfactory finite-sample performance of our procedures.  相似文献   

2.
The minimum disparity estimators proposed by Lindsay (1994) for discrete models form an attractive subclass of minimum distance estimators which achieve their robustness without sacrificing first order efficiency at the model. Similarly, disparity test statistics are useful robust alternatives to the likelihood ratio test for testing of hypotheses in parametric models; they are asymptotically equivalent to the likelihood ratio test statistics under the null hypothesis and contiguous alternatives. Despite their asymptotic optimality properties, the small sample performance of many of the minimum disparity estimators and disparity tests can be considerably worse compared to the maximum likelihood estimator and the likelihood ratio test respectively. In this paper we focus on the class of blended weight Hellinger distances, a general subfamily of disparities, and study the effects of combining two different distances within this class to generate the family of “combined” blended weight Hellinger distances, and identify the members of this family which generally perform well. More generally, we investigate the class of "combined and penal-ized" blended weight Hellinger distances; the penalty is based on reweighting the empty cells, following Harris and Basu (1994). It is shown that some members of the combined and penalized family have rather attractive properties  相似文献   

3.
A general class of minimum distance estimators for logistic regression models based on the ϕ-divergence measures is introduced: The minimum ϕ-divergence estimator, which is seen to be a generalization of the maximum likelihood estimator. Its asymptotic properties are studied as well as its behaviour in small samples throught a simulation study. This work was supported partially by Grant DGI (BMF2003-00892).  相似文献   

4.
For a given parametric probability model, we consider the risk of the maximum likelihood estimator with respect to α-divergence, which includes the special cases of Kullback–Leibler divergence, the Hellinger distance, and essentially χ2-divergence. The asymptotic expansion of the risk is given with respect to sample sizes up to order n? 2. Each term in the expansion is expressed with the geometrical properties of the Riemannian manifold formed by the parametric probability model.  相似文献   

5.
Jingjing Wu 《Statistics》2015,49(4):711-740
The successful application of the Hellinger distance approach to fully parametric models is well known. The corresponding optimal estimators, known as minimum Hellinger distance (MHD) estimators, are efficient and have excellent robustness properties [Beran R. Minimum Hellinger distance estimators for parametric models. Ann Statist. 1977;5:445–463]. This combination of efficiency and robustness makes MHD estimators appealing in practice. However, their application to semiparametric statistical models, which have a nuisance parameter (typically of infinite dimension), has not been fully studied. In this paper, we investigate a methodology to extend the MHD approach to general semiparametric models. We introduce the profile Hellinger distance and use it to construct a minimum profile Hellinger distance estimator of the finite-dimensional parameter of interest. This approach is analogous in some sense to the profile likelihood approach. We investigate the asymptotic properties such as the asymptotic normality, efficiency, and adaptivity of the proposed estimator. We also investigate its robustness properties. We present its small-sample properties using a Monte Carlo study.  相似文献   

6.
We propose penalized minimum φ-divergence estimator for parameter estimation and variable selection in logistic regression. Using an appropriate penalty function, we show that penalized φ-divergence estimator has oracle property. With probability tending to 1, penalized φ-divergence estimator identifies the true model and estimates nonzero coefficients as efficiently as if the sparsity of the true model was known in advance. The advantage of penalized φ-divergence estimator is that it produces estimates of nonzero parameters efficiently than penalized maximum likelihood estimator when sample size is small and is equivalent to it for large one. Numerical simulations confirm our findings.  相似文献   

7.
For the model of independence in a two way contingency table, shrinkage estimators based on minimum φφ-divergence estimators and φφ-divergence statistics are considered. These estimators are based on the James–Stein-type rule and incorporate the idea of preliminary test estimator. The asymptotic bias and risk are obtained under contiguous alternative hypotheses, and on the basis of them a comparison study is carried out.  相似文献   

8.
In this paper we study polytomous logistic regression model and the asymptotic properties of the minimum ϕ-divergence estimators for this model. A simulation study is conducted to analyze the behavior of these estimators as function of the power-divergence measure ϕ(λ) Research partially done when was visiting the Bowling Green State University as the Distinguished Lukacs Professor  相似文献   

9.
In testing of hypothesis, the robustness of the tests is an important concern. Generally, the maximum likelihood-based tests are most efficient under standard regularity conditions, but they are highly non-robust even under small deviations from the assumed conditions. In this paper, we have proposed generalized Wald-type tests based on minimum density power divergence estimators for parametric hypotheses. This method avoids the use of nonparametric density estimation and the bandwidth selection. The trade-off between efficiency and robustness is controlled by a tuning parameter β. The asymptotic distributions of the test statistics are chi-square with appropriate degrees of freedom. The performance of the proposed tests is explored through simulations and real data analysis.  相似文献   

10.
Efficiency and robustness are two fundamental concepts in parametric estimation problems. It was long thought that there was an inherent contradiction between the aims of achieving robustness and efficiency; that is, a robust estimator could not be efficient and vice versa. It is now known that the minimum Hellinger distance approached introduced by Beran [R. Beran, Annals of Statistics 1977;5:445–463] is one way of reconciling the conflicting concepts of efficiency and robustness. For parametric models, it has been shown that minimum Hellinger estimators achieve efficiency at the model density and simultaneously have excellent robustness properties. In this article, we examine the application of this approach in two semiparametric models. In particular, we consider a two‐component mixture model and a two‐sample semiparametric model. In each case, we investigate minimum Hellinger distance estimators of finite‐dimensional Euclidean parameters of particular interest and study their basic asymptotic properties. Small sample properties of the proposed estimators are examined using a Monte Carlo study. The results can be extended to semiparametric models of general form as well. The Canadian Journal of Statistics 37: 514–533; 2009 © 2009 Statistical Society of Canada  相似文献   

11.
Divergence measures play an important role in statistical theory, especially in large sample theories of estimation and testing. The underlying reason is that they are indices of statistical distance between probability distributions P and Q; the smaller these indices are the harder it is to discriminate between P and Q. Many divergence measures have been proposed since the publication of the paper of Kullback and Leibler (1951). Renyi (1961) gave the first generalization of Kullback-Leibler divergence, Jeffreys (1946) defined the J-divergences, Burbea and Rao (1982) introduced the R-divergences, Sharma and Mittal (1977) the (r,s)-divergences, Csiszar (1967) the ϕ-divergences, Taneja (1989) the generalized J-divergences and the generalized R-divergences and so on. In order to do a unified study of their statistical properties, here we propose a generalized divergence, called (h,ϕ)-divergence, which include as particular cases the above mentioned divergence measures. Under different assumptions, it is shown that the asymptotic distributions of the (h,ϕ)-divergence statistics are either normal or chi square. The chi square and the likelihood ratio test statistics are particular cases of the (h,ϕ)-divergence test statistics considered. From the previous results, asymptotic distributions of entropy statistics are derived too. Applications to testing statistical hypothesis in multinomial populations are given. The Pitman and Bahadur efficiencies of tests of goodness of fit and independence based on these statistics are obtained. To finish, apendices with the asymptotic variances of many well known divergence and entropy statistics are presented. The research in this paper was supported in part by DGICYT Grants N. PB91-0387 and N. PB91-0155. Their financial support is gratefully acknowledged.  相似文献   

12.
The present paper studies the minimum Hellinger distance estimator by recasting it as the maximum likelihood estimator in a data driven modification of the model density. In the process, the Hellinger distance itself is expressed as a penalized log likelihood function. The penalty is the sum of the model probabilities over the non-observed values of the sample space. A comparison of the modified model density with the original data provides insights into the robustness of the minimum Hellinger distance estimator. Adjustments of the amount of penalty leads to a class of minimum penalized Hellinger distance estimators, some members of which perform substantially better than the minimum Hellinger distance estimator at the model for small samples, without compromising the robustness properties of the latter.  相似文献   

13.
The maximum likelihood estimator (MLE) is asymptotically efficient for most parametric models under standard regularity conditions, but it has very poor robustness properties. On the other hand some of the minimum disparity estimators like the minimum Hellinger distance estimator (MHDE) have strong robustness features but their small sample efficiency at the model turns out to be very poor compared to the MLE. Methods based on the minimization of some combined disparities can substantially improve their small sample performances without affecting their robustness properties (Park et al., 1995). All studies involving the combined disparity have so far been empirical, and there are no results on the asymptotic properties of these estimators. In view of the usefulness of these procedures this is a major gap in theory, which we try to fill through the present work. Some illustrations of the performance of the estimators and the corresponding tests are also provided.  相似文献   

14.
To perform regression analysis in high dimensions, lasso or ridge estimation are a common choice. However, it has been shown that these methods are not robust to outliers. Therefore, alternatives as penalized M-estimation or the sparse least trimmed squares (LTS) estimator have been proposed. The robustness of these regression methods can be measured with the influence function. It quantifies the effect of infinitesimal perturbations in the data. Furthermore, it can be used to compute the asymptotic variance and the mean-squared error (MSE). In this paper we compute the influence function, the asymptotic variance and the MSE for penalized M-estimators and the sparse LTS estimator. The asymptotic biasedness of the estimators make the calculations non-standard. We show that only M-estimators with a loss function with a bounded derivative are robust against regression outliers. In particular, the lasso has an unbounded influence function.  相似文献   

15.
We consider the right truncated exponential distribution where the truncation point is unknown and show that the ML equation has a unique solution over an extended parameter space. In the case of the estimation of the truncation point T we show that the asymptotic distribution of the MLE is not centered at T. A modified MLE is introduced which outperforms all other considered estimators including the minimum variance unbiased estimator. Asymptotic as well as small sample properties of different estimators are investigated and compared. The truncated exponential distribution has an increasing failure rate, ideally suited for use as a survival distribution for biological and industrial data.  相似文献   

16.
Abstract

Constrained M (CM) estimates of multivariate location and scatter [Kent, J. T., Tyler, D. E. (1996). Constrained M-estimation for multivariate location and scatter. Ann. Statist. 24:1346–1370] are defined as the global minimum of an objective function subject to a constraint. These estimates combine the good global robustness properties of the S estimates and the good local robustness properties of the redescending M estimates. The CM estimates are not explicitly defined. Numerical methods have to be used to compute the CM estimates. In this paper, we give an algorithm to compute the CM estimates. Using the algorithm, we give a small simulation study to demonstrate the capability of the algorithm finding the CM estimates, and also to explore the finite sample behavior of the CM estimates. We also use the CM estimators to estimate the location and scatter parameters of some multivariate data sets to see the performance of the CM estimates dealing with the real data sets that may contain outliers.  相似文献   

17.
The restricted minimum φ-divergence estimator, [Pardo, J.A., Pardo, L. and Zografos, K., 2002, Minimum φ-divergence estimators with constraints in multinomial populations. Journal of Statistical Planning and Inference, 104, 221–237], is employed to obtain estimates of the cell frequencies of an I×I contingency table under hypotheses of symmetry, marginal homogeneity or quasi-symmetry. The associated φ-divergence statistics are distributed asymptotically as chi-squared distributions under the null hypothesis. The new estimators and test statistics contain, as particular cases, the classical estimators and test statistics previously presented in the literature for the cited problems. A simulation study is presented, for the symmetry problem, to choose the best function φ2 for estimation and the best function φ1 for testing.  相似文献   

18.
ABSTRACT

This article is concerned with some parametric and nonparametric estimators for the k-fold convolution of a distribution function. An alternative estimator is proposed and its unbiasedness, asymptotic unbiasedness, and consistency properties are investigated. The asymptotic normality of this estimator is established. Some applications of the estimator are given in renewal processes. Finally, the computational procedures are described and the relative performance of these estimators for small sample sizes is investigated by a simulation study.  相似文献   

19.
This article focuses on the minimum distance estimators under two newly introduced modifications of Cramér–von Mises distance. The generalized power form of Cramér–von Mises distance is defined together with the so-called Kolmogorov–Cramér distance which includes both standard Kolmogorov and Cramér–von Mises distances as limiting special cases. We prove the consistency of Kolmogorov-Cramér estimators in the (expected) L1-norm by direct technique employing domination relations between statistical distances. In our numerical simulation we illustrate the quality of consistency property for sample sizes of the most practical range from n = 10 to n = 500. We study dependence of consistency in L1-norm on ?-contamination neighborhood of the true model and further the robustness of these two newly defined estimators for normal families and contaminated samples. Numerical simulations are used to compare statistical properties of the minimum Kolmogorov–Cramér, generalized Cramér–von Mises, standard Kolmogorov, and Cramér–von Mises distance estimators of the normal family scale parameter. We deal with the corresponding order of consistency and robustness. The resulting graphs are presented and discussed for the cases of the contaminated and uncontaminated pseudo-random samples.  相似文献   

20.
The unbiased estimator of a population variance σ2, S 2 has traditionally been overemphasized, regardless of sample size. In this paper, alternative estimators of population variance are developed. These estimators are biased and have the minimum possible mean-squared error [and we define them as the “minimum mean-squared error biased estimators” (MBBE)]. The comparative merit of these estimators over the unbiased estimator is explored using relative efficiency (RE) (a ratio of mean-squared error values). It is found that, across all population distributions investigated, the RE of the MBBE is much higher for small samples and progressively diminishes to 1 with increasing sample size. The paper gives two applications involving the normal and exponential distributions.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号