New estimation and feature selection methods in mixture‐of‐experts models |
| |
Authors: | Abbas Khalili |
| |
Institution: | Department of Mathematics and Statistics, McGill University, Montreal, Quebec, Canada H3A 2K6 |
| |
Abstract: | We study estimation and feature selection problems in mixture‐of‐experts models. An $l_2$ ‐penalized maximum likelihood estimator is proposed as an alternative to the ordinary maximum likelihood estimator. The estimator is particularly advantageous when fitting a mixture‐of‐experts model to data with many correlated features. It is shown that the proposed estimator is root‐$n$ consistent, and simulations show its superior finite sample behaviour compared to that of the maximum likelihood estimator. For feature selection, two extra penalty functions are applied to the $l_2$ ‐penalized log‐likelihood function. The proposed feature selection method is computationally much more efficient than the popular all‐subset selection methods. Theoretically it is shown that the method is consistent in feature selection, and simulations support our theoretical results. A real‐data example is presented to demonstrate the method. The Canadian Journal of Statistics 38: 519–539; 2010 © 2010 Statistical Society of Canada |
| |
Keywords: | Mixture‐of‐experts regularization LASSO SCAD EM algorithms MSC 2000: Primary 62F12 secondary 62J02 62J07 |
|
|