Fully Bayesian logistic regression with hyper-LASSO priors for high-dimensional feature selection |
| |
Authors: | Longhai Li Weixin Yao |
| |
Affiliation: | 1. Department of Mathematics and Statistics, University of Saskatchewan, Saskatoon, SK, Canada;2. Department of Statistics, University of California at Riverside, Riverside, CA, USA |
| |
Abstract: | Feature selection arises in many areas of modern science. For example, in genomic research, we want to find the genes that can be used to separate tissues of different classes (e.g. cancer and normal). One approach is to fit regression/classification models with certain penalization. In the past decade, hyper-LASSO penalization (priors) have received increasing attention in the literature. However, fully Bayesian methods that use Markov chain Monte Carlo (MCMC) for regression/classification with hyper-LASSO priors are still in lack of development. In this paper, we introduce an MCMC method for learning multinomial logistic regression with hyper-LASSO priors. Our MCMC algorithm uses Hamiltonian Monte Carlo in a restricted Gibbs sampling framework. We have used simulation studies and real data to demonstrate the superior performance of hyper-LASSO priors compared to LASSO, and to investigate the issues of choosing heaviness and scale of hyper-LASSO priors. |
| |
Keywords: | High-dimensional feature selection non-convex penalties horseshoe heavy-tailed prior hyper-LASSO priors MCMC Hamiltonian Monte Carlo Gibbs sampling fully Bayesian |
|
|