首页 | 本学科首页   官方微博 | 高级检索  
     


Nonparametric empirical Bayes for the Dirichlet process mixture model
Authors:Jon D. McAuliffe  David M. Blei  Michael I. Jordan
Affiliation:(1) Statistics Department, University of California, Berkeley, CA, 94720;(2) Computer Science Department, Carnegie Mellon University, Pittsburgh, PA, 15213;(3) Statistics Department and Computer Science Division, University of California, Berkeley, CA, 94720
Abstract:
The Dirichlet process prior allows flexible nonparametric mixture modeling. The number of mixture components is not specified in advance and can grow as new data arrive. However, analyses based on the Dirichlet process prior are sensitive to the choice of the parameters, including an infinite-dimensional distributional parameter G 0. Most previous applications have either fixed G 0 as a member of a parametric family or treated G 0 in a Bayesian fashion, using parametric prior specifications. In contrast, we have developed an adaptive nonparametric method for constructing smooth estimates of G 0. We combine this method with a technique for estimating α, the other Dirichlet process parameter, that is inspired by an existing characterization of its maximum-likelihood estimator. Together, these estimation procedures yield a flexible empirical Bayes treatment of Dirichlet process mixtures. Such a treatment is useful in situations where smooth point estimates of G 0 are of intrinsic interest, or where the structure of G 0 cannot be conveniently modeled with the usual parametric prior families. Analysis of simulated and real-world datasets illustrates the robustness of this approach.
Keywords:Bayesian nonparametrics  Density estimation  Clusetring  Kernel density estimates  Gibbs sampling  Pólya urn schemes
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号