首页 | 本学科首页   官方微博 | 高级检索  
     


Using a Markov Chain to Construct a Tractable Approximation of an Intractable Probability Distribution
Authors:JAMES P. HOBERT  GALIN L. JONES   CHRISTIAN P. ROBERT
Affiliation:Department of Statistics, University of Florida; School of Statistics, University of Minnesota; UniversitéParis Dauphine &CREST, INSEE
Abstract:Abstract.  Let π denote an intractable probability distribution that we would like to explore. Suppose that we have a positive recurrent, irreducible Markov chain that satisfies a minorization condition and has π as its invariant measure. We provide a method of using simulations from the Markov chain to construct a statistical estimate of π from which it is straightforward to sample. We show that this estimate is 'strongly consistent' in the sense that the total variation distance between the estimate and π converges to 0 almost surely as the number of simulations grows. Moreover, we use some recently developed asymptotic results to provide guidance as to how much simulation is necessary. Draws from the estimate can be used to approximate features of π or as intelligent starting values for the original Markov chain. We illustrate our methods with two examples.
Keywords:burn-in    Gibbs sampler    minorization condition    mixture representation    Monte Carlo    regeneration    split chain
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号