Automatic choice of driving values in Monte Carlo likelihood approximation via posterior simulations |
| |
Authors: | Kuk Anthony Y. C. |
| |
Affiliation: | (1) Department of Statistics and Applied Probability, National University of Singapore, 3 Science Drive 2, Singapore, 117543 |
| |
Abstract: | For models with random effects or missing data, the likelihood function is sometimes intractable analytically but amenable to Monte Carlo approximation. To get a good approximation, the parameter value that drives the simulations should be sufficiently close to the maximum likelihood estimate (MLE) which unfortunately is unknown. Introducing a working prior distribution, we express the likelihood function as a posterior expectation and approximate it using posterior simulations. If the sample size is large, the sample information is likely to outweigh the prior specification and the posterior simulations will be concentrated around the MLE automatically, leading to good approximation of the likelihood near the MLE. For smaller samples, we propose to use the current posterior as the next prior distribution to make the posterior simulations closer to the MLE and hence improve the likelihood approximation. By using the technique of data duplication, we can simulate from the sharpened posterior distribution without actually updating the prior distribution. The suggested method works well in several test cases. A more complex example involving censored spatial data is also discussed. |
| |
Keywords: | Bayesian computation data duplication generalized linear mixed models Markov chain Monte Carlo missing data prior sharpening random effects |
本文献已被 SpringerLink 等数据库收录! |
|