Adaptive independent Metropolis-Hastings

40Citations
Citations of this article
41Readers
Mendeley users who have this article in their library.

Abstract

We propose an adaptive independent Metropolis-Hastings algorithm with the ability to learn from all previous proposals in the chain except the current location. It is an extension of the independent Metropolis-Hastings algorithm. Convergence is proved provided a strong Doeblin condition is satisfied, which essentially requires that all the proposal functions have uniformly heavier tails than the stationary distribution. The proof also holds if proposals depending on the current state are used intermittently, provided the information from these iterations is not used for adaption. The algorithm gives samples from the exact distribution within a finite number of iterations with probability arbitrarily close to 1. The algorithm is particularly useful when a large number of samples from the same distribution is necessary, like in Bayesian estimation, and in CPU intensive applications like, for example, in inverse problems and optimization. © Institute of Mathematical Statistics, 2009.

Cite

CITATION STYLE

APA

Holden, L., Hauge, R., & Holden, M. (2009). Adaptive independent Metropolis-Hastings. Annals of Applied Probability, 19(1), 395–413. https://doi.org/10.1214/08-AAP545

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free