# Distributed Markov chain Monte Carlo

We consider the design of Markov chain Monte Carlo (MCMC) methods for large-scale, distributed, heterogeneous compute facilities, with a focus on synthesising sample sets across multiple runs performed in parallel. While theory suggests that many independent Markov chains may be run and their samples pooled, the well-known practical problem of quasi-ergodicity, or poor mixing, frustrates this otherwise simple approach. Furthermore, without some mechanism for hastening the convergence of individual chains, overall speedup from parallelism is limited by the portion of each chain to be discarded as burn-in. Existing multiple-chain methods, such as parallel tempering and population MCMC, use a synchronous exchange of samples to expedite convergence. This work instead proposes mixing in an additional independent proposal, representing some hitherto best estimate or summary of the posterior, and cooperatively adapting this across chains. Such adaptation can be asynchronous, increases the ensemble’s robustness to quasi-ergodic behaviour in constituent chains, and may improve overall tolerance to fault.

### Citation

L.M. Murray (2010). Distributed Markov chain Monte Carlo. NIPS Workshop: Learning on Cores, Clusters and Clouds. url:http://lccc.eecs.berkeley.edu/Papers/dmcmc_short.pdf.

@Article{Murray2010,
title = {Distributed {M}arkov chain {M}onte {C}arlo},
author = {Lawrence Matthew Murray},
journal = {NIPS Workshop: Learning on Cores, Clusters and Clouds},
year = {2010},
url = {http://lccc.eecs.berkeley.edu/Papers/dmcmc_short.pdf}
}