Distributed Markov chain Monte Carlo
published: Jan. 13, 2011, recorded: December 2010, views: 298
Report a problem or upload filesIf you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
We consider the design of Markov chain Monte Carlo (MCMC) methods for large-scale, distributed, heterogeneous compute facilities, with a focus on synthesising sample sets across multiple runs performed in parallel. While theory suggests that many independent Markov chains may be run and their samples pooled, the well-known practical problem of quasi-ergodicity, or poor mixing, frustrates this otherwise simple approach. Furthermore, without some mechanism for hastening the convergence of individual chains, overall speedup from parallelism is limited by the portion of each chain to be discarded as burn-in. Existing multiple-chain methods, such as parallel tempering and population MCMC, use a synchronous exchange of samples to expedite convergence. This work instead proposes mixing in an additional independent proposal, representing some hitherto best estimate or summary of the posterior, and cooperatively adapting this across chains. Such adaptation can be asynchronous, increases the ensemble’s robustness to quasi-ergodic behaviour in constituent chains, and may improve overall tolerance to fault.
Link this pageWould you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !