TY - UNPB
T1 - Accelerating delayed-acceptance Markov chain Monte Carlo algorithms
AU - Wiqvist, Samuel
AU - Picchini, Umberto
AU - Forman, Julie Lyng
PY - 2018
Y1 - 2018
N2 - Delayed-acceptance Markov chain Monte Carlo (DA-MCMC) samples from a probability distribution, via a two-stages version of the Metropolis-Hastings algorithm, by combining the target distribution with a "surrogate" (i.e. an approximate and computationally cheaper version) of said distribution. DA-MCMC accelerates MCMC sampling in complex applications, while still targeting the exact distribution. We design a computationally faster DA-MCMC algorithm, which samples from an approximation of the target distribution. As a case study, we also introduce a novel stochastic differential equation model for protein folding data. We consider parameters inference in a Bayesian setting where a surrogate likelihood function is introduced in the delayed-acceptance scheme. In our applications we employ a Gaussian process as a surrogate likelihood, but other options are possible. In our accelerated algorithm the calculations in the "second stage" of the delayed-acceptance scheme are reordered in such as way that we can obtain a significant speed-up in the MCMC sampling, when the evaluation of the likelihood function is computationally intensive. We consider both simulations studies, and the analysis of real protein folding data. Simulation studies for the stochastic Ricker model and the novel stochastic differential equation model for protein-folding data, show that the speed-up is highly problem dependent. The more involved the computations of the likelihood function are, the higher the acceleration becomes when using our algorithm. Inference results for the standard delayed-acceptance algorithm and our approximated version are similar, indicating that our approximated algorithm can return reliable Bayesian inference.
AB - Delayed-acceptance Markov chain Monte Carlo (DA-MCMC) samples from a probability distribution, via a two-stages version of the Metropolis-Hastings algorithm, by combining the target distribution with a "surrogate" (i.e. an approximate and computationally cheaper version) of said distribution. DA-MCMC accelerates MCMC sampling in complex applications, while still targeting the exact distribution. We design a computationally faster DA-MCMC algorithm, which samples from an approximation of the target distribution. As a case study, we also introduce a novel stochastic differential equation model for protein folding data. We consider parameters inference in a Bayesian setting where a surrogate likelihood function is introduced in the delayed-acceptance scheme. In our applications we employ a Gaussian process as a surrogate likelihood, but other options are possible. In our accelerated algorithm the calculations in the "second stage" of the delayed-acceptance scheme are reordered in such as way that we can obtain a significant speed-up in the MCMC sampling, when the evaluation of the likelihood function is computationally intensive. We consider both simulations studies, and the analysis of real protein folding data. Simulation studies for the stochastic Ricker model and the novel stochastic differential equation model for protein-folding data, show that the speed-up is highly problem dependent. The more involved the computations of the likelihood function are, the higher the acceleration becomes when using our algorithm. Inference results for the standard delayed-acceptance algorithm and our approximated version are similar, indicating that our approximated algorithm can return reliable Bayesian inference.
KW - Bayesian inference
KW - Gaussian process
KW - pseudo marginal MCMC
KW - protein folding
KW - Stochastic differential equation
UR - https://arxiv.org/abs/1806.05982
M3 - Working paper
BT - Accelerating delayed-acceptance Markov chain Monte Carlo algorithms
ER -