arXiv Analytics

Sign in

arXiv:2505.10448 [stat.ML]AbstractReferencesReviewsResources

Efficient MCMC Sampling with Expensive-to-Compute and Irregular Likelihoods

Conor Rosato, Harvinder Lehal, Simon Maskell, Lee Devlin, Malcolm Strens

Published 2025-05-15Version 1

Bayesian inference with Markov Chain Monte Carlo (MCMC) is challenging when the likelihood function is irregular and expensive to compute. We explore several sampling algorithms that make use of subset evaluations to reduce computational overhead. We adapt the subset samplers for this setting where gradient information is not available or is unreliable. To achieve this, we introduce data-driven proxies in place of Taylor expansions and define a novel computation-cost aware adaptive controller. We undertake an extensive evaluation for a challenging disease modelling task and a configurable task with similar irregularity in the likelihood surface. We find our improved version of Hierarchical Importance with Nested Training Samples (HINTS), with adaptive proposals and a data-driven proxy, obtains the best sampling error in a fixed computational budget. We conclude that subset evaluations can provide cheap and naturally-tempered exploration, while a data-driven proxy can pre-screen proposals successfully in explored regions of the state space. These two elements combine through hierarchical delayed acceptance to achieve efficient, exact sampling.

Related articles:
arXiv:2304.02902 [stat.ML] (Published 2023-04-06)
Towards Efficient MCMC Sampling in Bayesian Neural Networks by Exploiting Symmetry
arXiv:2006.04295 [stat.ML] (Published 2020-06-08)
Efficient MCMC Sampling for Bayesian Matrix Factorization by Breaking Posterior Symmetries