Brooks Paige
University of Oxford
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Brooks Paige.
european conference on machine learning | 2015
David Tolpin; Jan-Willem van de Meent; Brooks Paige; Frank D. Wood
We introduce an adaptive output-sensitive Metropolis-Hastings algorithm for probabilistic models expressed as programs, Adaptive Lightweight Metropolis-Hastings (AdLMH). This algorithm extends Lightweight Metropolis-Hastings (LMH) by adjusting the probabilities of proposing random variables for modification to improve convergence of the program output. We show that AdLMH converges to the correct equilibrium distribution and compare convergence of AdLMH to that of LMH on several test problems to highlight different aspects of the adaptation scheme. We observe consistent improvement in convergence on the test problems.
european conference on machine learning | 2017
Ingmar Schuster; Heiko Strathmann; Brooks Paige; Dino Sejdinovic
We propose kernel sequential Monte Carlo (KSMC), a framework for sampling from static target densities. KSMC is a family of sequential Monte Carlo algorithms that are based on building emulator models of the current particle system in a reproducing kernel Hilbert space. We here focus on modelling nonlinear covariance structure and gradients of the target. The emulator’s geometry is adaptively updated and subsequently used to inform local proposals. Unlike in adaptive Markov chain Monte Carlo, continuous adaptation does not compromise convergence of the sampler. KSMC combines the strengths of sequental Monte Carlo and kernel methods: superior performance for multimodal targets and the ability to estimate model evidence as compared to Markov chain Monte Carlo, and the emulator’s ability to represent targets that exhibit high degrees of nonlinearity. As KSMC does not require access to target gradients, it is particularly applicable on targets whose gradients are unknown or prohibitively expensive. We describe necessary tuning details and demonstrate the benefits of the the proposed methodology on a series of challenging synthetic and real-world examples.
international conference on machine learning | 2014
Brooks Paige; Frank D. Wood
international conference on machine learning | 2016
Brooks Paige; Frank D. Wood
international conference on machine learning | 2017
Matt J. Kusner; Brooks Paige; José Miguel Hernández-Lobato
neural information processing systems | 2017
Siddharth Narayanaswamy; Brooks Paige; Jan-Willem van de Meent; Alban Desmaison; Noah D. Goodman; Pushmeet Kohli; Frank D. Wood; Philip H. S. Torr
neural information processing systems | 2014
Brooks Paige; Frank D. Wood; Arnaud Doucet; Yee Whye Teh
neural information processing systems | 2013
Benjamin Shababo; Brooks Paige; Ari Pakman; Liam Paninski
international conference on artificial intelligence and statistics | 2016
Jan-Willem Vandemeent; Brooks Paige; David Tolpin; Frank D. Wood
Archive | 2017
N. Siddharth; Brooks Paige; Alban Desmaison; Jan-Willem van de Meent; Frank D. Wood; Noah D. Goodman; Pushmeet Kohli; Philip H. S. Torr