Heiko Strathmann
University College London
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Heiko Strathmann.
Statistical Science | 2015
Anne-Marie Lyne; Mark A. Girolami; Yves F. Atchadé; Heiko Strathmann; Daniel Simpson
A large number of statistical models are “doubly-intractable”: the likelihood normalising term, which is a function of the model parameters, is intractable, as well as the marginal likelihood (model evidence). This means that standard inference techniques to sample from the posterior, such as Markov chain Monte Carlo (MCMC), cannot be used. Examples include, but are not confined to, massive Gaussian Markov random fields, autologistic models and Exponential random graph models. A number of approximate schemes based on MCMC techniques, Approximate Bayesian computation (ABC) or analytic approximations to the posterior have been suggested, and these are reviewed here. Exact MCMC schemes, which can be applied to a subset of doubly-intractable distributions, have also been developed and are described in this paper. As yet, no general method exists which can be applied to all classes of models with doubly-intractable posteriors. In addition, taking inspiration from the Physics literature, we study an alternative method based on representing the intractable likelihood as an infinite series. Unbiased estimates of the likelihood can then be obtained by finite time stochastic truncation of the series via Russian Roulette sampling, although the estimates are not necessarily positive. Results from the Quantum Chromodynamics literature are exploited to allow the use of possibly negative estimates in a pseudo-marginal MCMC scheme such that expectations with respect to the posterior distribution are preserved. The methodology is reviewed on well-known examples such as the parameters in Ising models, the posterior for Fisher–Bingham distributions on the d-Sphere and a largescale Gaussian Markov Random Field model describing the Ozone Column data. This leads to a critical assessment of the strengths and weaknesses of the methodology with pointers to ongoing research.
european conference on machine learning | 2017
Ingmar Schuster; Heiko Strathmann; Brooks Paige; Dino Sejdinovic
We propose kernel sequential Monte Carlo (KSMC), a framework for sampling from static target densities. KSMC is a family of sequential Monte Carlo algorithms that are based on building emulator models of the current particle system in a reproducing kernel Hilbert space. We here focus on modelling nonlinear covariance structure and gradients of the target. The emulator’s geometry is adaptively updated and subsequently used to inform local proposals. Unlike in adaptive Markov chain Monte Carlo, continuous adaptation does not compromise convergence of the sampler. KSMC combines the strengths of sequental Monte Carlo and kernel methods: superior performance for multimodal targets and the ability to estimate model evidence as compared to Markov chain Monte Carlo, and the emulator’s ability to represent targets that exhibit high degrees of nonlinearity. As KSMC does not require access to target gradients, it is particularly applicable on targets whose gradients are unknown or prohibitively expensive. We describe necessary tuning details and demonstrate the benefits of the the proposed methodology on a series of challenging synthetic and real-world examples.
Stat | 2017
Louis Ellam; Heiko Strathmann; Mark A. Girolami; Iain Murray
We propose a determinant-free approach for simulation-based Bayesian inference in high-dimensional Gaussian models. We introduce auxiliary variables with covariance equal to the inverse covariance of the model. The joint probability of the auxiliary model can be computed without evaluating determinants, which are often hard to compute in high dimensions. We develop a Markov chain Monte Carlo sampling scheme for the auxiliary model that requires no more than the application of inverse-matrix-square-roots and the solution of linear systems. These operations can be performed at large scales with rational approximations. We provide an empirical study on both synthetic and real-world data for sparse Gaussian processes and for large-scale Gaussian Markov random fields. Copyright
international conference on learning representations | 2017
Dougal J. Sutherland; Hsiao-Yu Fish Tung; Heiko Strathmann; Soumyajit De; Aaditya Ramdas; Alexander J. Smola; Arthur Gretton
arXiv: Methodology | 2013
Anne-Marie Lyne; Mark A. Girolami; Yves F. Atchadé; Heiko Strathmann; Daniel Simpson
arXiv: Machine Learning | 2015
Heiko Strathmann; Dino Sejdinovic; Mark A. Girolami
international conference on machine learning | 2014
Dino Sejdinovic; Heiko Strathmann; Maria Lomeli Garcia; Christophe Andrieu; Arthur Gretton
international conference on machine learning | 2016
Kacper P. Chwialkowski; Heiko Strathmann; Arthur Gretton
neural information processing systems | 2015
Heiko Strathmann; Dino Sejdinovic; Samuel Livingstone; Zoltán Szabó; Arthur Gretton
international conference on machine learning | 2014
Dino Sejdinovic; Heiko Strathmann; Maria Lomeli Garcia; Christophe Andrieu; Arthur Gretton