Theoretical Machine Learning Seminar

Beyond log-concavity: provable guarantees for sampling multi-modal distributions using simulated tempering Langevin Monte Carlo

A fundamental problem in Bayesian statistics is sampling from distributions that are only specified up to a partition function (constant of proportionality). In particular, we consider the problem of sampling from a distribution given access to the gradient of the log-pdf. For log-concave distributions, classical results due to Bakry and Emery show that natural continuous-time Markov chains called Langevin diffusions mix in polynomial time. But in practice, distributions are often multi-modal and hence non-log-concave, and can take exponential time to mix.We address this problem by combining Langevin diffusion with simulated tempering. The result is a Markov chain that mixes in polynomial rather than exponential time by transitioning between different temperatures of the distribution. We prove fast mixing for any distribution that is close to a mixture of gaussians of equal variance.Based on the paper https://arxiv.org/abs/1710.02736.Joint work with Rong Ge (Duke) and Andrej Risteski (MIT). To appear in NIPS AABI workshop.

Date & Time

November 27, 2017 | 12:30pm – 1:45pm

Location

White-Levy Room

Speakers

Holden Lee

Affiliation

Princeton University