Home > Research > Publications & Outputs > MCMC for variationally sparse Gaussian processes
View graph of relations

MCMC for variationally sparse Gaussian processes

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
  • James Hensman
  • Alexander G. De Matthews
  • Maurizio Filippone
  • Zoubin Ghahramani
Close
Publication date2015
Host publicationAdvances in Neural Information Processing Systems
PublisherNeural information processing systems foundation
Pages1648-1656
Number of pages9
Volume2015
<mark>Original language</mark>English
Event29th Annual Conference on Neural Information Processing Systems, NIPS 2015 - Montreal, Canada
Duration: 7/12/201512/12/2015

Conference

Conference29th Annual Conference on Neural Information Processing Systems, NIPS 2015
Country/TerritoryCanada
CityMontreal
Period7/12/1512/12/15

Conference

Conference29th Annual Conference on Neural Information Processing Systems, NIPS 2015
Country/TerritoryCanada
CityMontreal
Period7/12/1512/12/15

Abstract

Gaussian process (GP) models form a core part of probabilistic machine learning. Considerable research effort has been made into attacking three issues with GP models: how to compute efficiently when the number of data is large; how to approximate the posterior when the likelihood is not Gaussian and how to estimate covariance function parameter posteriors. This paper simultaneously addresses these, using a variational approximation to the posterior which is sparse in support of the function but otherwise free-form. The result is a Hybrid Monte-Carlo sampling scheme which allows for a non-Gaussian approximation over the function values and covariance parameters simultaneously, with efficient computations based on inducing-point sparse GPs. Code to replicate each experiment in this paper is available at github.com/sparseMCMC.