Home > Research > Publications & Outputs > Large-Scale Stochastic Sampling from the Probab...

Electronic data

  • 1806.07137v1

    Accepted author manuscript, 450 KB, PDF document

    Available under license: CC BY-NC: Creative Commons Attribution-NonCommercial 4.0 International License

Links

View graph of relations

Large-Scale Stochastic Sampling from the Probability Simplex

Research output: Contribution to conference - Without ISBN/ISSN Conference paperpeer-review

Published

Standard

Large-Scale Stochastic Sampling from the Probability Simplex. / Baker, Jack; Fearnhead, Paul; Fox, Emily B et al.
2018. 6722-6732 Paper presented at 32nd Neural Information Processing Systems Conference (NIPS 2018), Montreal, Canada.

Research output: Contribution to conference - Without ISBN/ISSN Conference paperpeer-review

Harvard

Baker, J, Fearnhead, P, Fox, EB & Nemeth, C 2018, 'Large-Scale Stochastic Sampling from the Probability Simplex', Paper presented at 32nd Neural Information Processing Systems Conference (NIPS 2018), Montreal, Canada, 3/12/18 - 8/12/18 pp. 6722-6732 . <https://dl.acm.org/citation.cfm?id=3327778>

APA

Baker, J., Fearnhead, P., Fox, E. B., & Nemeth, C. (2018). Large-Scale Stochastic Sampling from the Probability Simplex. 6722-6732 . Paper presented at 32nd Neural Information Processing Systems Conference (NIPS 2018), Montreal, Canada. https://dl.acm.org/citation.cfm?id=3327778

Vancouver

Baker J, Fearnhead P, Fox EB, Nemeth C. Large-Scale Stochastic Sampling from the Probability Simplex. 2018. Paper presented at 32nd Neural Information Processing Systems Conference (NIPS 2018), Montreal, Canada.

Author

Baker, Jack ; Fearnhead, Paul ; Fox, Emily B et al. / Large-Scale Stochastic Sampling from the Probability Simplex. Paper presented at 32nd Neural Information Processing Systems Conference (NIPS 2018), Montreal, Canada.11 p.

Bibtex

@conference{b643d9b64d334f1e8562d2dd23d3bfa2,
title = "Large-Scale Stochastic Sampling from the Probability Simplex",
abstract = "Stochastic gradient Markov chain Monte Carlo (SGMCMC) has become a popular method for scalable Bayesian inference. These methods are based on sampling a discrete-time approximation to a continuous time process, such as the Langevin diffusion. When applied to distributions defined on a constrained space, such as the simplex, the time-discretisation error can dominate when we are near the boundary of the space. We demonstrate that while current SGMCMC methods for the simplex perform well in certain cases, they struggle with sparse simplex spaces; when many of the components are close to zero. However, most popular large-scale applications of Bayesian inference on simplex spaces, such as network or topic models, are sparse. We argue that this poor performance is due to the biases of SGMCMC caused by the discretization error. To get around this, we propose the stochastic CIR process, which removes all discretization error and we prove that samples from the stochastic CIR process are asymptotically unbiased. Use of the stochastic CIR process within a SGMCMC algorithm is shown to give substantially better performance for a topic model and a Dirichlet process mixture model than existing SGMCMC approaches.",
keywords = "stat.CO, cs.LG, stat.ML",
author = "Jack Baker and Paul Fearnhead and Fox, {Emily B} and Christopher Nemeth",
year = "2018",
month = dec,
day = "3",
language = "English",
pages = "6722--6732 ",
note = "32nd Neural Information Processing Systems Conference (NIPS 2018) ; Conference date: 03-12-2018 Through 08-12-2018",
url = "https://nips.cc/",

}

RIS

TY - CONF

T1 - Large-Scale Stochastic Sampling from the Probability Simplex

AU - Baker, Jack

AU - Fearnhead, Paul

AU - Fox, Emily B

AU - Nemeth, Christopher

PY - 2018/12/3

Y1 - 2018/12/3

N2 - Stochastic gradient Markov chain Monte Carlo (SGMCMC) has become a popular method for scalable Bayesian inference. These methods are based on sampling a discrete-time approximation to a continuous time process, such as the Langevin diffusion. When applied to distributions defined on a constrained space, such as the simplex, the time-discretisation error can dominate when we are near the boundary of the space. We demonstrate that while current SGMCMC methods for the simplex perform well in certain cases, they struggle with sparse simplex spaces; when many of the components are close to zero. However, most popular large-scale applications of Bayesian inference on simplex spaces, such as network or topic models, are sparse. We argue that this poor performance is due to the biases of SGMCMC caused by the discretization error. To get around this, we propose the stochastic CIR process, which removes all discretization error and we prove that samples from the stochastic CIR process are asymptotically unbiased. Use of the stochastic CIR process within a SGMCMC algorithm is shown to give substantially better performance for a topic model and a Dirichlet process mixture model than existing SGMCMC approaches.

AB - Stochastic gradient Markov chain Monte Carlo (SGMCMC) has become a popular method for scalable Bayesian inference. These methods are based on sampling a discrete-time approximation to a continuous time process, such as the Langevin diffusion. When applied to distributions defined on a constrained space, such as the simplex, the time-discretisation error can dominate when we are near the boundary of the space. We demonstrate that while current SGMCMC methods for the simplex perform well in certain cases, they struggle with sparse simplex spaces; when many of the components are close to zero. However, most popular large-scale applications of Bayesian inference on simplex spaces, such as network or topic models, are sparse. We argue that this poor performance is due to the biases of SGMCMC caused by the discretization error. To get around this, we propose the stochastic CIR process, which removes all discretization error and we prove that samples from the stochastic CIR process are asymptotically unbiased. Use of the stochastic CIR process within a SGMCMC algorithm is shown to give substantially better performance for a topic model and a Dirichlet process mixture model than existing SGMCMC approaches.

KW - stat.CO

KW - cs.LG

KW - stat.ML

M3 - Conference paper

SP - 6722

EP - 6732

T2 - 32nd Neural Information Processing Systems Conference (NIPS 2018)

Y2 - 3 December 2018 through 8 December 2018

ER -