Final published version, 8.83 MB, PDF document
Available under license: CC BY: Creative Commons Attribution 4.0 International License
Accepted author manuscript
Licence: CC BY: Creative Commons Attribution 4.0 International License
Final published version
Research output: Contribution to Journal/Magazine › Journal article › peer-review
Research output: Contribution to Journal/Magazine › Journal article › peer-review
}
TY - JOUR
T1 - Coin Sampling
T2 - Gradient-Based Bayesian Inference without Learning Rates
AU - Sharrock, Louis
AU - Nemeth, Christopher
N1 - In: Proceedings of the 40th International Conference on Machine Learning (ICML), Hawaii, USA.
PY - 2023/7/23
Y1 - 2023/7/23
N2 - In recent years, particle-based variational inference (ParVI) methods such as Stein variational gradient descent (SVGD) have grown in popularity as scalable methods for Bayesian inference. Unfortunately, the properties of such methods invariably depend on hyperparameters such as the learning rate, which must be carefully tuned by the practitioner in order to ensure convergence to the target measure at a suitable rate. In this paper, we introduce a suite of new particle-based methods for scalable Bayesian inference based on coin betting, which are entirely learning-rate free. We illustrate the performance of our approach on a range of numerical examples, including several high-dimensional models and datasets, demonstrating comparable performance to other ParVI algorithms with no need to tune a learning rate.
AB - In recent years, particle-based variational inference (ParVI) methods such as Stein variational gradient descent (SVGD) have grown in popularity as scalable methods for Bayesian inference. Unfortunately, the properties of such methods invariably depend on hyperparameters such as the learning rate, which must be carefully tuned by the practitioner in order to ensure convergence to the target measure at a suitable rate. In this paper, we introduce a suite of new particle-based methods for scalable Bayesian inference based on coin betting, which are entirely learning-rate free. We illustrate the performance of our approach on a range of numerical examples, including several high-dimensional models and datasets, demonstrating comparable performance to other ParVI algorithms with no need to tune a learning rate.
M3 - Journal article
VL - 202
SP - 30850
EP - 30882
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
SN - 1938-7228
ER -