Abstract
Bayesian neural networks often approximate the weight-posterior with a Gaussian distribution. However, practical posteriors are often, even locally, highly non-Gaussian, and empirical performance deteriorates. We propose a simple parametric approximate posterior that adapts to the shape of the true posterior through a Riemannian metric that is determined by the log-posterior gradient. We develop a Riemannian Laplace approximation where samples naturally fall into weightregions with low negative log-posterior. We show that these samples can be drawn by solving a system of ordinary differential equations, which can be done efficiently by leveraging the structure of the Riemannian metric and automatic differentiation. Empirically, we demonstrate that our approach consistently improves over the conventional Laplace approximation across tasks. We further show that, unlike the conventional Laplace approximation, our method is not overly sensitive to the choice of prior, which alleviates a practical pitfall of current approaches.
Original language | English |
---|---|
Title of host publication | Proceedings of the 37th Conference on Neural Information Processing Systems |
Number of pages | 28 |
Volume | 36 |
Publisher | Neural Information Processing Systems Foundation |
Publication date | 2023 |
Publication status | Published - 2023 |
Event | 37th Annual Conference on Neural Information Processing Systems - Ernest N. Morial Convention Center, New Orleans, United States Duration: 10 Dec 2023 → 16 Dec 2023 Conference number: 37 |
Conference
Conference | 37th Annual Conference on Neural Information Processing Systems |
---|---|
Number | 37 |
Location | Ernest N. Morial Convention Center |
Country/Territory | United States |
City | New Orleans |
Period | 10/12/2023 → 16/12/2023 |