Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Compact Representation for Bayesian Neural Networks By Removing Permutation Symmetry (2401.00611v1)

Published 31 Dec 2023 in stat.ML, cs.AI, and cs.LG

Abstract: Bayesian neural networks (BNNs) are a principled approach to modeling predictive uncertainties in deep learning, which are important in safety-critical applications. Since exact Bayesian inference over the weights in a BNN is intractable, various approximate inference methods exist, among which sampling methods such as Hamiltonian Monte Carlo (HMC) are often considered the gold standard. While HMC provides high-quality samples, it lacks interpretable summary statistics because its sample mean and variance is meaningless in neural networks due to permutation symmetry. In this paper, we first show that the role of permutations can be meaningfully quantified by a number of transpositions metric. We then show that the recently proposed rebasin method allows us to summarize HMC samples into a compact representation that provides a meaningful explicit uncertainty estimate for each weight in a neural network, thus unifying sampling methods with variational inference. We show that this compact representation allows us to compare trained BNNs directly in weight space across sampling methods and variational inference, and to efficiently prune neural networks trained without explicit Bayesian frameworks by exploiting uncertainty estimates from HMC.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. Git re-basin: Merging models modulo permutation symmetries. In International Conference on Learning Representations, 2023.
  2. Weight uncertainty in neural network. In International Conference on Machine Learning, 2015.
  3. Weight-space symmetry in deep networks gives rise to permutation saddles, connected by equal-loss valleys across the loss landscape. arXiv preprint arXiv:1907.02911, 2019.
  4. The role of permutation invariance in linear mode connectivity of neural networks. In International Conference on Learning Representations, 2022.
  5. Linear mode connectivity and the lottery ticket hypothesis. In International Conference on Machine Learning, 2020.
  6. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In International Conference on Machine Learning, 2016.
  7. Robert Hecht-Nielsen. On the algebraic structure of feedforward network weight spaces. In Advanced Neural Computers, pages 129–135. Elsevier, 1990.
  8. What are bayesian neural network posteriors really like? In International Conference on Machine Learning, 2021.
  9. Anthony W Knapp. Basic algebra. Springer Science & Business Media, 2007.
  10. Being bayesian, even just a bit, fixes overconfidence in relu networks. In International Conference on Machine Learning, 2020.
  11. On symmetries in variational bayesian neural nets. NeurIPS 2021 Workshop on Bayesian Deep Learning, 2021.
  12. Simple and scalable predictive uncertainty estimation using deep ensembles. In Advances in Neural Information Processing Systems, 2017.
  13. Radford M Neal et al. Mcmc using hamiltonian dynamics. Handbook of markov chain monte carlo, 2(11):2, 2011.
  14. Improving the identifiability of neural networks for bayesian inference. NIPS Workshop on Bayesian Deep Learning, 2017.
  15. A scalable laplace approximation for neural networks. In International Conference on Learning Representations, 2018.
  16. Post-training neural network compression with variational bayesian quantization. NeurIPS 2022 Workshop on Challenges in Deploying and Monitoring Machine Learning Systems, 2022.
  17. Bayesian learning via stochastic gradient langevin dynamics. In International Conference on Machine Learning, 2011.
  18. Towards efficient mcmc sampling in bayesian neural networks by exploiting symmetry. arXiv preprint arXiv:2304.02902, 2023.
  19. Evaluating approximate inference in bayesian deep learning. In NeurIPS 2021 Competitions and Demonstrations Track, 2022.
  20. Variational bayesian quantization. In International Conference on Machine Learning, 2020.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Tim Z. Xiao (16 papers)
  2. Weiyang Liu (83 papers)
  3. Robert Bamler (33 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.