Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Generative Posterior Networks for Approximately Bayesian Epistemic Uncertainty Estimation (2312.17411v1)

Published 29 Dec 2023 in cs.LG and stat.ML

Abstract: In many real-world problems, there is a limited set of training data, but an abundance of unlabeled data. We propose a new method, Generative Posterior Networks (GPNs), that uses unlabeled data to estimate epistemic uncertainty in high-dimensional problems. A GPN is a generative model that, given a prior distribution over functions, approximates the posterior distribution directly by regularizing the network towards samples from the prior. We prove theoretically that our method indeed approximates the Bayesian posterior and show empirically that it improves epistemic uncertainty estimation and scalability over competing methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. Variational inference: A review for statisticians. Journal of the American statistical Association, 112(518):859–877, 2017.
  2. Weight uncertainty in neural network. In International Conference on Machine Learning, pp. 1613–1622. PMLR, 2015.
  3. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp. 1597–1607. PMLR, 2020.
  4. Efficient model-based reinforcement learning through optimistic policy search and planning. In Neural Information Processing Systems (NeurIPS), 2020. URL https://arxiv.org/abs/2006.08684.
  5. Aleatory or epistemic? does it matter? Structural safety, 31(2):105–112, 2009.
  6. Gal, Y. Uncertainty in deep learning. PhD thesis, University of Cambridge, 2016.
  7. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In international conference on machine learning, pp. 1050–1059. PMLR, 2016.
  8. Gpytorch: Blackbox matrix-matrix gaussian process inference with gpu acceleration. In Advances in Neural Information Processing Systems, 2018.
  9. On calibration of modern neural networks. In International Conference on Machine Learning, pp. 1321–1330. PMLR, 2017.
  10. Hamidieh, K. A data-driven statistical model for predicting the critical temperature of a superconductor. Computational Materials Science, 154:346–354, 2018.
  11. Bayesian deep ensembles via the neural tangent kernel. arXiv preprint arXiv:2007.05864, 2020.
  12. Gaussian processes for big data. arXiv preprint arXiv:1309.6835, 2013.
  13. Simple and principled uncertainty estimation with deterministic deep learning via distance awareness. Advances in Neural Information Processing Systems, 33:7498–7512, 2020.
  14. Uncertainty in neural networks: Approximately bayesian ensembling. In International conference on artificial intelligence and statistics, pp.  234–244. PMLR, 2020.
  15. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
  16. Practical bayesian optimization of machine learning algorithms. Advances in neural information processing systems, 25, 2012.
  17. Bayesian optimization with robust bayesian neural networks. Advances in neural information processing systems, 29:4134–4142, 2016.
  18. Confidence-calibrated adversarial training: Generalizing to unseen attacks. In International Conference on Machine Learning, pp. 9155–9166. PMLR, 2020.
  19. The calculation of posterior distributions by data augmentation. Journal of the American statistical Association, 82(398):528–540, 1987.
  20. Titsias, M. Variational learning of inducing variables in sparse gaussian processes. In Artificial intelligence and statistics, pp.  567–574. PMLR, 2009.
  21. Uncertainty estimation using a single deep deterministic neural network. In International conference on machine learning, pp. 9690–9700. PMLR, 2020.
  22. Kernel interpolation for scalable structured gaussian processes (kiss-gp). In International Conference on Machine Learning, pp. 1775–1784. PMLR, 2015.
  23. Deep kernel learning. In Artificial intelligence and statistics, pp.  370–378. PMLR, 2016.

Summary

We haven't generated a summary for this paper yet.