Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GFlowOut: Dropout with Generative Flow Networks (2210.12928v3)

Published 24 Oct 2022 in cs.LG and cs.AI

Abstract: Bayesian Inference offers principled tools to tackle many critical problems with modern neural networks such as poor calibration and generalization, and data inefficiency. However, scaling Bayesian inference to large architectures is challenging and requires restrictive approximations. Monte Carlo Dropout has been widely used as a relatively cheap way for approximate Inference and to estimate uncertainty with deep neural networks. Traditionally, the dropout mask is sampled independently from a fixed distribution. Recent works show that the dropout mask can be viewed as a latent variable, which can be inferred with variational inference. These methods face two important challenges: (a) the posterior distribution over masks can be highly multi-modal which can be difficult to approximate with standard variational inference and (b) it is not trivial to fully utilize sample-dependent information and correlation among dropout masks to improve posterior estimation. In this work, we propose GFlowOut to address these issues. GFlowOut leverages the recently proposed probabilistic framework of Generative Flow Networks (GFlowNets) to learn the posterior distribution over dropout masks. We empirically demonstrate that GFlowOut results in predictive distributions that generalize better to out-of-distribution data, and provide uncertainty estimates which lead to better performance in downstream tasks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (48)
  1. Adaptive dropout for training deep neural networks. Neural Information Processing Systems (NIPS), 2013.
  2. Flow network based generative models for non-iterative diverse candidate generation. Neural Information Processing Systems (NeurIPS), 2021a.
  3. Gflownet foundations. arXiv preprint arXiv:2111.09266, 2021b.
  4. Uncertainty as a form of transparency: Measuring, communicating, and using uncertainty. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, pp.  401–413, 2021.
  5. Learnable Bernoulli dropout for Bayesian deep learning. Artificial Intelligence and Statistics (AISTATS), 2020.
  6. Deep Gaussian processes. Artificial Intelligence and Statistics (AISTATS), 2013.
  7. Bayesian deep learning via subnetwork inference. International Conference on Machine Learning (ICML), 2021.
  8. Bayesian structure learning with generative flow networks. Uncertainty in Artificial Intelligence (UAI), 2022.
  9. Contextual dropout: An efficient sample-dependent dropout module. International Conference on Learning Representations (ICLR), 2021.
  10. On the expressiveness of approximate inference in Bayesian neural networks. Neural Information Processing Systems (NeurIPS), 2020.
  11. Deep ensembles: A loss landscape perspective. arXiv preprint arXiv:1912.02757, 2019.
  12. Dropout as a Bayesian approximation: Representing model uncertainty in deep learning. International Conference on Machine Learning (ICML), 2016.
  13. Concrete dropout. Neural Information Processing Systems (NIPS), 30, 2017.
  14. Dropblock: A regularization method for convolutional networks. Neural Information Processing Systems (NeurIPS), 2018.
  15. On calibration of modern neural networks. International Conference on Machine Learning (ICML), 2017.
  16. Deep residual learning for image recognition. Computer Vision and Pattern Recognition (CVPR), 2016.
  17. Benchmarking neural network robustness to common corruptions and perturbations. International Conference on Learning Representations (ICLR), 2019.
  18. Improving neural networks by preventing co-adaptation of feature detectors. arXiv preprint arXiv:1207.0580, 2012.
  19. Biological sequence design with gflownets. International Conference on Machine Learning (ICML), 2022.
  20. DEUP: Direct epistemic uncertainty prediction. Transactions on Machine Learning Research (TMLR), 2023.
  21. Variational dropout and the local reparameterization trick. Neural Information Processing Systems (NIPS), 2015.
  22. Accurate uncertainties for deep learning using calibrated regression. International Conference on Machine Learning (ICML), 2018.
  23. Is MC dropout bayesian? arXiv preprint arXiv:2110.04286, 2021.
  24. Meta dropout: Learning to perturb latent features for generalization. International Conference on Learning Representations (ICLR), 2020.
  25. Bayesian model selection, the marginal likelihood, and generalization. International Conference on Machine Learning (ICML), 2022.
  26. MacKay, D. J. A practical Bayesian framework for backpropagation networks. Neural Computation, 4(3):448–472, 1992.
  27. Learning GFlowNets from partial episodes for improved convergence and stability. International Conference on Machine Learning (ICML), 2023.
  28. Trajectory balance: Improved credit assignment in GFlowNets. Neural Information Processing Systems (NeurIPS), 2022.
  29. GFlowNets and variational inference. International Conference on Learning Representations (ICLR), 2023.
  30. Uncertainty Baselines: Benchmarks for uncertainty & robustness in deep learning. arXiv preprint arXiv:2106.04015, 2021.
  31. Neal, R. M. Bayesian learning for neural networks, volume 118. Springer Science & Business Media, 2012.
  32. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images. Computer Vision and Pattern Recognition (CVPR), 2015.
  33. Structured dropout variational inference for Bayesian neural networks. Neural Information Processing Systems (NeurIPS), 2021.
  34. Evaluating generalization in gflownets for molecule design. ICLR 2022 Machine Learning for Drug Discovery workshop, 2022.
  35. (more) efficient reinforcement learning via posterior sampling. Neural Information Processing Systems (NIPS), 2013.
  36. Can you trust your model’s uncertainty? evaluating predictive uncertainty under dataset shift. Neural Information Processing Systems (NeurIPS), 2019.
  37. Generative augmented flow networks. International Conference on Learning Representations (ICLR), 2023.
  38. Analysis on the dropout effect in convolutional neural networks. Asian Conference on Computer Vision, 2016.
  39. Autodropout: Learning dropout patterns to regularize deep networks. Association for the Advancement of Artificial Intelligence (AAAI), 2021.
  40. The eICU Collaborative Research Database, a freely available multi-center database for critical care research. Scientific data, 5(1):1–13, 2018.
  41. Evidential deep learning to quantify classification uncertainty. Neural Information Processing Systems (NeurIPS), 2018.
  42. Reinforcement learning: An introduction. MIT press, 2018.
  43. Bayesian deep learning and a probabilistic perspective of generalization. Neural Information Processing Systems (NeurIPS), 2020.
  44. Soft dropout and its variational bayes approximation. Machine Learning for Signal Processing (MLSP), 2019.
  45. Rx equalization for a high-speed channel based on bayesian active learning using dropout. Electrical Performance of Electronic Packaging and Systems (EPEPS), 2020.
  46. Deep modular co-attention networks for visual question answering. Computer Vision and Pattern Recognition (CVPR), 2019.
  47. Unifying generative models with gflownets. arXiv preprint arXiv:2209.02606, 2022a.
  48. Generative flow networks for discrete probabilistic modeling. International Conference on Machine Learning (ICML), 2022b.
Citations (17)

Summary

We haven't generated a summary for this paper yet.