Continual Learning via Sequential Function-Space Variational Inference (2312.17210v1)
Abstract: Sequential Bayesian inference over predictive functions is a natural framework for continual learning from streams of data. However, applying it to neural networks has proved challenging in practice. Addressing the drawbacks of existing techniques, we propose an optimization objective derived by formulating continual learning as sequential function-space variational inference. In contrast to existing methods that regularize neural network parameters directly, this objective allows parameters to vary widely during training, enabling better adaptation to new tasks. Compared to objectives that directly regularize neural network predictions, the proposed objective allows for more flexible variational distributions and more effective regularization. We demonstrate that, across a range of task sequences, neural networks trained via sequential function-space variational inference achieve better predictive accuracy than networks trained with related methods while depending less on maintaining a set of representative points from previous tasks.
- Uncertainty-based continual learning with adaptive regularization. In Advances in Neural Information Processing Systems.
- Memory aware synapses: learning what (not) to forget. In European Conference on Computer Vision.
- Measuring and regularizing networks in function space. In International Conference on Learning Representations.
- Streaming variational Bayes. In Advances in Neural Information Processing Systems.
- Streaming sparse Gaussian process approximations. In Advances in Neural Information Processing Systems.
- Understanding variational inference in function-space. In Symposium on Advances in Approximate Bayesian Inference.
- Dark experience for general continual learning: a strong, simple baseline. In Advances in Neural Information Processing Systems.
- Riemannian walk for incremental learning: understanding forgetting and intransigence. In European Conference on Computer Vision.
- Incremental variational sparse Gaussian process regression. In Advances in Neural Information Processing Systems.
- Elements of Information Theory. Wiley, New York.
- Csató, L. (2002). Gaussian processes: iterative sparse approximations. PhD thesis, Aston University.
- Sparse on-line Gaussian processes. Neural Computation.
- A continual learning survey: defying forgetting in classification tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence.
- Uncertainty-guided continual learning with Bayesian neural networks. In International Conference on Learning Representations.
- Towards robust evaluations of continual learning. ICML Workshop on Lifelong Learning: A Reinforcement Learning Approach.
- Online variational Bayesian learning. In NIPS Workshop on Online Learning.
- On-line variational Bayesian learning. In International Symposium on Independent Component Analysis and Blind Signal Separation.
- Less-forgetful learning for domain expansion in deep neural networks. In AAAI Conference on Artificial Intelligence.
- Variational auto-regressive Gaussian processes for continual learning. In International Conference on Machine Learning.
- Hierarchical Indian buffet neural networks for Bayesian continual learning. arXiv.
- Keep and learn: continual learning by constraining the latent space for knowledge preservation in neural networks. In Medical Image Computing and Computer Assisted Intervention.
- Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences.
- Human-level concept learning through probabilistic program induction. Science.
- Overcoming catastrophic forgetting by incremental moment matching. In Advances in Neural Information Processing Systems.
- Learning without forgetting. IEEE Transactions on Pattern Analysis and Machine Intelligence.
- Rotate your networks: better weight consolidation and less catastrophic forgetting. International Conference on Pattern Recognition.
- Generalized variational continual learning. In International Conference on Learning Representations.
- Gradient episodic memory for continual learning. In Advances in Neural Information Processing Systems.
- On sparse variational methods and the Kullback-Leibler divergence between stochastic processes. In International Conference on Artificial Intelligence and Statistics.
- Continual multi-task Gaussian processes. arXiv.
- Variational continual learning. In International Conference on Learning Representations.
- Continual deep learning by functional regularisation of memorable past. In Advances in Neural Information Processing Systems.
- Continual lifelong learning with neural networks: a review. Neural Networks.
- Continual learning by asymmetric loss approximation with single-side overestimation. In International Conference on Computer Vision.
- Online structured Laplace approximations for overcoming catastrophic forgetting. In Advances in Neural Information Processing Systems.
- Rethinking function-space variational inference in Bayesian neural networks. In Symposium on Advances in Approximate Bayesian Inference.
- Sato, M.-A. (2001). Online model selection based on the variational Bayes. Neural Computation.
- Progress & compress: a scalable framework for continual learning. In International Conference on Machine Learning.
- The Mathematical Theory of Communication. University of Illinois Press, Urbana and Chicago.
- Functional variational Bayesian neural networks. In International Conference on Learning Representations.
- Improving and understanding variational continual learning. In NeurIPS Workshop on Continual Learning.
- Functional regularisation for continual learning with Gaussian processes. In International Conference on Learning Representations.
- Online sparse multi-output Gaussian process regression and learning. IEEE Transactions on Signal and Information Processing over Networks.
- SOLA: continual learning with second-order loss approximation. arXiv.
- Optimization and generalization of regularization-based continual learning: a loss approximation viewpoint. arXiv.
- Continual learning through synaptic intelligence. In International Conference on Machine Learning.
- Tim G. J. Rudner (38 papers)
- Freddie Bickford Smith (7 papers)
- Qixuan Feng (5 papers)
- Yee Whye Teh (162 papers)
- Yarin Gal (170 papers)