Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

t-DGR: A Trajectory-Based Deep Generative Replay Method for Continual Learning in Decision Making (2401.02576v2)

Published 4 Jan 2024 in cs.LG, cs.AI, and cs.NE

Abstract: Deep generative replay has emerged as a promising approach for continual learning in decision-making tasks. This approach addresses the problem of catastrophic forgetting by leveraging the generation of trajectories from previously encountered tasks to augment the current dataset. However, existing deep generative replay methods for continual learning rely on autoregressive models, which suffer from compounding errors in the generated trajectories. In this paper, we propose a simple, scalable, and non-autoregressive method for continual learning in decision-making tasks using a generative model that generates task samples conditioned on the trajectory timestep. We evaluate our method on Continual World benchmarks and find that our approach achieves state-of-the-art performance on the average success rate metric among continual learning methods. Code is available at https://github.com/WilliamYue37/t-DGR.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. Mark Ring. Continual Learning in Reinforcement Environments. PhD thesis, University of Texas at Austin, 1994.
  2. Catastrophic interference in connectionist networks: The sequential learning problem. volume 24 of Psychology of Learning and Motivation, pages 109–165. Academic Press, 1989.
  3. Complementary learning systems within the hippocampus: A neural network modeling approach to understanding episodic memory consolidation. Psychological Review, 102(3):419–457, 1995.
  4. What learning systems do intelligent agents need? complementary learning systems theory updated. Trends in Cognitive Sciences, 20(7):512–534, 2016.
  5. Why there are complementary learning systems in the hippocampus and neocortex: Insights from the successes and failures of connectionist models of learning and memory. Psychological review, 102:419–57, 08 1995.
  6. Randall C. O’Reilly and Kenneth A. Norman. Hippocampal and neocortical contributions to memory: Advances in the complementary learning systems framework. Trends in Cognitive Sciences, 6(12):505–510, December 2002.
  7. Continual learning with deep generative replay, 2017.
  8. Continual world: A robotic benchmark for continual reinforcement learning, 2021.
  9. A comprehensive survey of continual learning: Theory, method and application, 2023.
  10. Gradient based sample selection for online continual learning. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019.
  11. Rainbow memory: Continual learning with a memory of diverse samples, 2021.
  12. Re-evaluating continual learning scenarios: A categorization and case for strong baselines, 2019.
  13. Gido M. van de Ven and Andreas S. Tolias. Three scenarios for continual learning, 2019.
  14. Dark experience for general continual learning: a strong, simple baseline, 2020.
  15. Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences, 114(13):3521–3526, mar 2017.
  16. Memory aware synapses: Learning what (not) to forget, 2018.
  17. Variational continual learning, 2018.
  18. Progress & compress: A scalable framework for continual learning, 2018.
  19. Packnet: Adding multiple tasks to a single network by iterative pruning, 2018.
  20. Uncertainty-based continual learning with adaptive regularization, 2019.
  21. Continual learning with node-importance based adaptive group sparse regularization, 2021.
  22. Lifelong learning with dynamically expandable networks, 2018.
  23. Generative adversarial networks, 2014.
  24. Auto-encoding variational bayes, 2022.
  25. Denoising diffusion probabilistic models, 2020.
  26. Deep unsupervised learning using nonequilibrium thermodynamics. CoRR, abs/1503.03585, 2015.
  27. Donald J. Newman. The double dixie cup problem. The American Mathematical Monthly, 67(1):58–61, 1960.
  28. Some new aspects of the coupon-collector’s problem, 2003.
  29. Cril: Continual robot imitation learning via generative and prediction model, 2021.
  30. Generative adversarial imitation learning, 2016.
  31. Inverse reinforcement learning. In Proceedings of the 17th International Conference on Machine Learning (ICML-2000), pages 663–670, 2000.
  32. Improved denoising diffusion probabilistic models, 2021.
  33. U-net: Convolutional networks for biomedical image segmentation, 2015.
  34. Meta-world: A benchmark and evaluation for multi-task and meta reinforcement learning, 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. William Yue (8 papers)
  2. Bo Liu (485 papers)
  3. Peter Stone (184 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com