Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
124 tokens/sec
GPT-4o
8 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Analysis of the Memorization and Generalization Capabilities of AI Agents: Are Continual Learners Robust? (2309.10149v2)

Published 18 Sep 2023 in cs.LG and cs.AI

Abstract: In continual learning (CL), an AI agent (e.g., autonomous vehicles or robotics) learns from non-stationary data streams under dynamic environments. For the practical deployment of such applications, it is important to guarantee robustness to unseen environments while maintaining past experiences. In this paper, a novel CL framework is proposed to achieve robust generalization to dynamic environments while retaining past knowledge. The considered CL agent uses a capacity-limited memory to save previously observed environmental information to mitigate forgetting issues. Then, data points are sampled from the memory to estimate the distribution of risks over environmental change so as to obtain predictors that are robust with unseen changes. The generalization and memorization performance of the proposed framework are theoretically analyzed. This analysis showcases the tradeoff between memorization and generalization with the memory size. Experiments show that the proposed algorithm outperforms memory-based CL baselines across all environments while significantly improving the generalization performance on unseen target environments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. D. Lopez-Paz and M. Ranzato, “Gradient episodic memory for continual learning,” in Advances in neural information processing systems, 2017, vol. 30.
  2. “Edge continual learning for dynamic digital twins over wireless networks,” in IEEE 23rd International Workshop on Signal Processing Advances in Wireless Communication (SPAWC), Oulu, Finland, 2022, pp. 1–5.
  3. “Probable domain generalization via quantile risk minimization,” in Advances in Neural Information Processing Systems, 2022, pp. 17340–17358.
  4. “Experience replay for continual learning,” in Advances in Neural Information Processing Systems, 2019.
  5. “Dark experience for general continual learning: a strong, simple baseline,” in Advances in neural information processing systems, 2020, pp. 15920–15930.
  6. “Gradient based sample selection for online continual learning,” in Advances in neural information processing systems, 2019.
  7. “Using hindsight to anchor past knowledge in continual learning,” in Proceedings of the AAAI conference on artificial intelligence, 2021, vol. 35, pp. 6993–7001.
  8. “Information-theoretic online memory selection for continual learning,” arXiv preprint arXiv:2204.04763, 2022.
  9. “Progress & compress: A scalable framework for continual learning,” in International conference on machine learning, 2018, pp. 4528–4537.
  10. “Looking back on learned experiences for class/task incremental learning,” in International Conference on Learning Representations, 2022.
  11. “The ideal continual learner: An agent that never forgets,” in International Conference on Machine Learning, 2023, pp. 27585–27610.
  12. K. Raghavan and P. Balaprakash, “Formalizing the generalization-forgetting trade-off in continual learning,” in Advances in Neural Information Processing Systems, 2021, pp. 17284–17297.
  13. “Theory on forgetting and generalization of continual learning,” arXiv preprint arXiv:2302.05836, 2023.
  14. “Online continual learning with natural distribution shifts: An empirical study with visual data,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 8281–8290.
  15. “Continual active adaptation to evolving distributional shifts,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 3444–3450.
  16. “Out-of-distribution forgetting: vulnerability of continual learning to intra-class distribution shift,” arXiv preprint arXiv:2306.00427, 2023.
  17. “Stochastic convex optimization.,” in COLT, 2009, vol. 2, pp. 1–5.
  18. “Optimal continual learning has perfect memory and is np-hard,” in International Conference on Machine Learning, 2020, pp. 5327–5337.
  19. “Communication-efficient learning of deep networks from decentralized data,” in Artificial intelligence and statistics, 2017, pp. 1273–1282.
  20. “Scaffold: Stochastic controlled averaging for federated learning,” in International conference on machine learning, 2020, pp. 5132–5143.
Citations (1)

Summary

We haven't generated a summary for this paper yet.