Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning a Low-Rank Feature Representation: Achieving Better Trade-Off between Stability and Plasticity in Continual Learning (2312.08740v1)

Published 14 Dec 2023 in cs.LG and cs.CV

Abstract: In continual learning, networks confront a trade-off between stability and plasticity when trained on a sequence of tasks. To bolster plasticity without sacrificing stability, we propose a novel training algorithm called LRFR. This approach optimizes network parameters in the null space of the past tasks' feature representation matrix to guarantee the stability. Concurrently, we judiciously select only a subset of neurons in each layer of the network while training individual tasks to learn the past tasks' feature representation matrix in low-rank. This increases the null space dimension when designing network parameters for subsequent tasks, thereby enhancing the plasticity. Using CIFAR-100 and TinyImageNet as benchmark datasets for continual learning, the proposed approach consistently outperforms state-of-the-art methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (17)
  1. “Understanding the role of training regimes in continual learning,” in Proc. Adv. Neural Inf. Process. Syst, 2020, vol. 33, pp. 7308–7320.
  2. “Reconciling meta-learning and continual learning with online mixtures of tasks,” in Proc. Adv. Neural Inf. Process. Syst, 2019, vol. 32, pp. 9122–9133.
  3. “Packnet: Adding multiple tasks to a single network by iterative pruning,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., 2018, pp. 7765–7773.
  4. “Selective experience replay for lifelong learning,” in Proc. AAAI Conf. Artif. Intell., 2018, pp. 3302–3309.
  5. “Experience replay for continual learning,” in Proc. Adv. Neural Inf. Process. Systs., 2019, vol. 32, pp. 350–360.
  6. “Overcoming catastrophic forgetting in neural networks,” in Proc. Nat. Acad. Sci., 2017, vol. 114, pp. 3521–3526.
  7. “Gradient episodic memory for continual learning,” in Proc. Adv. Neural Inf. Process. Systs., 2017, vol. 30, pp. 6467–6476.
  8. “Gradient projection memory for continual learning,” in Proc. Int. Conf. Learn. Represent., 2021.
  9. “Training networks in null space of feature covariance for continual learning,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., 2021, pp. 184–193.
  10. “Efficient lifelong learning with a-GEM,” in Proc. Int. Conf. Learn. Represent., 2019.
  11. “Balancing stability and plasticity through advanced null space in continual learning,” in Proc. Eur. Conf. Comput. Vis., 2022, pp. 219–236.
  12. “Rethinking the value of network pruning,” in Proc. Int. Conf. Learn. Represent., 2019.
  13. “The lottery ticket hypothesis: Finding sparse, trainable neural networks,” in Proc. Int. Conf. Learn. Represent., 2019.
  14. “Learning efficient convolutional networks through network slimming,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 2736–2744.
  15. “Memory aware synapses: Learning what (not) to forget,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 139–154.
  16. “Improved schemes for episodic memory-based lifelong learning,” in Proc. Adv. Neural Inf. Process. Systs., 2020, vol. 33, pp. 1023–1035.
  17. “Continual learning of context-dependent processing in neural networks,” Nat. Mach. Intell., vol. 1, no. 8, pp. 364–372, 2019.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Zhenrong Liu (5 papers)
  2. Yang Li (1142 papers)
  3. Yi Gong (53 papers)
  4. Yik-Chung Wu (79 papers)