Learning a Low-Rank Feature Representation: Achieving Better Trade-Off between Stability and Plasticity in Continual Learning (2312.08740v1)
Abstract: In continual learning, networks confront a trade-off between stability and plasticity when trained on a sequence of tasks. To bolster plasticity without sacrificing stability, we propose a novel training algorithm called LRFR. This approach optimizes network parameters in the null space of the past tasks' feature representation matrix to guarantee the stability. Concurrently, we judiciously select only a subset of neurons in each layer of the network while training individual tasks to learn the past tasks' feature representation matrix in low-rank. This increases the null space dimension when designing network parameters for subsequent tasks, thereby enhancing the plasticity. Using CIFAR-100 and TinyImageNet as benchmark datasets for continual learning, the proposed approach consistently outperforms state-of-the-art methods.
- “Understanding the role of training regimes in continual learning,” in Proc. Adv. Neural Inf. Process. Syst, 2020, vol. 33, pp. 7308–7320.
- “Reconciling meta-learning and continual learning with online mixtures of tasks,” in Proc. Adv. Neural Inf. Process. Syst, 2019, vol. 32, pp. 9122–9133.
- “Packnet: Adding multiple tasks to a single network by iterative pruning,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., 2018, pp. 7765–7773.
- “Selective experience replay for lifelong learning,” in Proc. AAAI Conf. Artif. Intell., 2018, pp. 3302–3309.
- “Experience replay for continual learning,” in Proc. Adv. Neural Inf. Process. Systs., 2019, vol. 32, pp. 350–360.
- “Overcoming catastrophic forgetting in neural networks,” in Proc. Nat. Acad. Sci., 2017, vol. 114, pp. 3521–3526.
- “Gradient episodic memory for continual learning,” in Proc. Adv. Neural Inf. Process. Systs., 2017, vol. 30, pp. 6467–6476.
- “Gradient projection memory for continual learning,” in Proc. Int. Conf. Learn. Represent., 2021.
- “Training networks in null space of feature covariance for continual learning,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., 2021, pp. 184–193.
- “Efficient lifelong learning with a-GEM,” in Proc. Int. Conf. Learn. Represent., 2019.
- “Balancing stability and plasticity through advanced null space in continual learning,” in Proc. Eur. Conf. Comput. Vis., 2022, pp. 219–236.
- “Rethinking the value of network pruning,” in Proc. Int. Conf. Learn. Represent., 2019.
- “The lottery ticket hypothesis: Finding sparse, trainable neural networks,” in Proc. Int. Conf. Learn. Represent., 2019.
- “Learning efficient convolutional networks through network slimming,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 2736–2744.
- “Memory aware synapses: Learning what (not) to forget,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 139–154.
- “Improved schemes for episodic memory-based lifelong learning,” in Proc. Adv. Neural Inf. Process. Systs., 2020, vol. 33, pp. 1023–1035.
- “Continual learning of context-dependent processing in neural networks,” Nat. Mach. Intell., vol. 1, no. 8, pp. 364–372, 2019.
- Zhenrong Liu (5 papers)
- Yang Li (1142 papers)
- Yi Gong (53 papers)
- Yik-Chung Wu (79 papers)