Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Cross-Domain Continual Learning (2402.12490v1)

Published 19 Feb 2024 in cs.LG, cs.AI, and cs.CV

Abstract: Continual learning is a process that involves training learning agents to sequentially master a stream of tasks or classes without revisiting past data. The challenge lies in leveraging previously acquired knowledge to learn new tasks efficiently, while avoiding catastrophic forgetting. Existing methods primarily focus on single domains, restricting their applicability to specific problems. In this work, we introduce a novel approach called Cross-Domain Continual Learning (CDCL) that addresses the limitations of being limited to single supervised domains. Our method combines inter- and intra-task cross-attention mechanisms within a compact convolutional network. This integration enables the model to maintain alignment with features from previous tasks, thereby delaying the data drift that may occur between tasks, while performing unsupervised cross-domain (UDA) between related domains. By leveraging an intra-task-specific pseudo-labeling method, we ensure accurate input pairs for both labeled and unlabeled samples, enhancing the learning process. To validate our approach, we conduct extensive experiments on public UDA datasets, showcasing its positive performance on cross-domain continual learning challenges. Additionally, our work introduces incremental ideas that contribute to the advancement of this field. We make our code and models available to encourage further exploration and reproduction of our results: \url{https://github.com/Ivsucram/CDCL}

Definition Search Book Streamline Icon: https://streamlinehq.com
References (53)
  1. Gradient Based Sample Selection for Online Continual Learning. Curran Associates Inc., Red Hook, NY, USA, 2019.
  2. A theory of learning from different domains. Mach. Learn., 79(1–2):151–175, May 2010.
  3. Class-incremental continual learning into the extended der-verse. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  4. Dark experience for general continual learning: a strong, simple baseline. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 15920–15930. Curran Associates, Inc., 2020.
  5. From babies to robots: The contribution of developmental robotics to developmental psychology. Child Development Perspectives, 12(3):183–188, Sept. 2018.
  6. The art of adaptive pattern recognition by a self-organizing neural network. Computer, 21(3):77–88, 1988.
  7. Riemannian walk for incremental learning: Understanding forgetting and intransigence. In Computer Vision – ECCV 2018: 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part XI, page 556–572, Berlin, Heidelberg, 2018. Springer-Verlag.
  8. Using hindsight to anchor past knowledge in continual learning, 2020.
  9. Efficient lifelong learning with a-gem. ArXiv, abs/1812.00420, 2019.
  10. Crdoco: Pixel-level domain transfer with cross-domain consistency. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
  11. Effective pseudo-labeling based on heatmap for unsupervised domain adaptation in cell detection. Medical Image Analysis, 79:102436, 2022.
  12. Towards unsupervised domain adaptation via domain-transformer, 2022.
  13. Coatnet: Marrying convolution and attention for all data sizes. arXiv preprint arXiv:2106.04803, 2021.
  14. Acdc: Online unsupervised cross-domain adaptation, 2021.
  15. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 248–255, 2009.
  16. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021.
  17. Levit: A vision transformer in convnet’s clothing for faster inference. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 12259–12269, October 2021.
  18. Neuroscience-inspired artificial intelligence. Neuron, 95(2):245–258, 2017.
  19. Escaping the big data paradigm with compact transformers. 2021.
  20. J. J. Hull. A database for handwritten text recognition research. IEEE Transactions on Pattern Analysis and Machine Intelligence, 16(5):550–554, 1994.
  21. Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences, 114(13):3521–3526, 2017.
  22. A two-stream continual learning system with variational domain-agnostic feature replay. IEEE Transactions on Neural Networks and Learning Systems, 33(9):4466–4478, 2022.
  23. MNIST handwritten digit database. ATT Labs, 2010.
  24. Dong-Hyun Lee. Pseudo-label : The simple and efficient semi-supervised learning method for deep neural networks. ICML 2013 Workshop : Challenges in Representation Learning (WREPL), 07 2013.
  25. Cross-domain adaptive clustering for semi-supervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 2505–2514, June 2021.
  26. Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation. In Proceedings of the 37th International Conference on Machine Learning, ICML’20. JMLR.org, 2020.
  27. Robust training under label noise by over-parameterization. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato, editors, Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 14153–14172. PMLR, 17–23 Jul 2022.
  28. Gradient episodic memory for continual learning. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
  29. Decoupled weight decay regularization. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
  30. Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization. Proceedings of the National Academy of Sciences, 115(44):E10467–E10475, 2018.
  31. Augmentation strategies for learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 8022–8031, June 2021.
  32. Transfer learning via dimensionality reduction. In Proceedings of the 23rd National Conference on Artificial Intelligence - Volume 2, AAAI’08, page 677–682. AAAI Press, 2008.
  33. Moment matching for multi-source domain adaptation. In Proceedings of the IEEE International Conference on Computer Vision, pages 1406–1415, 2019.
  34. Visda: The visual domain adaptation challenge, 2017.
  35. Progressive neural networks. ArXiv, abs/1606.04671, 2016.
  36. Adapting visual category models to new domains. volume 6314, pages 213–226, 09 2010.
  37. Progress andamp; compress: A scalable framework for continual learning. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 4528–4537. PMLR, 10–15 Jul 2018.
  38. Continual learning with deep generative replay. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
  39. On generalizing beyond domains in cross-domain continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9265–9274, June 2022.
  40. Development of multisensory integration from the perspective of the individual neuron. Nature reviews. Neuroscience, 15:520–35, 07 2014.
  41. Lifelong robot learning. Robotics and Autonomous Systems, 15(1):25–46, 1995. The Biology and Technology of Intelligent Autonomous Agents.
  42. Gido M van de Ven and Andreas S Tolias. Generative replay with feedback connections as a general strategy for continual learning. arXiv preprint arXiv:1809.10635, 2018.
  43. Scaling local self-attention for parameter efficient visual backbones, 2021.
  44. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, page 6000–6010, Red Hook, NY, USA, 2017. Curran Associates Inc.
  45. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5018–5027, 2017.
  46. Symmetric cross entropy for robust learning with noisy labels. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pages 322–330, Los Alamitos, CA, USA, nov 2019. IEEE Computer Society.
  47. Continual learning with lifelong vision transformer. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 171–181, 2022.
  48. Autonomous cross domain adaptation under extreme label scarcity. IEEE Transactions on Neural Networks and Learning Systems, pages 1–12, 2022.
  49. Cdtrans: Cross-domain transformer for unsupervised domain adaptation, 2021.
  50. TVT: transferable vision transformer for unsupervised domain adaptation. CoRR, abs/2108.05988, 2021.
  51. Atpl: Mutually enhanced adversarial training and pseudo labeling for unsupervised domain adaptation. Knowledge-Based Systems, 250:108831, 2022.
  52. Continual learning through synaptic intelligence. In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 3987–3995. PMLR, 06–11 Aug 2017.
  53. Ca-uda: Class-aware unsupervised domain adaptation with optimal assignment and pseudo-label refinement, 2022.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Marcus de Carvalho (7 papers)
  2. Mahardhika Pratama (59 papers)
  3. Jie Zhang (847 papers)
  4. Chua Haoyan (1 paper)
  5. Edward Yapp (2 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets