Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

FedET: A Communication-Efficient Federated Class-Incremental Learning Framework Based on Enhanced Transformer (2306.15347v1)

Published 27 Jun 2023 in cs.LG and cs.AI

Abstract: Federated Learning (FL) has been widely concerned for it enables decentralized learning while ensuring data privacy. However, most existing methods unrealistically assume that the classes encountered by local clients are fixed over time. After learning new classes, this assumption will make the model's catastrophic forgetting of old classes significantly severe. Moreover, due to the limitation of communication cost, it is challenging to use large-scale models in FL, which will affect the prediction accuracy. To address these challenges, we propose a novel framework, Federated Enhanced Transformer (FedET), which simultaneously achieves high accuracy and low communication cost. Specifically, FedET uses Enhancer, a tiny module, to absorb and communicate new knowledge, and applies pre-trained Transformers combined with different Enhancers to ensure high precision on various tasks. To address local forgetting caused by new classes of new tasks and global forgetting brought by non-i.i.d (non-independent and identically distributed) class imbalance across different local clients, we proposed an Enhancer distillation method to modify the imbalance between old and new knowledge and repair the non-i.i.d. problem. Experimental results demonstrate that FedET's average accuracy on representative benchmark datasets is 14.1% higher than the state-of-the-art method, while FedET saves 90% of the communication cost compared to the previous method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. SS-IL: separated softmax for incremental learning. In 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, October 10-17, 2021 Ahn et al. [2021], pages 824–833.
  2. Do deep nets really need to be deep? In Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada Ba and Caruana [2014], pages 2654–2662.
  3. Mixtext: Linguistically-informed interpolation of hidden space for semi-supervised text classification. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020 Chen et al. [2020], pages 2147–2157.
  4. Episodic memory in lifelong language learning. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada Lange et al. [2022], pages 13122–13131.
  5. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA Deng et al. [2009], pages 248–255.
  6. Federated class-incremental learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022 Dong et al. [2022], pages 10154–10163.
  7. No one left behind: Real-world federated class-incremental learning. abs/2302.00903, 2023.
  8. Podnet: Pooled outputs distillation for small-tasks incremental learning. In Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part XX Douillard et al. [2020], pages 86–102.
  9. Memory efficient continual learning for neural text classification. abs/2203.04640, 2022.
  10. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016 He et al. [2016], pages 770–778.
  11. Masked autoencoders are scalable vision learners. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022 He et al. [2022], pages 15979–15988.
  12. Federated reconnaissance: Efficient, distributed, class-incremental learning. abs/2109.00150, 2021.
  13. Federated learning with dynamic transformer for text to speech. In Interspeech 2021, 22nd Annual Conference of the International Speech Communication Association, Brno, Czechia, 30 August - 3 September 2021 Hong et al. [2021], pages 3590–3594.
  14. Distilling causal effect of data in class-incremental learning. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, virtual, June 19-25, 2021 Hu et al. [2021], pages 3957–3966.
  15. Continual learning for text classification with information disentanglement based regularization. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online, June 6-11, 2021 Huang et al. [2021], pages 2736–2746.
  16. Learning multiple layers of features from tiny images. Citeseer, 2009.
  17. Recurrent convolutional neural networks for text classification. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, January 25-30, 2015, Austin, Texas, USA Lai et al. [2015], pages 2267–2273.
  18. A continual learning survey: Defying forgetting in classification tasks. 44(7):3366–3385, 2022.
  19. Mnemonics training: Multi-class incremental learning without forgetting. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020 Liu et al. [2020], pages 12242–12251.
  20. Decoupled weight decay regularization. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019 Loshchilov and Hutter [2019].
  21. Quantization and knowledge distillation for efficient federated learning on edge devices. In 22nd IEEE International Conference on High Performance Computing and Communications, HPCC 2020, Yanuca Island, December 14-16, 2020 Qu et al. [2020], pages 967–972.
  22. icarl: Incremental classifier and representation learning. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017 Rebuffi et al. [2017], pages 5533–5542.
  23. Adapterdrop: On the efficiency of adapters in transformers. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021 Rücklé et al. [2021], pages 7930–7946.
  24. Incremental learning of object detectors without catastrophic forgetting. In IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, October 22-29, 2017 Shmelkov et al. [2017], pages 3420–3429.
  25. On learning the geodesic path for incremental learning. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, virtual, June 19-25, 2021 Simon et al. [2021], pages 1591–1600.
  26. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA Tan and Le [2019], pages 6105–6114.
  27. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, EMNLP 2020 - Demos, Online, November 16-20, 2020 Wolf et al. [2020], pages 38–45.
  28. Large scale incremental learning. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019 Wu et al. [2019], pages 374–382.
  29. Federated Learning. Synthesis Lectures on Artificial Intelligence and Machine Learning. Morgan & Claypool Publishers, 2019.
  30. Learning and evaluating general linguistic intelligence. abs/1901.11373, 2019.
  31. Federated continual learning with weighted inter-client transfer. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event Yoon et al. [2021], pages 12073–12086.
  32. Class-incremental learning via deep model consolidation. In IEEE Winter Conference on Applications of Computer Vision, WACV 2020, Snowmass Village, CO, USA, March 1-5, 2020 Zhang et al. [2020], pages 1120–1129.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Chenghao Liu (61 papers)
  2. Xiaoyang Qu (41 papers)
  3. Jianzong Wang (144 papers)
  4. Jing Xiao (267 papers)
Citations (23)