Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lightweight Unsupervised Federated Learning with Pretrained Vision Language Model (2404.11046v1)

Published 17 Apr 2024 in cs.AI, cs.CV, and cs.LG

Abstract: Federated learning aims to tackle the ``isolated data island" problem, where it trains a collective model from physically isolated clients while safeguarding the privacy of users' data. However, supervised federated learning necessitates that each client labels their data for training, which can be both time-consuming and resource-intensive, and may even be impractical for edge devices. Moreover, the training and transmission of deep models present challenges to the computation and communication capabilities of the clients. To address these two inherent challenges in supervised federated learning, we propose a novel lightweight unsupervised federated learning approach that leverages unlabeled data on each client to perform lightweight model training and communication by harnessing pretrained vision-LLMs, such as CLIP. By capitalizing on the zero-shot prediction capability and the well-trained image encoder of the pre-trained CLIP model, we have carefully crafted an efficient and resilient self-training approach. This method refines the initial zero-shot predicted pseudo-labels of unlabeled instances through the sole training of a linear classifier on top of the fixed image encoder. Additionally, to address data heterogeneity within each client, we propose a class-balanced text feature sampling strategy for generating synthetic instances in the feature space to support local training. Experiments are conducted on multiple benchmark datasets. The experimental results demonstrate that our proposed method greatly enhances model performance in comparison to CLIP's zero-shot predictions and even outperforms supervised federated learning benchmark methods given limited computational and communication overhead.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. AltCLIP: Altering the language encoder in CLIP for extended language capabilities. In Findings of the Association for Computational Linguistics: ACL 2023, 2023.
  2. Contrastive language and vision learning of general fashion concepts. Scientific Reports, 2022.
  3. Cinic-10 is not imagenet or cifar-10. arXiv preprint arXiv:1810.03505, 2018.
  4. Semifl: Semi-supervised federated learning for unlabeled clients with alternate training. In NeurIPS, 2022.
  5. Using language to extend to unseen domains. In ICLR, 2023.
  6. Fednew: A communication-efficient and privacy-preserving newton-type method for federated learning. In ICML, 2022.
  7. Open-vocabulary object detection via vision and language knowledge distillation. In ICLR, 2022.
  8. pfedprompt: Learning personalized prompt for vision-language models in federated learning. In Proceedings of the ACM Web Conference 2023, 2023a.
  9. Fedbr: Improving federated learning on heterogeneous data via local learning bias reduction. In ICML, 2023b.
  10. Measuring the effects of non-identical data distribution for federated visual classification. arXiv preprint arXiv:1909.06335, 2019.
  11. A visual–language foundation model for pathology image analysis using medical twitter. Nature Medicine, 2023.
  12. Federated semi-supervised learning with inter-client consistency & disjoint learning. In ICLR, 2021.
  13. Scaling up visual and vision-language representation learning with noisy text supervision. In ICML, 2021.
  14. Scaffold: Stochastic controlled averaging for federated learning. In ICML, 2020.
  15. Multi-level branched regularization for federated learning. In ICML, 2022.
  16. Learning multiple layers of features from tiny images. 2009.
  17. Preservation of the global knowledge by not-true distillation in federated learning. In NeurIPS, 2022.
  18. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In ICML, 2022.
  19. Model-contrastive federated learning. In CVPR, 2021.
  20. Federated optimization in heterogeneous networks. In MLSys, 2020.
  21. Federated learning with positive and unlabeled data. In ICML, 2022.
  22. Federated learning from only unlabeled data with class-conditional-sharing clients. In ICLR, 2022.
  23. Image segmentation using text and image prompts. In CVPR, 2022.
  24. No fear of heterogeneity: Classifier calibration for federated learning with non-iid data. In NeurIPS, 2021.
  25. Communication-efficient learning of deep networks from decentralized data. In AISTATS, 2017.
  26. Visual classification via description from large language models. In ICLR, 2023.
  27. Learning transferable visual models from natural language supervision. In ICML, 2021.
  28. Adaptive federated optimization. In ICLR, 2021.
  29. Flava: A foundational language and vision alignment model. In CVPR, 2022.
  30. Dynamic regularized sharpness aware minimization in federated learning: Approaching global consistency and smooth landscape. In ICML, 2023.
  31. Federated learning from pre-trained models: A contrastive learning approach. In NeurIPS, 2022.
  32. Attention is all you need. In NeurIPS, 2017.
  33. Tackling the objective inconsistency problem in heterogeneous federated optimization. In NeurIPS, 2020.
  34. Simvlm: Simple visual language model pretraining with weak supervision. In ICLR, 2022.
  35. Feddisco: Federated learning with discrepancy-aware collaboration. In ICML, 2023.
  36. Improving semi-supervised federated learning by reducing the gradient diversity of models. In 2021 IEEE International Conference on Big Data (Big Data), 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Hao Yan (109 papers)
  2. Yuhong Guo (52 papers)

Summary

We haven't generated a summary for this paper yet.