DiPrompT: Disentangled Prompt Tuning for Multiple Latent Domain Generalization in Federated Learning (2403.08506v1)
Abstract: Federated learning (FL) has emerged as a powerful paradigm for learning from decentralized data, and federated domain generalization further considers the test dataset (target domain) is absent from the decentralized training data (source domains). However, most existing FL methods assume that domain labels are provided during training, and their evaluation imposes explicit constraints on the number of domains, which must strictly match the number of clients. Because of the underutilization of numerous edge devices and additional cross-client domain annotations in the real world, such restrictions may be impractical and involve potential privacy leaks. In this paper, we propose an efficient and novel approach, called Disentangled Prompt Tuning (DiPrompT), a method that tackles the above restrictions by learning adaptive prompts for domain generalization in a distributed manner. Specifically, we first design two types of prompts, i.e., global prompt to capture general knowledge across all clients and domain prompts to capture domain-specific knowledge. They eliminate the restriction on the one-to-one mapping between source domains and local clients. Furthermore, a dynamic query metric is introduced to automatically search the suitable domain label for each sample, which includes two-substep text-image alignments based on prompt tuning without labor-intensive annotation. Extensive experiments on multiple datasets demonstrate that our DiPrompT achieves superior domain generalization performance over state-of-the-art FL methods when domain labels are not provided, and even outperforms many centralized learning methods using domain labels.
- Benchmarking algorithms for federated domain generalization. arXiv preprint arXiv:2307.04942, 2023.
- Multi-domain synchronous refinement network for unsupervised cross-domain person re-identification. In 2021 IEEE International Conference on Multimedia and Expo (ICME), pages 1–6, 2021a.
- Mfi: Multi-range feature interchange for video action recognition. In 2020 25th International Conference on Pattern Recognition (ICPR), pages 6664–6671. IEEE, 2021b.
- Combating data imbalances in federated semi-supervised learning with dual regulators. 2024.
- Domain generalization by mutual-information regularization with pre-trained models. In European Conference on Computer Vision, pages 440–457. Springer, 2022.
- Prompt learning with optimal transport for vision-language models. 2023.
- Learning to learn with variational information bottleneck for domain generalization. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part X 16, pages 200–216, 2020.
- Unbiased metric learning: On the utilization of multiple datasets and web images for softening bias. In Proceedings of the IEEE International Conference on Computer Vision, pages 1657–1664, 2013.
- Loss function learning for domain generalization by implicit gradient. In International Conference on Machine Learning, pages 7002–7016, 2022.
- Promptfl: Let federated participants cooperatively learn prompts instead of models–federated learning in age of foundation model. arXiv preprint arXiv:2208.11625, 2022.
- Learn from others and be yourself in heterogeneous federated learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10143–10153, 2022.
- Deeper, broader and artier domain generalization. In Proceedings of the IEEE international conference on computer vision, pages 5542–5550, 2017.
- Learning to generalize: Meta-learning for domain generalization. In Proceedings of the AAAI conference on artificial intelligence, 2018a.
- Model-contrastive federated learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10713–10722, 2021a.
- Groupformer: Group activity recognition with clustered spatial-temporal transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 13668–13677, 2021b.
- Probing visual-audio representation for video highlight detection via hard-pairs guided contrastive learning. arXiv preprint arXiv:2206.10157, 2022.
- Federated optimization in heterogeneous networks. Proceedings of Machine learning and systems, 2:429–450, 2020.
- Domain generalization via conditional invariant representations. In Proceedings of the AAAI conference on artificial intelligence, 2018b.
- Mitigating both covariate and conditional shift for domain generalization. In 2022 IEEE 8th International Conference on Cloud Computing and Intelligent Systems (CCIS), pages 437–443, 2022.
- Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35, 2023.
- Feddg: Federated domain generalization on medical image segmentation via episodic learning in continuous frequency space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1013–1023, 2021.
- Fedclip: Fast generalization and personalization for clip in federated learning. arXiv preprint arXiv:2302.13485, 2023.
- Domain generalization using a mixture of multiple latent domains. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 11749–11756, 2020.
- Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pages 1273–1282. PMLR, 2017.
- Attention diversification for domain generalization. In European conference on computer vision, pages 322–340. Springer, 2022.
- Domain invariant representation learning with domain density transformations. Advances in Neural Information Processing Systems, 34:5264–5275, 2021.
- Fedsr: A simple and effective domain generalization method for federated learning. Advances in Neural Information Processing Systems, 35:38831–38843, 2022.
- Language models as knowledge bases? arXiv preprint arXiv:1909.01066, 2019.
- E-bert: Efficient-yet-effective entity embeddings for bert. arXiv preprint arXiv:1911.03681, 2019.
- Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748–8763. PMLR, 2021.
- Fishr: Invariant gradient variances for out-of-distribution generalization. In International Conference on Machine Learning, pages 18347–18377, 2022.
- Autoprompt: Eliciting knowledge from language models with automatically generated prompts. pages 4222–4235, 2020.
- Clipood: Generalizing clip to out-of-distributions. arXiv preprint arXiv:2302.00864, 2023.
- Prototypical networks for few-shot learning. In Advances in Neural Information Processing Systems, pages 4077–4087, 2017.
- Gradient masked averaging for federated learning. arXiv preprint arXiv:2201.11986, 2022.
- Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5018–5027, 2017.
- Unsupervised domain adaptive learning via synthetic data for person re-identification. arXiv preprint arXiv:2109.05542, 2021.
- Federated machine learning: Concept and applications. ACM Transactions on Intelligent Systems and Technology (TIST), 10(2):1–19, 2019.
- Visual-language prompt tuning with knowledge-guided context optimization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6757–6767, 2023.
- Pcl: Proxy-based contrastive learning for domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7097–7107, 2022.
- Federated learning with domain generalization. arXiv preprint arXiv:2111.10487, 2021.
- Domain generalization with mixstyle. arXiv preprint arXiv:2104.02008, 2021.
- Domain generalization: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022a.
- Conditional prompt learning for vision-language models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16816–16825, 2022b.
- Learning to prompt for vision-language models. International Journal of Computer Vision, 130(9):2337–2348, 2022c.