Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 165 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 24 tok/s Pro
GPT-4o 112 tok/s Pro
Kimi K2 208 tok/s Pro
GPT OSS 120B 466 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Federated Wasserstein Distance (2310.01973v1)

Published 3 Oct 2023 in cs.LG and cs.DC

Abstract: We introduce a principled way of computing the Wasserstein distance between two distributions in a federated manner. Namely, we show how to estimate the Wasserstein distance between two samples stored and kept on different devices/clients whilst a central entity/server orchestrates the computations (again, without having access to the samples). To achieve this feat, we take advantage of the geometric properties of the Wasserstein distance -- in particular, the triangle inequality -- and that of the associated {\em geodesics}: our algorithm, FedWad (for Federated Wasserstein Distance), iteratively approximates the Wasserstein distance by manipulating and exchanging distributions from the space of geodesics in lieu of the input samples. In addition to establishing the convergence properties of FedWad, we provide empirical results on federated coresets and federate optimal transport dataset distance, that we respectively exploit for building a novel federated model and for boosting performance of popular federated learning algorithms.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. Geometric approximation via coresets. Combinatorial and computational geometry, 52(1):1–30, 2005.
  2. Barycenters in the wasserstein space. SIAM Journal on Mathematical Analysis, 43(2):904–924, 2011.
  3. Geometric dataset distances via optimal transport. Advances in Neural Information Processing Systems, 33:21428–21439, 2020.
  4. Gradient flows: in metric spaces and in the space of probability measures. Springer Science & Business Media, 2005.
  5. Federated learning with personalization layers. arXiv preprint arXiv:1912.00818, 2019.
  6. Diffeomorphic density matching by optimal information transport. SIAM Journal on Imaging Sciences, 8(3):1718–1751, 2015.
  7. Wasserstein measure coresets. arXiv preprint arXiv:1805.07412, 2018.
  8. Exploiting Shared Representations for Personalized Federated Learning. In International Conference on Machine Learning, pp. 2089–2099, 2021.
  9. Learning wasserstein embeddings. In International Conference on Learning Representations (ICLR), 2018.
  10. On the rate of convergence in wasserstein distance of the empirical measure. Probability theory and related fields, 162(3-4):707–738, 2015.
  11. Sketching data sets for large-scale learning: Keeping only what you need. IEEE Signal Processing Magazine, 38(5):12–36, 2021.
  12. Manifold interpolating optimal-transport flows for trajectory inference. Advances in Neural Information Processing Systems, 35:29705–29718, 2022.
  13. Advances and Open Problems in Federated Learning. Foundations and Trends in Machine Learning, 14(1–2):1–210, 2021.
  14. Optimal mass transport: Signal processing and machine-learning applications. IEEE signal processing magazine, 34(4):43–59, 2017.
  15. Differentially private optimal transport: Application to domain adaptation. In IJCAI, pp.  2852–2858, 2019.
  16. Secure efficient federated knn for recommendation systems. In Advances in Natural Computation, Fuzzy Systems and Knowledge Discovery, pp.  1808–1819. Springer, 2021.
  17. Transport based image morphing with intensity modulation. In Scale Space and Variational Methods in Computer Vision: 6th International Conference, SSVM 2017, Kolding, Denmark, June 4-8, 2017, Proceedings, pp.  563–577. Springer, 2017.
  18. Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pp.  1273–1282. PMLR, 2017.
  19. On coresets for logistic regression. Advances in Neural Information Processing Systems, 31, 2018.
  20. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
  21. Computational optimal transport: With applications to data science. Foundations and Trends® in Machine Learning, 11(5-6):355–607, 2019.
  22. Jeff M Phillips. Coresets and sketches. arXiv preprint arXiv:1601.00617, 2016.
  23. Private nearest neighbors classification in federated databases. IACR Cryptol. ePrint Arch., 2018:289, 2018.
  24. Ulrike Von Luxburg. A tutorial on spectral clustering. Statistics and computing, 17:395–416, 2007.
  25. A Field Guide to Federated Optimization. arXiv preprint arXiv:2107.06917, 2021.
  26. Approximate k-nearest neighbor query over spatial data federation. In Database Systems for Advanced Applications: 28th International Conference, DASFAA 2023, Tianjin, China, April 17–20, 2023, Proceedings, Part I, pp.  351–368. Springer, 2023.
Citations (4)

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube