Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CAESAR: Enhancing Federated RL in Heterogeneous MDPs through Convergence-Aware Sampling with Screening (2403.20156v2)

Published 29 Mar 2024 in cs.LG and cs.AI

Abstract: In this study, we delve into Federated Reinforcement Learning (FedRL) in the context of value-based agents operating across diverse Markov Decision Processes (MDPs). Existing FedRL methods typically aggregate agents' learning by averaging the value functions across them to improve their performance. However, this aggregation strategy is suboptimal in heterogeneous environments where agents converge to diverse optimal value functions. To address this problem, we introduce the Convergence-AwarE SAmpling with scReening (CAESAR) aggregation scheme designed to enhance the learning of individual agents across varied MDPs. CAESAR is an aggregation strategy used by the server that combines convergence-aware sampling with a screening mechanism. By exploiting the fact that agents learning in identical MDPs are converging to the same optimal value function, CAESAR enables the selective assimilation of knowledge from more proficient counterparts, thereby significantly enhancing the overall learning efficiency. We empirically validate our hypothesis and demonstrate the effectiveness of CAESAR in enhancing the learning efficiency of agents, using both a custom-built GridWorld environment and the classical FrozenLake-v1 task, each presenting varying levels of environmental heterogeneity.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. Openai gym. In arXiv preprint arXiv:1606.01540.
  2. Chapter 14 - Federated sequential decision making: Bayesian optimization, reinforcement learning, and beyond. In Federated Learning, Lam M. Nguyen, Trong Nghia Hoang, and Pin-Yu Chen (Eds.). Academic Press, 257–279. https://doi.org/10.1016/B978-0-44-319037-7.00023-5
  3. Fault-Tolerant Federated Reinforcement Learning with Theoretical Guarantee. In 35th Conference on Neural Information Processing Systems (NeurIPS).
  4. FedHQL: Federated Heterogeneous Q-Learning. arXiv:2301.11135.
  5. Federated Reinforcement Learning for the Building Facilities. In 2022 IEEE International Conference on Omni-layer Intelligent Systems (COINS). 1–6. https://doi.org/10.1109/COINS54846.2022.9854959
  6. Federated reinforcement learning with environment heterogeneity. In International Conference on Artificial Intelligence and Statistics (AISTATS).
  7. Watkins Christopher JCH and Peter Dayan. 1992. Q-learning. In Machine learning 8.
  8. Decentralized Federated Policy Gradient with Byzantine Fault-Tolerance and Provably Fast Convergence. (2024). https://arxiv.org/abs/2401.03489
  9. Federated Reinforcement Learning: Linear Speedup Under Markovian Sampling. In Proceedings of the 39th International Conference on Machine Learning (ICML).
  10. Federated transfer reinforcement learning for autonomous driving. In Federated and Transfer Learning. Springer, 357–371.
  11. Lifelong Federated Reinforcement Learning: A Learning Architecture for Navigation in Cloud Robotic Systems. IEEE Robotics and Automation Letters 4, 4 (2019), 4555–4562. https://doi.org/10.1109/LRA.2019.2931179
  12. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS).
  13. Federated Reinforcement Learning for Fast Personalization. In 2019 IEEE Second International Conference on Artificial Intelligence and Knowledge Engineering (AIKE). 123–127. https://doi.org/10.1109/AIKE.2019.00031
  14. Federated reinforcement learning: Techniques, applications, and open challenges. (2021).
  15. Towards Understanding Asynchronous Advantage Actor-critic: Convergence and Linear Speedup. IEEE Transactions on Signal Processing (2023).
  16. Richard Sutton and Andrew Barto. 2018. Reinforcement learning: An introduction. MIT press.
  17. Federated deep reinforcement learning for Internet of Things with decentralized cooperative edge caching. IEEE Internet of Things Journal 7, 10 (2020), 9441–9455.
  18. The Blessing of Heterogeneity in Federated Q-Learning: Linear Speedup and Beyond. In Proceedings of the 40th International Conference on Machine Learning.
  19. A resource-constrained and privacy-preserving edge-computing-enabled clinical decision system: A federated reinforcement learning approach. IEEE Internet of Things Journal 8, 11 (2021), 9122–9138.
  20. Federated Learning. Synthesis Lectures on Artificial Intelligence and Machine Learning (2020), 121–131.
  21. When Deep Reinforcement Learning Meets Federated Learning: Intelligent Multitimescale Resource Management for Multiaccess Edge Computing in 5G Ultradense Network. IEEE Internet of Things Journal 8, 4 (2020), 2238–2251.
  22. Resilient Mechanism Against Byzantine Failure for Distributed Deep Reinforcement Learning. In 2022 IEEE 33rd International Symposium on Software Reliability Engineering (ISSRE). IEEE, 378–389.
  23. Federated deep reinforcement learning. arXiv:1901.08277.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets