Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Decentralized Learning Strategies for Estimation Error Minimization with Graph Neural Networks (2404.03227v2)

Published 4 Apr 2024 in eess.SP and cs.LG

Abstract: We address the challenge of sampling and remote estimation for autoregressive Markovian processes in a multi-hop wireless network with statistically-identical agents. Agents cache the most recent samples from others and communicate over wireless collision channels governed by an underlying graph topology. Our goal is to minimize time-average estimation error and/or age of information with decentralized scalable sampling and transmission policies, considering both oblivious (where decision-making is independent of the physical processes) and non-oblivious policies (where decision-making depends on physical processes). We prove that in oblivious policies, minimizing estimation error is equivalent to minimizing the age of information. The complexity of the problem, especially the multi-dimensional action spaces and arbitrary network topologies, makes theoretical methods for finding optimal transmission policies intractable. We optimize the policies using a graphical multi-agent reinforcement learning framework, where each agent employs a permutation-equivariant graph neural network architecture. Theoretically, we prove that our proposed framework exhibits desirable transferability properties, allowing transmission policies trained on small- or moderate-size networks to be executed effectively on large-scale topologies. Numerical experiments demonstrate that (i) Our proposed framework outperforms state-of-the-art baselines; (ii) The trained policies are transferable to larger networks, and their performance gains increase with the number of agents; (iii) The training procedure withstands non-stationarity even if we utilize independent learning techniques; and, (iv) Recurrence is pivotal in both independent learning and centralized training and decentralized execution, and improves the resilience to non-stationarity in independent learning.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (57)
  1. X. Chen, X. Liao and S. Saeedi-Bidokhti, “Real-time sampling and estimation on random access channels: Age of information and beyond,” in IEEE International Conference on Computer Communications (INFOCOM), 2021.
  2. R. V. Ramakanth, V. Tripathi, and E. Modiano, “Monitoring Correlated Sources: AoI-based Scheduling is Nearly Optimal,” arXiv, 2023.
  3. Y. Sun, Y. Polyanskiy, and E. Uysal-Biyikoglu, “Sampling of the Wiener process for remote estimation over a channel with random delay,” IEEE Transactions on Information Theory, vol. 66, no. 2, pp. 1118 – 1135, 2020.
  4. T. Z. Ornee and Y. Sun, “Sampling for remote estimation for Ornstein-Uhlenbeck process through queues: age of information and beyond,” IEEE/ACM Transactions on Networking, vol. 29, no. 5, pp. 1962 – 1975, 2021.
  5. S. Kaul, R. D. Yates, and M. Gruteser, “Real-time status: How often should one update?” in IEEE International Conference on Computer Communications (INFOCOM), 2012.
  6. Y. Sun, Y. Polyanskiy, and E. Uysal-Biyikoglu, “Remote Estimation of the Wiener Process over a Channel with Random Delay,” IEEE Transactions on Information Theory, vol. 66, no. 2, pp. 1118 – 1135, 2020.
  7. R. Talak and E. Modiano, “Age-Delay Tradeoffs in Queueing Systems,” IEEE Transactions on Information Theory, vol. 67, no. 3, pp. 1743 – 1758, 2021.
  8. X. Chen, K. Gatsis, H. Hassani and S. Saeedi-Bidokhti, “Age of information in random access channels,” IEEE Transactions on Information Theory, vol. 68, no. 10, pp. 6548 – 6568, 2022.
  9. O. T. Yavascan and E. Uysal, “Analysis of slotted ALOHA with an age threshold,” IEEE Journal on Selected Areas in Communications, vol. 39, no. 5, pp. 1456 – 1470, 2021.
  10. S. Kang, A. Eryilmaz, and N. B. Shroff, “Remote Tracking of Distributed Dynamic Sources Over a Random Access Channel With One-Bit Updates,” IEEE Transactions on Network Science and Engineering, vol. 10, no. 4, pp. 1931 – 1941, 2023.
  11. V. Tripathi, R. Talak, and E. Modiano, “Information freshness in multihop wireless networks,” IEEE Transactions on Information Theory, vol. 31, no. 2, pp. 784 – 799, 2023.
  12. N. Jones and E. Modiano, “Minimizing age of information in spatially distributed random access wireless networks,” arXiv, Dec 2022.
  13. G. Papoudakis, F. Christinos, L. Schafer, and etc, “Benchmarking multi-agent deep reinforcement learning algorithms in cooperative tasks,” in Advances in Neural Information Processing Systems, 2021.
  14. F. Gama, J. Bruna, and A. Ribeiro, “Stability properties of graph neural networks,” IEEE Transactions on Signal Processing, vol. 68, pp. 5680 – 5695, 2020.
  15. L. Ruiz, F. Gama, and A. Ribeiro, “Gated Graph Recurrent Neural Networks,” IEEE Transactions on Signal Processing, vol. 68, pp. 6303 – 6318, 2020.
  16. S. Kaul, M. Gruteser, V. Rai, et. al, “Minimizing age of information in vehicular networks,” 2011 8th Annual IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc Communications and Networks, 2011.
  17. I. Kadota and E. Modiano, “Minimizing age of information in wireless networks with stochastic arrivals,” IEEE Transactions on Mobile Computing, vol. 20, no. 3, pp. 1173 – 1185, 2021.
  18. R. D. Yates and S. K. Kaul, “The age of information: real-time status updating by multiple sources,” IEEE Transactions on Information Theory, vol. 65, no. 3, pp. 1807 – 1827, 2019.
  19. I. Kadota and E. Modiano, “Age of information in random access networks with stochastic arrivals,” in IEEE International Conference on Computer Communications, 2021.
  20. S. Farazi, A. G. Klein and D. R. Brown, ““Fundamental bounds on the age of information in general multi-hop interference networks,” in IEEE Conference on Computer Communications Workshops, 2019.
  21. B. Buyukates, A. Soysal, and S. Ulukus, “Age of information in multihop multicast networks,” Journal of Communications and Networks, vol. 21, no. 3, pp. 256 – 267, 2019.
  22. V. Tripathi, R. Talak, and E. Modiano, “Information freshness in multihop wireless networks,” IEEE/ACM Transactions on Networking, vol. 31, no. 2, pp. 784 – 799, 2022.
  23. E. Tolstaya, L. Butler, D. Mox, et. al, “Learning connectivity for data distribution in robot teams,” in IEEE/RSJ International Conference on Intelligent Robots and Systems, 2021.
  24. Y. Sun, Y. Polyanskiy, and E. Uysal-Biyikoglu, “Remote estimation of the Wiener process over a channel with random delay,” in IEEE International Symposium on Information Theory, 2017.
  25. T. Z. Ornee and Y. Sun, “Performance bounds for sampling and remote estimation of Gauss-Markov processes over a noisy channel with random delay,” in IEEE International Workshop on Signal Processing Advances in Wireless Communications, 2021.
  26. C. Tsai and C. Wang, “Unifying AoI Minimization and Remote Estimation—Optimal Sensor/Controller Coordination With Random Two-Way Delay,” IEEE/ACM Transactions on Networking, vol. 30, no. 1, pp. 229 – 242, 2022.
  27. C. Kam, S. Kompella, G. D. Nguyen, et. al, “Towards an Effective Age of Information: Remote Estimation of a Markov Source,” in IEEE Conference on Computer Communications Workshops, 2018.
  28. A. Maatouk, S. Kriouile, M. Assaad, et. al, “The Age of Incorrect Information: A New Performance Metric for Status Updates,” IEEE/ACM Transactions on Networking, vol. 28, no. 5, pp. 2215 – 2228, 2020.
  29. S. Saha, H. Makkar, V. Sukumaran, et. al, “On the Relationship Between Mean Absolute Error and Age of Incorrect Information in the Estimation of a Piecewise Linear Signal Over Noisy Channels,” IEEE Communications Letters, vol. 26, no. 11, pp. 2576 – 2580, 2022.
  30. P. Hernandez-Leal, B. Kartal and M. E. Taylor, “A survey and critique of multiagent deep reinforcement learning,” in International Conference on Autonomous Agents and Multi-Agent Systems, 2019.
  31. M. Tan, “Multi-agent reinforcement learning: Independent vs. cooperative agents,” in International Conference on Machine Learning, 1993.
  32. R. Lowe, Y. Wu, A. Tamar, and etc, “Multi-agent actor-critic for mixed cooperative-competitive environments,” in Advances in Neural Information Processing Systems, 2017.
  33. P. Sunehag, G. Lever, A. Gruslys, and etc, “Value-decomposition networks for cooperative multi-agent learning,” in International Conference on Autonomous Agents and Multi-Agent Systems, 2018.
  34. A. Feriani and E. Hossain, “Single and Multi-Agent Deep Reinforcement Learning for AI-Enabled Wireless Networks: A Tutorial,” IEEE Communications Survey & Tutorials, vol. 33, no. 2, pp. 1226 – 1252, 2021.
  35. N. Naderializadeh, J. Sydir, M. Simsek , and etc, “Resource Management in Wireless Networks via Multi-Agent Deep Reinforcement Learning,” IEEE Transactions on Wireless Communications, vol. 20, no. 6, pp. 3507 – 3523, 2021.
  36. Y. Nasir and D. Guo, “Multi-Agent Deep Reinforcement Learning for Dynamic Power Allocation in Wireless Networks,” IEEE Journal on Selected Areas in Communications, vol. 37, no. 10, pp. 2239 – 2250, 2019.
  37. Y. Zhang, B. Di, Z. Zheng, J. Lin, and etc, “Distributed Multi-Cloud Multi-Access Edge Computing by Multi-Agent Reinforcement Learning,” IEEE Transactions on Wireless Communications, vol. 20, no. 4, pp. 2565 – 2578, 2021.
  38. H. Li, Y. Wu, and M. Chen, “Adaptive Fault-Tolerant Tracking Control for Discrete-Time Multiagent Systems via Reinforcement Learning Algorithm,” IEEE Transactions on Cybernetics, vol. 51, no. 3, pp. 1163 – 1174, 2021.
  39. N. Garg and T. Ratnarajah, “Cooperative Scenarios for Multi-Agent Reinforcement Learning in Wireless Edge Caching,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021.
  40. K. Menda, Y. Chen, J. Grana, and etc, “Deep Reinforcement Learning for Event-Driven Multi-Agent Decision Processes,” IEEE Transactions on Intelligent Transaportation Systems, vol. 20, no. 4, pp. 1259 – 1268, 2019.
  41. M. Gori, G. Mondardini, and F. Scarselli, “A new model for learning in graph domains,” in IEEE International Joint Conference on Neural Networks, 2005.
  42. R. Pascanu, C. Gulcehre, K. Cho, and etc, “How to construct deep recurrent neural networks,” in International Conference on Learning Representations, 2014.
  43. M. Schuster and K. K. Paliwal, “Bidirectional recurrent neural networks,” IEEE Transactions on Signal Processing, vol. 45, no. 11, pp. 2673 – 2681, 1997.
  44. Y. Seo, M. Defferrard, P. Vandergheynst, and etc, “Structured sequence modeling with graph convolutional recurrent networks,” in International Conference on Neural Information Processing, 2018.
  45. Y. Li, R. Yu, C. Shahabi, and etc, “Diffusion convolutional recurrent neural network: Data-driven traffic forecasting,” in International Conference on Learning Representations, 2018.
  46. C. Capanema, G. Oliveira, F. Silva, and etc, “Combining recurrent and Graph Neural Networks to predict the next place’s category,” Ad Hoc Networks, vol. 138, no. 103016, 2023.
  47. L. Ruiz, L. F. O. Chamon, and A. Ribeiro, “Transferability Properties of Graph Neural Networks,” arXiv, Dec 2021.
  48. V. Mnih, A. Badia, M. Mirza, and etc, “Asynchronous methods for deep reinforcement learning,” in International Conference on Machine Learning, 2016.
  49. P. Dhariwal, C. Hesse, O. Klimov, and etc, “Openai baselines,” OpenAI, Tech. Rep., Jan 2020. [Online]. Available: https://github.com/openai/baselines
  50. F. Gama, A. G. Marques, G. Leus, and A. Ribeiro, “Convolutional Graph Neural Networks,” in The 53rd Asilomar Conference on Circuits, Systems and Computers (ACSSC), 2019.
  51. J. Du, J. Shi, S. Kar, and etc, “On graph convolution for graph CNNs,” in IEEE Data Science Workshop (DSW), 2018.
  52. T. Rashid, M. Samvelyan, C. Schroeder-De-Witt, and etc, “QMIX: monotonic value function factorisation for deep multi-agent reinforcement learning,” in International Conference on Machine Learning, 2018.
  53. I. Liu, R. A. Yeh, and A. G. Schwing, “Pic: Permutation invariant critic for multi-agent deep reinforcement learning,” arXiv, 2019.
  54. S. Knight, H. X. Nguyen, N. Falkner, and et.al., “The Internet Topology Zoo,” IEEE Journal on Selected Areas in Communications, vol. 29, no. 9, pp. 1765 – 1775, 2011.
  55. [Online]. Available: pytorch-geometric.readthedocs.io/en/latest/
  56. B. Gao and L. Pavel, “On the Properties of the Softmax Function with Application in Game Theory and Reinforcement Learning,” arXiv, Apr 2017.
  57. A. B. Aleksandrov and V. V. Peller, “Operator Lipschitz functions,” Russian Mathematical Surveys, vol. 71, no. 4, pp. 605 – 702, 2016.
Citations (1)

Summary

We haven't generated a summary for this paper yet.