Decentralized Learning Strategies for Estimation Error Minimization with Graph Neural Networks (2404.03227v2)
Abstract: We address the challenge of sampling and remote estimation for autoregressive Markovian processes in a multi-hop wireless network with statistically-identical agents. Agents cache the most recent samples from others and communicate over wireless collision channels governed by an underlying graph topology. Our goal is to minimize time-average estimation error and/or age of information with decentralized scalable sampling and transmission policies, considering both oblivious (where decision-making is independent of the physical processes) and non-oblivious policies (where decision-making depends on physical processes). We prove that in oblivious policies, minimizing estimation error is equivalent to minimizing the age of information. The complexity of the problem, especially the multi-dimensional action spaces and arbitrary network topologies, makes theoretical methods for finding optimal transmission policies intractable. We optimize the policies using a graphical multi-agent reinforcement learning framework, where each agent employs a permutation-equivariant graph neural network architecture. Theoretically, we prove that our proposed framework exhibits desirable transferability properties, allowing transmission policies trained on small- or moderate-size networks to be executed effectively on large-scale topologies. Numerical experiments demonstrate that (i) Our proposed framework outperforms state-of-the-art baselines; (ii) The trained policies are transferable to larger networks, and their performance gains increase with the number of agents; (iii) The training procedure withstands non-stationarity even if we utilize independent learning techniques; and, (iv) Recurrence is pivotal in both independent learning and centralized training and decentralized execution, and improves the resilience to non-stationarity in independent learning.
- X. Chen, X. Liao and S. Saeedi-Bidokhti, “Real-time sampling and estimation on random access channels: Age of information and beyond,” in IEEE International Conference on Computer Communications (INFOCOM), 2021.
- R. V. Ramakanth, V. Tripathi, and E. Modiano, “Monitoring Correlated Sources: AoI-based Scheduling is Nearly Optimal,” arXiv, 2023.
- Y. Sun, Y. Polyanskiy, and E. Uysal-Biyikoglu, “Sampling of the Wiener process for remote estimation over a channel with random delay,” IEEE Transactions on Information Theory, vol. 66, no. 2, pp. 1118 – 1135, 2020.
- T. Z. Ornee and Y. Sun, “Sampling for remote estimation for Ornstein-Uhlenbeck process through queues: age of information and beyond,” IEEE/ACM Transactions on Networking, vol. 29, no. 5, pp. 1962 – 1975, 2021.
- S. Kaul, R. D. Yates, and M. Gruteser, “Real-time status: How often should one update?” in IEEE International Conference on Computer Communications (INFOCOM), 2012.
- Y. Sun, Y. Polyanskiy, and E. Uysal-Biyikoglu, “Remote Estimation of the Wiener Process over a Channel with Random Delay,” IEEE Transactions on Information Theory, vol. 66, no. 2, pp. 1118 – 1135, 2020.
- R. Talak and E. Modiano, “Age-Delay Tradeoffs in Queueing Systems,” IEEE Transactions on Information Theory, vol. 67, no. 3, pp. 1743 – 1758, 2021.
- X. Chen, K. Gatsis, H. Hassani and S. Saeedi-Bidokhti, “Age of information in random access channels,” IEEE Transactions on Information Theory, vol. 68, no. 10, pp. 6548 – 6568, 2022.
- O. T. Yavascan and E. Uysal, “Analysis of slotted ALOHA with an age threshold,” IEEE Journal on Selected Areas in Communications, vol. 39, no. 5, pp. 1456 – 1470, 2021.
- S. Kang, A. Eryilmaz, and N. B. Shroff, “Remote Tracking of Distributed Dynamic Sources Over a Random Access Channel With One-Bit Updates,” IEEE Transactions on Network Science and Engineering, vol. 10, no. 4, pp. 1931 – 1941, 2023.
- V. Tripathi, R. Talak, and E. Modiano, “Information freshness in multihop wireless networks,” IEEE Transactions on Information Theory, vol. 31, no. 2, pp. 784 – 799, 2023.
- N. Jones and E. Modiano, “Minimizing age of information in spatially distributed random access wireless networks,” arXiv, Dec 2022.
- G. Papoudakis, F. Christinos, L. Schafer, and etc, “Benchmarking multi-agent deep reinforcement learning algorithms in cooperative tasks,” in Advances in Neural Information Processing Systems, 2021.
- F. Gama, J. Bruna, and A. Ribeiro, “Stability properties of graph neural networks,” IEEE Transactions on Signal Processing, vol. 68, pp. 5680 – 5695, 2020.
- L. Ruiz, F. Gama, and A. Ribeiro, “Gated Graph Recurrent Neural Networks,” IEEE Transactions on Signal Processing, vol. 68, pp. 6303 – 6318, 2020.
- S. Kaul, M. Gruteser, V. Rai, et. al, “Minimizing age of information in vehicular networks,” 2011 8th Annual IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc Communications and Networks, 2011.
- I. Kadota and E. Modiano, “Minimizing age of information in wireless networks with stochastic arrivals,” IEEE Transactions on Mobile Computing, vol. 20, no. 3, pp. 1173 – 1185, 2021.
- R. D. Yates and S. K. Kaul, “The age of information: real-time status updating by multiple sources,” IEEE Transactions on Information Theory, vol. 65, no. 3, pp. 1807 – 1827, 2019.
- I. Kadota and E. Modiano, “Age of information in random access networks with stochastic arrivals,” in IEEE International Conference on Computer Communications, 2021.
- S. Farazi, A. G. Klein and D. R. Brown, ““Fundamental bounds on the age of information in general multi-hop interference networks,” in IEEE Conference on Computer Communications Workshops, 2019.
- B. Buyukates, A. Soysal, and S. Ulukus, “Age of information in multihop multicast networks,” Journal of Communications and Networks, vol. 21, no. 3, pp. 256 – 267, 2019.
- V. Tripathi, R. Talak, and E. Modiano, “Information freshness in multihop wireless networks,” IEEE/ACM Transactions on Networking, vol. 31, no. 2, pp. 784 – 799, 2022.
- E. Tolstaya, L. Butler, D. Mox, et. al, “Learning connectivity for data distribution in robot teams,” in IEEE/RSJ International Conference on Intelligent Robots and Systems, 2021.
- Y. Sun, Y. Polyanskiy, and E. Uysal-Biyikoglu, “Remote estimation of the Wiener process over a channel with random delay,” in IEEE International Symposium on Information Theory, 2017.
- T. Z. Ornee and Y. Sun, “Performance bounds for sampling and remote estimation of Gauss-Markov processes over a noisy channel with random delay,” in IEEE International Workshop on Signal Processing Advances in Wireless Communications, 2021.
- C. Tsai and C. Wang, “Unifying AoI Minimization and Remote Estimation—Optimal Sensor/Controller Coordination With Random Two-Way Delay,” IEEE/ACM Transactions on Networking, vol. 30, no. 1, pp. 229 – 242, 2022.
- C. Kam, S. Kompella, G. D. Nguyen, et. al, “Towards an Effective Age of Information: Remote Estimation of a Markov Source,” in IEEE Conference on Computer Communications Workshops, 2018.
- A. Maatouk, S. Kriouile, M. Assaad, et. al, “The Age of Incorrect Information: A New Performance Metric for Status Updates,” IEEE/ACM Transactions on Networking, vol. 28, no. 5, pp. 2215 – 2228, 2020.
- S. Saha, H. Makkar, V. Sukumaran, et. al, “On the Relationship Between Mean Absolute Error and Age of Incorrect Information in the Estimation of a Piecewise Linear Signal Over Noisy Channels,” IEEE Communications Letters, vol. 26, no. 11, pp. 2576 – 2580, 2022.
- P. Hernandez-Leal, B. Kartal and M. E. Taylor, “A survey and critique of multiagent deep reinforcement learning,” in International Conference on Autonomous Agents and Multi-Agent Systems, 2019.
- M. Tan, “Multi-agent reinforcement learning: Independent vs. cooperative agents,” in International Conference on Machine Learning, 1993.
- R. Lowe, Y. Wu, A. Tamar, and etc, “Multi-agent actor-critic for mixed cooperative-competitive environments,” in Advances in Neural Information Processing Systems, 2017.
- P. Sunehag, G. Lever, A. Gruslys, and etc, “Value-decomposition networks for cooperative multi-agent learning,” in International Conference on Autonomous Agents and Multi-Agent Systems, 2018.
- A. Feriani and E. Hossain, “Single and Multi-Agent Deep Reinforcement Learning for AI-Enabled Wireless Networks: A Tutorial,” IEEE Communications Survey & Tutorials, vol. 33, no. 2, pp. 1226 – 1252, 2021.
- N. Naderializadeh, J. Sydir, M. Simsek , and etc, “Resource Management in Wireless Networks via Multi-Agent Deep Reinforcement Learning,” IEEE Transactions on Wireless Communications, vol. 20, no. 6, pp. 3507 – 3523, 2021.
- Y. Nasir and D. Guo, “Multi-Agent Deep Reinforcement Learning for Dynamic Power Allocation in Wireless Networks,” IEEE Journal on Selected Areas in Communications, vol. 37, no. 10, pp. 2239 – 2250, 2019.
- Y. Zhang, B. Di, Z. Zheng, J. Lin, and etc, “Distributed Multi-Cloud Multi-Access Edge Computing by Multi-Agent Reinforcement Learning,” IEEE Transactions on Wireless Communications, vol. 20, no. 4, pp. 2565 – 2578, 2021.
- H. Li, Y. Wu, and M. Chen, “Adaptive Fault-Tolerant Tracking Control for Discrete-Time Multiagent Systems via Reinforcement Learning Algorithm,” IEEE Transactions on Cybernetics, vol. 51, no. 3, pp. 1163 – 1174, 2021.
- N. Garg and T. Ratnarajah, “Cooperative Scenarios for Multi-Agent Reinforcement Learning in Wireless Edge Caching,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021.
- K. Menda, Y. Chen, J. Grana, and etc, “Deep Reinforcement Learning for Event-Driven Multi-Agent Decision Processes,” IEEE Transactions on Intelligent Transaportation Systems, vol. 20, no. 4, pp. 1259 – 1268, 2019.
- M. Gori, G. Mondardini, and F. Scarselli, “A new model for learning in graph domains,” in IEEE International Joint Conference on Neural Networks, 2005.
- R. Pascanu, C. Gulcehre, K. Cho, and etc, “How to construct deep recurrent neural networks,” in International Conference on Learning Representations, 2014.
- M. Schuster and K. K. Paliwal, “Bidirectional recurrent neural networks,” IEEE Transactions on Signal Processing, vol. 45, no. 11, pp. 2673 – 2681, 1997.
- Y. Seo, M. Defferrard, P. Vandergheynst, and etc, “Structured sequence modeling with graph convolutional recurrent networks,” in International Conference on Neural Information Processing, 2018.
- Y. Li, R. Yu, C. Shahabi, and etc, “Diffusion convolutional recurrent neural network: Data-driven traffic forecasting,” in International Conference on Learning Representations, 2018.
- C. Capanema, G. Oliveira, F. Silva, and etc, “Combining recurrent and Graph Neural Networks to predict the next place’s category,” Ad Hoc Networks, vol. 138, no. 103016, 2023.
- L. Ruiz, L. F. O. Chamon, and A. Ribeiro, “Transferability Properties of Graph Neural Networks,” arXiv, Dec 2021.
- V. Mnih, A. Badia, M. Mirza, and etc, “Asynchronous methods for deep reinforcement learning,” in International Conference on Machine Learning, 2016.
- P. Dhariwal, C. Hesse, O. Klimov, and etc, “Openai baselines,” OpenAI, Tech. Rep., Jan 2020. [Online]. Available: https://github.com/openai/baselines
- F. Gama, A. G. Marques, G. Leus, and A. Ribeiro, “Convolutional Graph Neural Networks,” in The 53rd Asilomar Conference on Circuits, Systems and Computers (ACSSC), 2019.
- J. Du, J. Shi, S. Kar, and etc, “On graph convolution for graph CNNs,” in IEEE Data Science Workshop (DSW), 2018.
- T. Rashid, M. Samvelyan, C. Schroeder-De-Witt, and etc, “QMIX: monotonic value function factorisation for deep multi-agent reinforcement learning,” in International Conference on Machine Learning, 2018.
- I. Liu, R. A. Yeh, and A. G. Schwing, “Pic: Permutation invariant critic for multi-agent deep reinforcement learning,” arXiv, 2019.
- S. Knight, H. X. Nguyen, N. Falkner, and et.al., “The Internet Topology Zoo,” IEEE Journal on Selected Areas in Communications, vol. 29, no. 9, pp. 1765 – 1775, 2011.
- [Online]. Available: pytorch-geometric.readthedocs.io/en/latest/
- B. Gao and L. Pavel, “On the Properties of the Softmax Function with Application in Game Theory and Reinforcement Learning,” arXiv, Apr 2017.
- A. B. Aleksandrov and V. V. Peller, “Operator Lipschitz functions,” Russian Mathematical Surveys, vol. 71, no. 4, pp. 605 – 702, 2016.