DeepMPR: Enhancing Opportunistic Routing in Wireless Networks through Multi-Agent Deep Reinforcement Learning (2306.09637v1)
Abstract: Opportunistic routing relies on the broadcast capability of wireless networks. It brings higher reliability and robustness in highly dynamic and/or severe environments such as mobile or vehicular ad-hoc networks (MANETs/VANETs). To reduce the cost of broadcast, multicast routing schemes use the connected dominating set (CDS) or multi-point relaying (MPR) set to decrease the network overhead and hence, their selection algorithms are critical. Common MPR selection algorithms are heuristic, rely on coordination between nodes, need high computational power for large networks, and are difficult to tune for network uncertainties. In this paper, we use multi-agent deep reinforcement learning to design a novel MPR multicast routing technique, DeepMPR, which is outperforming the OLSR MPR selection algorithm while it does not require MPR announcement messages from the neighbors. Our evaluation results demonstrate the performance gains of our trained DeepMPR multicast forwarding policy compared to other popular techniques.
- S. Biswas and R. Morris, “Opportunistic routing in multi-hop wireless networks,” ACM SIGCOMM Computer Communication Review, vol. 34, no. 1, pp. 69–74, 2004.
- ——, “Exor: opportunistic multi-hop routing for wireless networks,” in ACM SIGCOMM Computer Communication Review, vol. 35, no. 4. ACM, 2005, pp. 133–144.
- J. A. Macker, “Simplified multicast forwarding,” Internet Engineering Task Force (IETF), vol. 83, no. 1, pp. 1–27, 2012. [Online]. Available: https://www.rfc-editor.org/rfc/rfc6610
- X. Hu, I. Stojmenovic, and L. Zhang, “A connected dominating set based multicast routing protocol in wireless ad hoc networks,” IEEE Transactions on Mobile Computing, vol. 3, no. 4, pp. 337–346, 2004.
- T. Clausen and P. Jacquet, “Optimized link state routing protocol (OLSR),” RFC 3626, Oct. 2003.
- A. Qayyum, L. Viennot, and A. Laouiti, “Multipoint relaying for flooding broadcast messages in mobile wireless networks,” in Proceedings of the 35th annual Hawaii international conference on system sciences. IEEE, 2002, pp. 3866–3875.
- L. Maccari, M. Maischberger, and R. Lo Cigno, “Where have all the MPRs gone? on the optimal selection of multi-point relays,” Ad Hoc Networks, vol. 77, pp. 69–83, 2018.
- D. Gantsou, P. Sondi, and S. Hanafi, “Revisiting multipoint relay selection in the optimized link state routing protocol,” International Journal of Communication Networks and Distributed Systems, vol. 2, no. 1, pp. 4–15, 2009.
- A. Singh and W. N. Bhukya, “An evolutionary approach to multi-point relays selection in mobile ad hoc networks,” in Pattern Recognition and Machine Intelligence: 8th International Conference, PReMI 2019, Tezpur, India, December 17-20, 2019, Proceedings, Part I. Springer, 2019, pp. 375–384.
- S. Chen and K. Nahrstedt, “Distributed quality-of-service routing in ad hoc networks,” IEEE Journal on Selected Areas in Communications, vol. 17, no. 8, pp. 1488–1505, August 1999.
- R. Bellman, “A markovian decision process,” Indiana University Mathematics Journal, vol. 6, no. 4, pp. 679–684, April 1957.
- V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves, M. Riedmiller, A. K. Fidjeland, G. Ostrovski et al., “Human-level control through deep reinforcement learning,” nature, vol. 518, no. 7540, pp. 529–533, 2015.
- C. Yu, J. Lan, Z. Guo, and Y. Hu, “DROM: Optimizing the routing in software-defined networks with deep reinforcement learning,” IEEE Access, vol. 6, pp. 64 533–64 539, 2018.
- G. Stampa, M. Arias, D. Sánchez-Charles, V. Muntés-Mulero, and A. Cabellos, “A deep-reinforcement learning approach for software-defined networking routing optimization,” arXiv preprint arXiv:1709.07080, 2017.
- W. Cui and W. Yu, “Scalable deep reinforcement learning for routing and spectrum access in physical layer,” IEEE Transactions on Communications, vol. 69, no. 12, pp. 8200–8213, 2021.
- S. Kaviani, B. Ryu, E. Ahmed, K. Larson, A. Le, A. Yahja, and J. H. Kim, “DeepCQ+: Robust and scalable routing with multi-agent deep reinforcement learning for highly dynamic networks,” in MILCOM 2021-2021 IEEE Military Communications Conference (MILCOM). IEEE, 2021, pp. 31–36.
- B. Liang and Z. J. Haas, “Predictive distance-based mobility management for pcs networks,” in IEEE INFOCOM’99. Conference on Computer Communications. Proceedings. Eighteenth Annual Joint Conference of the IEEE Computer and Communications Societies. The Future is Now (Cat. No. 99CH36320), vol. 3. IEEE, 1999, pp. 1377–1384.
- J. Ariyakhajorn, P. Wannawilai, and C. Sathitwiriyawong, “A comparative study of Random Waypoint and Gauss-Markov mobility models in the performance evaluation of MANET,” in 2006 International Symposium on Communications and Information Technologies, 2006, pp. 894–899.
- R. S. Sutton, “Learning to predict by the methods of temporal differences,” Machine learning, vol. 3, no. 1, pp. 9–44, 1988.
- J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” arXiv preprint arXiv:1707.06347, 2017.
- L. Ouyang, J. Wu, X. Jiang, D. Almeida, C. Wainwright, P. Mishkin, C. Zhang, S. Agarwal, K. Slama, A. Ray et al., “Training language models to follow instructions with human feedback,” Advances in Neural Information Processing Systems, vol. 35, pp. 27 730–27 744, 2022.
- J. K. Terry, N. Grammel, A. Hari, L. Santos, B. Black, and D. Manocha, “Parameter sharing is surprisingly useful for multi-agent deep reinforcement learning,” arXiv preprint arXiv:2005.13625, 2020.
- P. Moritz, R. Nishihara, S. Wang, A. Tumanov, R. Liaw, E. Liang, M. Elibol, Z. Yang, W. Paul, M. I. Jordan et al., “Ray: A distributed framework for emerging {{\{{AI}}\}} applications,” in 13th USENIX Symposium on Operating Systems Design and Implementation (OSDI 18), 2018, pp. 561–577.
- E. Liang, R. Liaw, R. Nishihara, P. Moritz, R. Fox, K. Goldberg, J. Gonzalez, M. Jordan, and I. Stoica, “Rllib: Abstractions for distributed reinforcement learning,” in International Conference on Machine Learning. PMLR, 2018, pp. 3053–3062.