Towards Auto-Modeling of Formal Verification for NextG Protocols: A Multimodal cross- and self-attention Large Language Model Approach (2312.17353v2)
Abstract: This paper introduces Auto-modeling of Formal Verification with Real-world Prompting for 5G and NextG protocols (AVRE), a novel system designed for the formal verification of Next Generation (NextG) communication protocols, addressing the increasing complexity and scalability challenges in network protocol design and verification. Utilizing LLMs, AVRE transforms protocol descriptions into dependency graphs and formal models, efficiently resolving ambiguities and capturing design intent. The system integrates a transformer model with LLMs to autonomously establish quantifiable dependency relationships through cross- and self-attention mechanisms. Enhanced by iterative feedback from the HyFuzz experimental platform, AVRE significantly advances the accuracy and relevance of formal verification in complex communication protocols, offering a groundbreaking approach to validating sophisticated communication systems. We compare CAL's performance with state-of-the-art LLM-based models and traditional time sequence models, demonstrating its superiority in accuracy and robustness, achieving an accuracy of 95.94\% and an AUC of 0.98. This NLP-based approach enables, for the first time, the creation of exploits directly from design documents, making remarkable progress in scalable system verification and validation.
- 3GPP, “5g evolution toward 5g advanced: An overview of 3gpp releases 17 and 18,” IEEE Xplore, 2022.
- “Nr; radio resource control (rrc); protocol specification,” 3rd Generation Partnership Project (3GPP), Technical Specification (TS) 38.331, 2021, available: 3GPP website. [Online]. Available: https://www.3gpp.org
- “Open ran – 5g hacking just got a lot more interesting,” Karsten Nohl, Security Research Labs, Tech. Rep., 2022.
- D. Dolev and A. Yao, “On the security of public key protocols,” IEEE Transactions on information theory, vol. 29, no. 2, pp. 198–208, 1983.
- I. Cervesato, “The dolev-yao intruder is the most powerful attacker,” in 16th Annual Symposium on Logic in Computer Science—LICS, vol. 1. Citeseer, 2001, pp. 1–2.
- M. Ammann, L. Hirschi, and S. Kremer, “Dy fuzzing: Formal dolev-yao models meet cryptographic protocol fuzz testing,” in 45th IEEE Symposium on Security and Privacy, 2024.
- I. Rakotonirina, G. Barthe, and C. Schneidewind, “Decision and complexity of dolev-yao hyperproperties (technical report),” in Symposium on Principles of Programming Languages (POPL), 2024.
- R. Kande, H. Pearce, B. Tan, B. Dolan-Gavitt, S. Thakur, R. Karri, and J. Rajendran, “Llm-assisted generation of hardware assertions,” arXiv preprint arXiv:2306.14027, 2023.
- M. Cosler, C. Hahn, D. Mendoza, F. Schmitt, and C. Trippel, “nl2spec: Interactively translating unstructured natural language to temporal logics with large language models,” arXiv preprint arXiv:2303.04864, 2023.
- M. Orenes-Vera, M. Martonosi, and D. Wentzlaff, “Using llms to facilitate formal verification of rtl,” arXiv e-prints, pp. arXiv–2309, 2023.
- S. Yuan, J. Yang, S. Arya, C. Lipizzi, and Y. Wang, “From ambiguity to explicitness: Nlp-assisted 5g specification abstraction for formal analysis,” arXiv preprint arXiv:2308.03277, 2023.
- J. Yang and Y. Wang, “A nextg hybrid testing platform for multi-step deep fuzzing and performance assessment from virtualization to over-the-air,” in IEEE Cloudnet, 2023.
- R. Drechsler, M. Soeken, and R. Wille, “Formal specification level: Towards verification-driven design based on natural language processing,” in Proceeding of the 2012 Forum on Specification and Design Languages. IEEE, 2012, pp. 53–58.
- L. Banarescu, C. Bonial, S. Cai, M. Georgescu, K. Griffitt, U. Hermjakob, K. Knight, P. Koehn, M. Palmer, and N. Schneider, “Abstract meaning representation for sembanking,” in Proceedings of the 7th linguistic annotation workshop and interoperability with discourse, 2013, pp. 178–186.
- L. Dong and M. Lapata, “Language to logical form with neural attention,” arXiv preprint arXiv:1601.01280, 2016.
- S. Reddy, D. Chen, and C. D. Manning, “Coqa: A conversational question answering challenge,” Transactions of the Association for Computational Linguistics, vol. 7, pp. 249–266, 2019.
- C. Hahn, F. Schmitt, J. J. Tillman, N. Metzger, J. Siber, and B. Finkbeiner, “Formal specifications from natural language,” arXiv preprint arXiv:2206.01962, 2022.
- A. Saparov and H. He, “Language models are greedy reasoners: A systematic formal analysis of chain-of-thought,” arXiv preprint arXiv:2210.01240, 2022.
- P. Srikumar, “Fast and wrong: The case for formally specifying hardware with llms,” in Proceedings of the International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS). ACM. ACM Press, 2023.
- N. Tihanyi, T. Bisztray, R. Jain, M. A. Ferrag, L. C. Cordeiro, and V. Mavroeidis, “The formai dataset: Generative ai in so ware security through the lens of formal verification,” PROMISE’23, p. 33, 2023.
- Y. Charalambous, N. Tihanyi, R. Jain, Y. Sun, M. A. Ferrag, and L. C. Cordeiro, “A new era in software security: Towards self-healing software via large language models and formal verification,” arXiv preprint arXiv:2305.14752, 2023.
- S. Abdelnabi, K. Greshake, S. Mishra, C. Endres, T. Holz, and M. Fritz, “Not what you’ve signed up for: Compromising real-world llm-integrated applications with indirect prompt injection,” in Proceedings of the 16th ACM Workshop on Artificial Intelligence and Security, 2023, pp. 79–90.
- S. Ouyang, J. M. Zhang, M. Harman, and M. Wang, “Llm is like a box of chocolates: the non-determinism of chatgpt in code generation,” arXiv preprint arXiv:2308.02828, 2023.
- K. Krishna, Y. Chang, J. Wieting, and M. Iyyer, “Rankgen: Improving text generation with large ranking models,” arXiv preprint arXiv:2205.09726, 2022.
- Y. Wang, A. Gorski, and A. P. da Silva, “Development of a data-driven mobile 5g testbed: Platform for experimental research,” in 2021 IEEE International Mediterranean Conference on Communications and Networking (MeditCom). IEEE, 2021, pp. 324–329.
- J. Yang, S. Arya, and Y. Wang, “Formal-guided fuzz testing: Targeting security assurance from specification to implementation for 5g and beyond,” arXiv preprint arXiv:2307.11247, 2023.
- J. Yang and Y. Wang, “Formal and fuzzing amplification: Targeting vulnerability detection in 5g and beyond,” arXiv preprint arXiv:2307.05758, 2023.
- D. Dauphinais, M. Zylka, H. Spahic, F. Shaik, J. Yang, I. Cruz, J. Gibson, and Y. Wang, “Automated vulnerability testing and detection digital twin framework for 5g systems,” in 2023 IEEE 9th International Conference on Network Softwarization (NetSoft). IEEE, 2023, pp. 308–310.
- “3gpp ts 38.331: Radio resource control (rrc),” 3GPP, Report, 2022.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention is all you need,” 2017. [Online]. Available: https://arxiv.org/pdf/1706.03762.pdf
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
- J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “Bert: Pre-training of deep bidirectional transformers for language understanding,” arXiv preprint arXiv:1810.04805, 2018.
- T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell et al., “Language models are few-shot learners,” arXiv preprint arXiv:2005.14165, 2020.