GoEX: Perspectives and Designs Towards a Runtime for Autonomous LLM Applications (2404.06921v1)
Abstract: LLMs are evolving beyond their classical role of providing information within dialogue systems to actively engaging with tools and performing actions on real-world applications and services. Today, humans verify the correctness and appropriateness of the LLM-generated outputs (e.g., code, functions, or actions) before putting them into real-world execution. This poses significant challenges as code comprehension is well known to be notoriously difficult. In this paper, we study how humans can efficiently collaborate with, delegate to, and supervise autonomous LLMs in the future. We argue that in many cases, "post-facto validation" - verifying the correctness of a proposed action after seeing the output - is much easier than the aforementioned "pre-facto validation" setting. The core concept behind enabling a post-facto validation system is the integration of an intuitive undo feature, and establishing a damage confinement for the LLM-generated actions as effective strategies to mitigate the associated risks. Using this, a human can now either revert the effect of an LLM-generated output or be confident that the potential risk is bounded. We believe this is critical to unlock the potential for LLM agents to interact with applications and services with limited (post-facto) human involvement. We describe the design and implementation of our open-source runtime for executing LLM actions, Gorilla Execution Engine (GoEX), and present open research questions towards realizing the goal of LLMs and applications interacting with each other with minimal human supervision. We release GoEX at https://github.com/ShishirPatil/gorilla/.
- GPT-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
- The dataflow model: a practical approach to balancing correctness, latency, and cost in massive-scale, unbounded, out-of-order data processing. VLDB, 2015.
- Many-shot jailbreaking. 2024.
- AWS. Testing your AWS KMS API calls. https://docs.aws.amazon.com/kms/latest/developerguide/programming-dryrun.html.
- Discovering latent knowledge in language models without supervision. arXiv preprint arXiv:2212.03827, 2022.
- Apache Flink: Stream and batch processing in a single engine. The Bulletin of the Technical Committee on Data Engineering, 2015.
- A survey on evaluation of large language models. ACM TIST, 2023.
- StruQ: Defending against prompt injection with structured queries. arXiv preprint arXiv:2402.06363, 2024.
- Protection in the Hydra operating system. SOSP, 1975.
- Not what you’ve signed up for: Compromising real-world LLM-integrated applications with indirect prompt injection. In ACM AISec, 2023.
- Principles of transaction-oriented database recovery. ACM CSUR, 1983.
- Exploiting programmatic behavior of LLMs: Dual-use through standard security attacks. arXiv preprint arXiv:2302.05733, 2023.
- Kubernetes. kubectl usage conventions. https://kubernetes.io/docs/reference/kubectl/conventions/.
- Theodore A Linden. Operating system structures to support security and reliable software. CSUR, 1976.
- Prompt injection attack against LLM-integrated applications. arXiv preprint arXiv:2306.05499, 2023.
- Microsoft. Presidio - data protection and de-identification SDK. https://github.com/microsoft/presidio.
- OpenAI. ChatGPT plugins, 2023. https://openai.com/blog/chatgpt-plugins.
- TALM: Tool augmented language models, 2022.
- Generative agents: Interactive simulacra of human behavior. In UIST, 2023.
- Gorilla: Large language model connected with massive APIs. arXiv preprint arXiv:2305.15334, 2023.
- Ignore previous prompt: Attack techniques for language models. arXiv preprint arXiv:2211.09527, 2022.
- Jatmo: Prompt injection defense by task-specific finetuning. arXiv preprint arXiv:2312.17673, 2023.
- ToolLLM: Facilitating large language models to master 16000+ real-world APIs, 2023.
- Ros: an open-source robot operating system. In ICRA workshop on open source software, volume 3, pp. 5. Kobe, Japan, 2009.
- Direct preference optimization: Your language model is secretly a reward model. NeurIPS, 2024.
- A survey of hallucination in large foundation models. arXiv preprint arXiv:2309.05922, 2023.
- Site isolation: Process separation for web sites within the browser. In USENIX Security, 2019.
- Code Llama: Open foundation models for code. arXiv preprint arXiv:2308.12950, 2023.
- Toolformer: Language models can teach themselves to use tools. arXiv preprint arXiv:2302.04761, 2023.
- Ignore this title and hackaprompt: Exposing systemic vulnerabilities of LLMs through a global scale prompt hacking competition. arXiv preprint arXiv:2311.16119, 2023.
- Xuchen Suo. Signed-Prompt: A new approach to prevent prompt injection attacks against LLM-integrated applications. arXiv preprint arXiv:2401.07612, 2024.
- Tensor trust: Interpretable prompt injection attacks from an online game. arXiv preprint arXiv:2311.01011, 2023.
- DecodingTrust: A comprehensive assessment of trustworthiness in GPT models. arXiv preprint arXiv:2306.11698, 2023a.
- Voyager: An open-ended embodied agent with large language models, 2023b.
- The Cambridge CAP computer and its operating system. 1979.
- Autogen: Enabling next-gen LLM applications via multi-agent conversation framework. arXiv preprint arXiv:2308.08155, 2023.
- SecGPT: An execution isolation architecture for llm-based systems. arXiv preprint arXiv:2403.04960, 2024.
- ReAct: Synergizing reasoning and acting in language models. arXiv preprint arXiv:2210.03629, 2022.
- Benchmarking and defending against indirect prompt injection attacks on large language models. arXiv preprint arXiv:2312.14197, 2023.
- Assessing prompt injection risks in 200+ custom gpts. arXiv preprint arXiv:2311.11538, 2023.
- Siren’s song in the AI ocean: a survey on hallucination in large language models. arXiv preprint arXiv:2309.01219, 2023.
- Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593, 2019.
- Shishir G. Patil (8 papers)
- Tianjun Zhang (38 papers)
- Vivian Fang (5 papers)
- Noppapon C. (1 paper)
- Roy Huang (1 paper)
- Aaron Hao (1 paper)
- Martin Casado (2 papers)
- Joseph E. Gonzalez (167 papers)
- Raluca Ada Popa (20 papers)
- Ion Stoica (177 papers)