Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RoTBench: A Multi-Level Benchmark for Evaluating the Robustness of Large Language Models in Tool Learning (2401.08326v3)

Published 16 Jan 2024 in cs.CL and cs.AI

Abstract: Tool learning has generated widespread interest as a vital means of interaction between LLMs and the physical world. Current research predominantly emphasizes LLMs' capacity to utilize tools in well-structured environments while overlooking their stability when confronted with the inevitable noise of the real world. To bridge this gap, we introduce RoTBench, a multi-level benchmark for evaluating the robustness of LLMs in tool learning. Specifically, we establish five external environments, each featuring varying levels of noise (i.e., Clean, Slight, Medium, Heavy, and Union), providing an in-depth analysis of the model's resilience across three critical phases: tool selection, parameter identification, and content filling. Experiments involving six widely-used models underscore the urgent necessity for enhancing the robustness of LLMs in tool learning. For instance, the performance of GPT-4 even drops significantly from 80.00 to 58.10 when there is no substantial change in manual accuracy. More surprisingly, the noise correction capability inherent in the GPT family paradoxically impedes its adaptability in the face of mild noise. In light of these findings, we propose RoTTuning, a strategy that enriches the diversity of training environments to bolster the robustness of LLMs in tool learning. The code and data are available at https://github.com/Junjie-Ye/RoTBench.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Junjie Ye (66 papers)
  2. Yilong Wu (11 papers)
  3. Songyang Gao (28 papers)
  4. Sixian Li (12 papers)
  5. Guanyu Li (10 papers)
  6. Xiaoran Fan (23 papers)
  7. Qi Zhang (784 papers)
  8. Tao Gui (127 papers)
  9. Xuanjing Huang (287 papers)
  10. Caishuang Huang (13 papers)
Citations (9)