Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

What Affects the Stability of Tool Learning? An Empirical Study on the Robustness of Tool Learning Frameworks (2407.03007v1)

Published 3 Jul 2024 in cs.CL and cs.AI

Abstract: Tool learning methods have enhanced the ability of LLMs to interact with real-world applications. Many existing works fine-tune LLMs or design prompts to enable LLMs to select appropriate tools and correctly invoke them to meet user requirements. However, it is observed in previous works that the performance of tool learning varies from tasks, datasets, training settings, and algorithms. Without understanding the impact of these factors, it can lead to inconsistent results, inefficient model deployment, and suboptimal tool utilization, ultimately hindering the practical integration and scalability of LLMs in real-world scenarios. Therefore, in this paper, we explore the impact of both internal and external factors on the performance of tool learning frameworks. Through extensive experiments on two benchmark datasets, we find several insightful conclusions for future work, including the observation that LLMs can benefit significantly from increased trial and exploration. We believe our empirical study provides a new perspective for future tool learning research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Chengrui Huang (4 papers)
  2. Zhengliang Shi (15 papers)
  3. Yuntao Wen (2 papers)
  4. Xiuying Chen (80 papers)
  5. Peng Han (37 papers)
  6. Shen Gao (49 papers)
  7. Shuo Shang (30 papers)