Yujia Qin [email protected] Tsinghua University, Shihao Liang Tsinghua University, Yining Ye Tsinghua University, Kunlun Zhu Tsinghua University, Lan Yan Tsinghua University, Yaxi Lu, Yankai Lin Tsinghua University Renmin University of China, Xin Cong Tsinghua University, Xiangru Tang Yale University 5 WeChat AI Tencent Inc. 6 Zhihu Inc, Bill Qian Yale University 5 WeChat AI Tencent Inc. 6 Zhihu Inc, Sihan Zhao Tsinghua University, Lauren Hong Tsinghua University, Runchu Tian Tsinghua University, Ruobing Xie, Jie Zhou, Mark Gerstein Yale University 5 WeChat AI Tencent Inc. 6 Zhihu Inc, Dahai Li ModelBest Inc, Zhiyuan Liu Tsinghua University, Maosong Sun Tsinghua University, • • Movies, Tool API API Collection Instruction Generation Solution Path Annotation LLaMA ToolLLaMA RapidAPI (2023)
The paper presents ToolLLM, a framework designed to enhance the tool-use capabilities of large language models (LLMs) like LLaMA by enabling them to effectively interact with APIs. Recognizing that open-source LLMs have limitations in tool use compared to closed-source models like ChatGPT, the authors developed a new instruction-tuning dataset called ToolBench, which includes 16,464 real-world RESTful APIs across 49 categories. The dataset construction involves API collection from RapidAPI Hub, instruction generation using ChatGPT, and solution path annotation using a depth-first search-based decision tree algorithm. Additionally, ToolEval, an automatic evaluator, is introduced to assess model performance based on pass and win rates. Experimentation demonstrates that the fine-tuned ToolLLaMA model exhibits strong performance in executing complex instructions and generalizing to unseen APIs, while outperforming various baseline models.
This paper employs the following methods:
The following datasets were used in this research:
The authors identified the following limitations: