Yaowei Zheng School of Computer Science and Engineering Beihang University China, Richong Zhang [email protected] School of Computer Science and Engineering Beihang University China, Junhao Zhang [email protected], Yanhan Ye [email protected] School of Computer Science and Engineering Beihang University China, Zheyan Luo School of Computer Science and Engineering Beihang University China, Zhangchi Feng School of Computer Science and Engineering Beihang University China, Yongqiang Ma School of Computer Science and Engineering Beihang University China School of Software and Microelectronics Peking University China (2024)
The paper presents LLAMAFACTORY, a unified framework for efficient fine-tuning of over 100 large language models (LLMs). The framework integrates various cutting-edge training methods and provides an intuitive web UI, LLAMABOARD, for users to customize fine-tuning without coding. LLAMAFACTORY emphasizes efficiency in training through scalable modules—Model Loader, Data Worker, and Trainer— designed to standardize model and data processing. The authors empirically validate the framework's effectiveness on language modeling and text generation tasks, reporting significant improvements in memory usage and training throughput, making LLM fine-tuning accessible to a broader audience.
This paper employs the following methods:
The following datasets were used in this research:
The authors identified the following limitations: