LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models. Zheng, Y., Zhang, R., Zhang, J., Ye, Y., Luo, Z., & Ma, Y. March, 2024. arXiv:2403.13372 [cs]
LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models [link]Paper  doi  abstract   bibtex   
Efficient fine-tuning is vital for adapting large language models (LLMs) to downstream tasks. However, it requires non-trivial efforts to implement these methods on different models. We present LlamaFactory, a unified framework that integrates a suite of cutting-edge efficient training methods. It allows users to flexibly customize the fine-tuning of 100+ LLMs without the need for coding through the built-in web UI LlamaBoard. We empirically validate the efficiency and effectiveness of our framework on language modeling and text generation tasks. It has been released at https://github.com/hiyouga/LLaMA-Factory and already received over 13,000 stars and 1,600 forks.
@misc{zheng_llamafactory_2024,
	title = {{LlamaFactory}: {Unified} {Efficient} {Fine}-{Tuning} of 100+ {Language} {Models}},
	shorttitle = {{LlamaFactory}},
	url = {http://arxiv.org/abs/2403.13372},
	doi = {10.48550/arXiv.2403.13372},
	abstract = {Efficient fine-tuning is vital for adapting large language models (LLMs) to downstream tasks. However, it requires non-trivial efforts to implement these methods on different models. We present LlamaFactory, a unified framework that integrates a suite of cutting-edge efficient training methods. It allows users to flexibly customize the fine-tuning of 100+ LLMs without the need for coding through the built-in web UI LlamaBoard. We empirically validate the efficiency and effectiveness of our framework on language modeling and text generation tasks. It has been released at https://github.com/hiyouga/LLaMA-Factory and already received over 13,000 stars and 1,600 forks.},
	urldate = {2024-03-24},
	publisher = {arXiv},
	author = {Zheng, Yaowei and Zhang, Richong and Zhang, Junhao and Ye, Yanhan and Luo, Zheyan and Ma, Yongqiang},
	month = mar,
	year = {2024},
	note = {arXiv:2403.13372 [cs]},
}

Downloads: 0