适配器(计算)
融合
计算机科学
计算机硬件
哲学
语言学
作者
Dehong Gao,Yufei Ma,Sen Liu,Mengfei Song,Linbo Jin,Wen Jiang,Xin Wang,Ning Wei,Shanqing Yu,Qi Xuan,Xiaoyan Cai,Libin Yang
标识
DOI:10.1016/j.knosys.2024.112043
摘要
We present FashionGPT, a series of fine-tuned Large Language Models (LLMs) which demonstrate outstanding performance and stand at first place in HuggingFace Open LLM Leaderboard twice. In contrast to conventional dataset fusion fine-tuning, we propose a novel instruction fine-tuning paradigm, called multiple LoRA-adapter fusion fine-tuning. This paradigm involves fine-tuning multiple independent LoRA-adapters based on distinct datasets, which are subsequently fused using learnable weights to create a versatile large language model. Extensive experiments demonstrate that the LLMs fine-tuned with the LoRA-adapter fusion approaches outperform those fine-tuned with the dataset fusion approaches. The FashionGPT series is released in https://huggingface.co/ICBU-NPU/ and only for research use.
科研通智能强力驱动
Strongly Powered by AbleSci AI