LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report
LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report
Low Rank Adaptation (LoRA) has emerged as one of the most widely adopted methods for Parameter Efficient Fine-Tuning (PEFT) of Large Language Models (LLMs). LoRA reduces the number of trainable parameters and memory usage while achieving comparable performance to full fine-tuning. We aim to assess the viability of training and serving LLMs fine-tuned with LoRA in real-world applications. First, we measure the quality of LLMs fine-tuned with quantized low rank adapters across 10 base models and 31 tasks for a total of 310 models. We find that 4-bit LoRA fine-tuned models outperform base models by 34 points and GPT-4 by 10 points on average. Second, we investigate the most effective base models for fine-tuning and assess the correlative and predictive capacities of task complexity heuristics in forecasting the outcomes of fine-tuning. Finally, we evaluate the latency and concurrency capabilities of LoRAX, an open-source Multi-LoRA inference server that facilitates the deployment of multiple LoRA fine-tuned models on a single GPU using shared base model weights and dynamic adapter loading. LoRAX powers LoRA Land, a web application that hosts 25 LoRA fine-tuned Mistral-7B LLMs on a single NVIDIA A100 GPU with 80GB memory. LoRA Land highlights the quality and cost-effectiveness of employing multiple specialized LLMs over a single, general-purpose LLM.
Justin Zhao、Geoffrey Angus、Timothy Wang、Travis Addair、Devvret Rishi、Wael Abid、Piero Molino、Jeffery Kinnison、Alex Sherstinsky、Arnav Garg
计算技术、计算机技术
Justin Zhao,Geoffrey Angus,Timothy Wang,Travis Addair,Devvret Rishi,Wael Abid,Piero Molino,Jeffery Kinnison,Alex Sherstinsky,Arnav Garg.LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report[EB/OL].(2024-04-29)[2025-07-16].https://arxiv.org/abs/2405.00732.点此复制
评论