|国家预印本平台
首页|Slimming Down LLMs Without Losing Their Minds

Slimming Down LLMs Without Losing Their Minds

Slimming Down LLMs Without Losing Their Minds

来源:Arxiv_logoArxiv
英文摘要

This paper investigates and validates the impact of fine-tuning on large language model performance, focusing on parameter-efficient methods (LoRA and QLoRA). We evaluate model capabilities across three key domains: (1) commonsense reasoning (HellaSwag), (2) mathematical reasoning (GSM8K), and (3) multi-domain knowledge (MMLU-CS). Our findings demonstrate that: (1) LoRA-based methods effectively improve task-specific performance while maintaining computational efficiency, and (2) performance strongly depends on alignment between fine-tuning dataset and benchmark tasks. The study provides both theoretical insights into parameter-efficient mechanisms and practical guidance for developers implementing efficient LLM adaptation with limited resources.

Qingda、Mai

Michael

计算技术、计算机技术

Qingda,Mai.Slimming Down LLMs Without Losing Their Minds[EB/OL].(2025-06-12)[2025-07-09].https://arxiv.org/abs/2506.10885.点此复制

评论