|国家预印本平台
首页|Memory-Efficient Split Federated Learning for LLM Fine-Tuning on Heterogeneous Mobile Devices

Memory-Efficient Split Federated Learning for LLM Fine-Tuning on Heterogeneous Mobile Devices

Memory-Efficient Split Federated Learning for LLM Fine-Tuning on Heterogeneous Mobile Devices

来源:Arxiv_logoArxiv
英文摘要

In this paper, we propose an edge-assisted split federated learning framework to facilitate large language model (LLM) fine-tuning on heterogeneous mobile devices while alleviating memory pressures on both mobile devices and the edge server. Specifically, mobile devices perform low-rank adaptation (LoRA) fine-tuning on only a subset of lower layers of the pre-trained LLM, tailored to their individual capacities. On the server, a full LLM is maintained, and the corresponding LoRA modules are selectively fine-tuned in a sequential manner for each device. To further enhance training efficiency, we propose a server-side training scheduling method that optimizes the processing order of devices for accelerating fine-tuning. Extensive experiments demonstrate that compared to the baselines, our scheme can reduce 79\% memory footprint and 6\% training time while achieving comparable performance.

Xiaopei Chen、Liang Li、Fei Ji、Wen Wu

计算技术、计算机技术

Xiaopei Chen,Liang Li,Fei Ji,Wen Wu.Memory-Efficient Split Federated Learning for LLM Fine-Tuning on Heterogeneous Mobile Devices[EB/OL].(2025-06-03)[2025-07-16].https://arxiv.org/abs/2506.02940.点此复制

评论