|国家预印本平台
首页|InstantFT: An FPGA-Based Runtime Subsecond Fine-tuning of CNN Models

InstantFT: An FPGA-Based Runtime Subsecond Fine-tuning of CNN Models

InstantFT: An FPGA-Based Runtime Subsecond Fine-tuning of CNN Models

来源:Arxiv_logoArxiv
英文摘要

Training deep neural networks (DNNs) requires significantly more computation and memory than inference, making runtime adaptation of DNNs challenging on resource-limited IoT platforms. We propose InstantFT, an FPGA-based method for ultra-fast CNN fine-tuning on IoT devices, by optimizing the forward and backward computations in parameter-efficient fine-tuning (PEFT). Experiments on datasets with concept drift demonstrate that InstantFT fine-tunes a pre-trained CNN 17.4x faster than existing Low-Rank Adaptation (LoRA)-based approaches, while achieving comparable accuracy. Our FPGA-based InstantFT reduces the fine-tuning time to just 0.36s and improves energy-efficiency by 16.3x, enabling on-the-fly adaptation of CNNs to non-stationary data distributions.

Keisuke Sugiura、Hiroki Matsutani

半导体技术电子技术应用

Keisuke Sugiura,Hiroki Matsutani.InstantFT: An FPGA-Based Runtime Subsecond Fine-tuning of CNN Models[EB/OL].(2025-06-06)[2025-06-17].https://arxiv.org/abs/2506.06505.点此复制

评论