Online Neural Model Fine-Tuning in Massive MIMO CSI Feedback: Taming The Communication Cost of Model Updates
Online Neural Model Fine-Tuning in Massive MIMO CSI Feedback: Taming The Communication Cost of Model Updates
Efficient channel state information (CSI) compression is essential in frequency division duplexing (FDD) massive multiple-input multiple-output (MIMO) systems due to the significant feedback overhead. Recently, deep learning-based compression techniques have demonstrated superior performance across various data types, including CSI. However, these methods often suffer from performance degradation when the data distribution shifts, primarily due to limited generalization capabilities. To address this challenge, we propose an online model fine-tuning approach for CSI feedback in massive MIMO systems. We consider full-model fine-tuning, where both the encoder and decoder are jointly updated using recent CSI samples. A key challenge in this setup is the transmission of updated decoder parameters, which introduces additional feedback overhead. To mitigate this bottleneck, we incorporate the bit-rate of model updates into the fine-tuning objective and entropy code the updates jointly with the compressed CSI. To reduce the bit-rate, we design an efficient prior distribution that encourages the network to update only the most significant weights, thereby minimizing the overall model update cost. Our results show that full-model fine-tuning significantly enhances the rate-distortion (RD) performance of neural CSI compression despite the additional communication cost of model updates. Moreover, we investigate the impact of update frequency in dynamic wireless environments and identify an optimal fine-tuning interval that achieves the best RD trade-off.
Tommy Svensson、Mehdi Sattari、Deniz G??nd??z
通信无线通信
Tommy Svensson,Mehdi Sattari,Deniz G??nd??z.Online Neural Model Fine-Tuning in Massive MIMO CSI Feedback: Taming The Communication Cost of Model Updates[EB/OL].(2025-06-19)[2025-07-01].https://arxiv.org/abs/2501.18250.点此复制
评论