Comba: Improving Bilinear RNNs with Closed-loop Control
Comba: Improving Bilinear RNNs with Closed-loop Control
Recent efficient sequence modeling methods such as Gated DeltaNet, TTT, and RWKV-7 have achieved performance improvements by supervising the recurrent memory management through Delta learning rule. Unlike previous state-space models (e.g., Mamba) and gated linear attentions (e.g., GLA), these models introduce interactions between the recurrent state and the key vector, structurally resembling bilinear systems. In this paper, we first introduce the concept of Bilinear RNNs with a comprehensive analysis on the advantages and limitations of these models. Then, based on closed-loop control theory, we propose a novel Bilinear RNN variant named Comba, which adopts a scalar-plus-low-rank state transition, with both state feedback and output feedback corrections. We also implement a hardware-efficient chunk-wise parallel kernel in Triton and train models with 340M/1.3B parameters on large-scale corpus. Comba demonstrates superior performance and computation efficiency in both language and vision modeling.
Jiaxi Hu、Yongqi Pan、Jusen Du、Disen Lan、Xiaqiang Tang、Qingsong Wen、Yuxuan Liang、Weigao Sun
自动化基础理论计算技术、计算机技术
Jiaxi Hu,Yongqi Pan,Jusen Du,Disen Lan,Xiaqiang Tang,Qingsong Wen,Yuxuan Liang,Weigao Sun.Comba: Improving Bilinear RNNs with Closed-loop Control[EB/OL].(2025-06-21)[2025-07-02].https://arxiv.org/abs/2506.02475.点此复制
评论