|国家预印本平台
首页|Comba: Improving Bilinear RNNs with Closed-loop Control

Comba: Improving Bilinear RNNs with Closed-loop Control

Comba: Improving Bilinear RNNs with Closed-loop Control

来源:Arxiv_logoArxiv
英文摘要

Recent efficient sequence modeling methods such as Gated DeltaNet, TTT, and RWKV-7 have achieved performance improvements by supervising the recurrent memory management through Delta learning rule. Unlike previous state-space models (e.g., Mamba) and gated linear attentions (e.g., GLA), these models introduce interactions between the recurrent state and the key vector, structurally resembling bilinear systems. In this paper, we first introduce the concept of Bilinear RNNs with a comprehensive analysis on the advantages and limitations of these models. Then, based on closed-loop control theory, we propose a novel Bilinear RNN variant named Comba, which adopts a scalar-plus-low-rank state transition, with both state feedback and output feedback corrections. We also implement a hardware-efficient chunk-wise parallel kernel in Triton and train models with 340M/1.3B parameters on large-scale corpus. Comba demonstrates superior performance and computation efficiency in both language and vision modeling.

Jiaxi Hu、Yongqi Pan、Jusen Du、Disen Lan、Xiaqiang Tang、Qingsong Wen、Yuxuan Liang、Weigao Sun

自动化基础理论计算技术、计算机技术

Jiaxi Hu,Yongqi Pan,Jusen Du,Disen Lan,Xiaqiang Tang,Qingsong Wen,Yuxuan Liang,Weigao Sun.Comba: Improving Bilinear RNNs with Closed-loop Control[EB/OL].(2025-06-21)[2025-07-02].https://arxiv.org/abs/2506.02475.点此复制

评论