|国家预印本平台
首页|Efficient and Adaptive Simultaneous Speech Translation with Fully Unidirectional Architecture

Efficient and Adaptive Simultaneous Speech Translation with Fully Unidirectional Architecture

Efficient and Adaptive Simultaneous Speech Translation with Fully Unidirectional Architecture

来源:Arxiv_logoArxiv
英文摘要

Simultaneous speech translation (SimulST) produces translations incrementally while processing partial speech input. Although large language models (LLMs) have showcased strong capabilities in offline translation tasks, applying them to SimulST poses notable challenges. Existing LLM-based SimulST approaches either incur significant computational overhead due to repeated encoding of bidirectional speech encoder, or they depend on a fixed read/write policy, limiting the efficiency and performance. In this work, we introduce Efficient and Adaptive Simultaneous Speech Translation (EASiST) with fully unidirectional architecture, including both speech encoder and LLM. EASiST includes a multi-latency data curation strategy to generate semantically aligned SimulST training samples and redefines SimulST as an interleaved generation task with explicit read/write tokens. To facilitate adaptive inference, we incorporate a lightweight policy head that dynamically predicts read/write actions. Additionally, we employ a multi-stage training strategy to align speech-text modalities and optimize both translation and policy behavior. Experiments on the MuST-C En$\rightarrow$De and En$\rightarrow$Es datasets demonstrate that EASiST offers superior latency-quality trade-offs compared to several strong baselines.

Biao Fu、Donglei Yu、Minpeng Liao、Chengxi Li、Yidong Chen、Kai Fan、Xiaodong Shi

计算技术、计算机技术

Biao Fu,Donglei Yu,Minpeng Liao,Chengxi Li,Yidong Chen,Kai Fan,Xiaodong Shi.Efficient and Adaptive Simultaneous Speech Translation with Fully Unidirectional Architecture[EB/OL].(2025-04-16)[2025-05-06].https://arxiv.org/abs/2504.11809.点此复制

评论