CMU's IWSLT 2025 Simultaneous Speech Translation System
CMU's IWSLT 2025 Simultaneous Speech Translation System
This paper presents CMU's submission to the IWSLT 2025 Simultaneous Speech Translation (SST) task for translating unsegmented English speech into Chinese and German text in a streaming manner. Our end-to-end speech-to-text system integrates a chunkwise causal Wav2Vec 2.0 speech encoder, an adapter, and the Qwen2.5-7B-Instruct as the decoder. We use a two-stage simultaneous training procedure on robust speech segments curated from LibriSpeech, CommonVoice, and VoxPopuli datasets, utilizing standard cross-entropy loss. Our model supports adjustable latency through a configurable latency multiplier. Experimental results demonstrate that our system achieves 44.3 BLEU for English-to-Chinese and 25.1 BLEU for English-to-German translations on the ACL60/60 development set, with computation-aware latencies of 2.7 seconds and 2.3 seconds, and theoretical latencies of 2.2 and 1.7 seconds, respectively.
Siqi Ouyang、Xi Xu、Lei Li
常用外国语计算技术、计算机技术
Siqi Ouyang,Xi Xu,Lei Li.CMU's IWSLT 2025 Simultaneous Speech Translation System[EB/OL].(2025-06-16)[2025-06-27].https://arxiv.org/abs/2506.13143.点此复制
评论