|国家预印本平台
首页|Analyzing the Importance of Blank for CTC-Based Knowledge Distillation

Analyzing the Importance of Blank for CTC-Based Knowledge Distillation

Analyzing the Importance of Blank for CTC-Based Knowledge Distillation

来源:Arxiv_logoArxiv
英文摘要

With the rise of large pre-trained foundation models for automatic speech recognition new challenges appear. While the performance of these models is good, runtime and cost of inference increases. One approach to make use of their strength while retaining efficiency is to distill their knowledge to smaller models during training. In this work, we explore different CTC-based distillation variants, focusing on blank token handling. We show that common approaches like blank elimination do not always work off the shelf. We explore new blank selection patterns as a potential sweet spot between standard knowledge distillation and blank elimination mechanisms. Through the introduction of a symmetric selection method, we are able to remove the CTC loss during knowledge distillation with minimal to no performance degradation. With this, we make the training independent from target labels, potentially allowing for distillation on untranscribed audio data.

Benedikt Hilmes、Nick Rossenbach、Ralf Schlüter

计算技术、计算机技术

Benedikt Hilmes,Nick Rossenbach,Ralf Schlüter.Analyzing the Importance of Blank for CTC-Based Knowledge Distillation[EB/OL].(2025-06-02)[2025-06-20].https://arxiv.org/abs/2506.01503.点此复制

评论