An approach to measuring the performance of Automatic Speech Recognition (ASR) models in the context of Large Language Model (LLM) powered applications
An approach to measuring the performance of Automatic Speech Recognition (ASR) models in the context of Large Language Model (LLM) powered applications
Automatic Speech Recognition (ASR) plays a crucial role in human-machine interaction and serves as an interface for a wide range of applications. Traditionally, ASR performance has been evaluated using Word Error Rate (WER), a metric that quantifies the number of insertions, deletions, and substitutions in the generated transcriptions. However, with the increasing adoption of large and powerful Large Language Models (LLMs) as the core processing component in various applications, the significance of different types of ASR errors in downstream tasks warrants further exploration. In this work, we analyze the capabilities of LLMs to correct errors introduced by ASRs and propose a new measure to evaluate ASR performance for LLM-powered applications.
Sujith Pulikodan、Sahapthan K、Prasanta Kumar Ghosh、Visruth Sanka、Nihar Desai
计算技术、计算机技术自动化技术、自动化技术设备
Sujith Pulikodan,Sahapthan K,Prasanta Kumar Ghosh,Visruth Sanka,Nihar Desai.An approach to measuring the performance of Automatic Speech Recognition (ASR) models in the context of Large Language Model (LLM) powered applications[EB/OL].(2025-07-22)[2025-08-10].https://arxiv.org/abs/2507.16456.点此复制
评论