ASRJam: Human-Friendly AI Speech Jamming to Prevent Automated Phone Scams
ASRJam: Human-Friendly AI Speech Jamming to Prevent Automated Phone Scams
Large Language Models (LLMs), combined with Text-to-Speech (TTS) and Automatic Speech Recognition (ASR), are increasingly used to automate voice phishing (vishing) scams. These systems are scalable and convincing, posing a significant security threat. We identify the ASR transcription step as the most vulnerable link in the scam pipeline and introduce ASRJam, a proactive defence framework that injects adversarial perturbations into the victim's audio to disrupt the attacker's ASR. This breaks the scam's feedback loop without affecting human callers, who can still understand the conversation. While prior adversarial audio techniques are often unpleasant and impractical for real-time use, we also propose EchoGuard, a novel jammer that leverages natural distortions, such as reverberation and echo, that are disruptive to ASR but tolerable to humans. To evaluate EchoGuard's effectiveness and usability, we conducted a 39-person user study comparing it with three state-of-the-art attacks. Results show that EchoGuard achieved the highest overall utility, offering the best combination of ASR disruption and human listening experience.
Freddie Grabovski、Gilad Gressel、Yisroel Mirsky
通信无线通信
Freddie Grabovski,Gilad Gressel,Yisroel Mirsky.ASRJam: Human-Friendly AI Speech Jamming to Prevent Automated Phone Scams[EB/OL].(2025-06-10)[2025-06-24].https://arxiv.org/abs/2506.11125.点此复制
评论