|国家预印本平台
首页|Scaling and Distilling Transformer Models for sEMG

Scaling and Distilling Transformer Models for sEMG

Scaling and Distilling Transformer Models for sEMG

来源:Arxiv_logoArxiv
英文摘要

Surface electromyography (sEMG) signals offer a promising avenue for developing innovative human-computer interfaces by providing insights into muscular activity. However, the limited volume of training data and computational constraints during deployment have restricted the investigation of scaling up the model size for solving sEMG tasks. In this paper, we demonstrate that vanilla transformer models can be effectively scaled up on sEMG data and yield improved cross-user performance up to 110M parameters, surpassing the model size regime investigated in other sEMG research (usually <10M parameters). We show that >100M-parameter models can be effectively distilled into models 50x smaller with minimal loss of performance (<1.5% absolute). This results in efficient and expressive models suitable for complex real-time sEMG tasks in real-world environments.

Nicholas Mehlman、Jean-Christophe Gagnon-Audet、Michael Shvartsman、Kelvin Niu、Alexander H. Miller、Shagun Sodhani

计算技术、计算机技术

Nicholas Mehlman,Jean-Christophe Gagnon-Audet,Michael Shvartsman,Kelvin Niu,Alexander H. Miller,Shagun Sodhani.Scaling and Distilling Transformer Models for sEMG[EB/OL].(2025-07-29)[2025-08-21].https://arxiv.org/abs/2507.22094.点此复制

评论