|国家预印本平台
首页|Beyond Gloss: A Hand-Centric Framework for Gloss-Free Sign Language Translation

Beyond Gloss: A Hand-Centric Framework for Gloss-Free Sign Language Translation

Beyond Gloss: A Hand-Centric Framework for Gloss-Free Sign Language Translation

来源:Arxiv_logoArxiv
英文摘要

Sign Language Translation (SLT) is a challenging task that requires bridging the modality gap between visual and linguistic information while capturing subtle variations in hand shapes and movements. To address these challenges, we introduce \textbf{BeyondGloss}, a novel gloss-free SLT framework that leverages the spatio-temporal reasoning capabilities of Video Large Language Models (VideoLLMs). Since existing VideoLLMs struggle to model long videos in detail, we propose a novel approach to generate fine-grained, temporally-aware textual descriptions of hand motion. A contrastive alignment module aligns these descriptions with video features during pre-training, encouraging the model to focus on hand-centric temporal dynamics and distinguish signs more effectively. To further enrich hand-specific representations, we distill fine-grained features from HaMeR. Additionally, we apply a contrastive loss between sign video representations and target language embeddings to reduce the modality gap in pre-training. \textbf{BeyondGloss} achieves state-of-the-art performance on the Phoenix14T and CSL-Daily benchmarks, demonstrating the effectiveness of the proposed framework. We will release the code upon acceptance of the paper.

Sobhan Asasi、Mohamed Ilyas Lakhal、Ozge Mercanoglu Sincan、Richard Bowden

计算技术、计算机技术

Sobhan Asasi,Mohamed Ilyas Lakhal,Ozge Mercanoglu Sincan,Richard Bowden.Beyond Gloss: A Hand-Centric Framework for Gloss-Free Sign Language Translation[EB/OL].(2025-07-31)[2025-08-07].https://arxiv.org/abs/2507.23575.点此复制

评论