|国家预印本平台
首页|A Survey on Multimodal Music Emotion Recognition

A Survey on Multimodal Music Emotion Recognition

A Survey on Multimodal Music Emotion Recognition

来源:Arxiv_logoArxiv
英文摘要

Multimodal music emotion recognition (MMER) is an emerging discipline in music information retrieval that has experienced a surge in interest in recent years. This survey provides a comprehensive overview of the current state-of-the-art in MMER. Discussing the different approaches and techniques used in this field, the paper introduces a four-stage MMER framework, including multimodal data selection, feature extraction, feature processing, and final emotion prediction. The survey further reveals significant advancements in deep learning methods and the increasing importance of feature fusion techniques. Despite these advancements, challenges such as the need for large annotated datasets, datasets with more modalities, and real-time processing capabilities remain. This paper also contributes to the field by identifying critical gaps in current research and suggesting potential directions for future research. The gaps underscore the importance of developing robust, scalable, a interpretable models for MMER, with implications for applications in music recommendation systems, therapeutic tools, and entertainment.

Rashini Liyanarachchi、Aditya Joshi、Erik Meijering

计算技术、计算机技术

Rashini Liyanarachchi,Aditya Joshi,Erik Meijering.A Survey on Multimodal Music Emotion Recognition[EB/OL].(2025-04-26)[2025-06-06].https://arxiv.org/abs/2504.18799.点此复制

评论