4.6 Article

End-to-end music emotion variation detection using iteratively reconstructed deep features

期刊

MULTIMEDIA TOOLS AND APPLICATIONS
卷 81, 期 4, 页码 5017-5031

出版社

SPRINGER
DOI: 10.1007/s11042-021-11584-7

关键词

Music emotion recognition; Arousal; Valence; End-to-end deep learning; Bi-directional gated recurrent unit; Iterative reconstruction

向作者/读者索取更多资源

This study proposes a deep neural network-based solution for automatic music emotion recognition, which can extract emotion features directly from raw audio waveform and achieve high regression accuracy on the DEAM dataset.
Automatic music emotion recognition (MER) has received increased attention in areas of music information retrieval and user interface development. Music emotion variation detection (or dynamic MER) captures also temporal changes of emotion, and emotional content in music is expressed as a series of valence-arousal predictions. One of the issues in MER is extraction of emotional characteristics from audio signal. We propose a deep neural network based solution for mining music emotion-related salient features directly from raw audio waveform. The proposed architecture is based on stacking one-dimensional convolution layer, autoencoder-based layer with iterative reconstruction, and bidirectional gated recurrent unit. The tests on the DEAM dataset have shown that the proposed solution, in comparison with other state-of-the-art systems, can bring a significant improvement of the regression accuracy, notably for the valence dimension. It is shown that the proposed iterative reconstruction layer is able to enhance the discriminative properties of the features and further increase regression accuracy.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据