Journal
NEUROCOMPUTING
Volume 292, Issue -, Pages 104-110Publisher
ELSEVIER
DOI: 10.1016/j.neucom.2018.02.076
Keywords
Music auto-tagging; Deep learning; Music information retrieval; Recurrent Neural Network
Categories
Funding
- National Natural Science Foundation of China [11572084, 11472061, 71371046]
- Fundamental Research Funds for the Central Universities
- DHU Distinguished Young Professor Program [16D210404]
- Fundamental Research Funds for the Central Universities [CUSF-DH-D-2018097]
Ask authors/readers for more resources
Musical tags are used to describe music and are cruxes of music information retrieval. Existing methods for music auto-tagging usually consist of preprocessing phase (feature extraction) and machine learning phase. However, the preprocessing phase of most existing method is suffered either information loss or non-sufficient features, while the machine learning phase depends on heavily the feature extracted in the preprocessing phase, lacking the ability to make use of information. To solve this problem, we propose a content-based automatic tagging algorithm using deep Recurrent Neural Network (RNN) with scattering transformed inputs in this paper. Acting as the first phase, scattering transform extracts features from the raw data, meanwhile retains much more information than traditional methods such as mel-frequency cepstral coefficient (MFCC) and mel-frequency spectrogram. Five-layer RNNs with Gated Recurrent Unit (GRU) and sigmoid output layer are used as the second phase of our algorithm, which are extremely powerful machine learning tools capable of making full use of data fed to them. To evaluate the performance of the architecture, we experiment on Magnatagatune dataset using the measurement of the area under the ROC-curve (AUC-ROC). Experimental results show that the tagging performance can be boosted by the proposed method compared with the state-of-the-art models. Additionally, our architecture results in faster training speed and less memory usage. (c) 2018 Elsevier B.V. All rights reserved.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available