4.6 Article

Music auto-tagging using deep Recurrent Neural Networks

Journal

NEUROCOMPUTING
Volume 292, Issue -, Pages 104-110

Publisher

ELSEVIER
DOI: 10.1016/j.neucom.2018.02.076

Keywords

Music auto-tagging; Deep learning; Music information retrieval; Recurrent Neural Network

Funding

  1. National Natural Science Foundation of China [11572084, 11472061, 71371046]
  2. Fundamental Research Funds for the Central Universities
  3. DHU Distinguished Young Professor Program [16D210404]
  4. Fundamental Research Funds for the Central Universities [CUSF-DH-D-2018097]

Ask authors/readers for more resources

Musical tags are used to describe music and are cruxes of music information retrieval. Existing methods for music auto-tagging usually consist of preprocessing phase (feature extraction) and machine learning phase. However, the preprocessing phase of most existing method is suffered either information loss or non-sufficient features, while the machine learning phase depends on heavily the feature extracted in the preprocessing phase, lacking the ability to make use of information. To solve this problem, we propose a content-based automatic tagging algorithm using deep Recurrent Neural Network (RNN) with scattering transformed inputs in this paper. Acting as the first phase, scattering transform extracts features from the raw data, meanwhile retains much more information than traditional methods such as mel-frequency cepstral coefficient (MFCC) and mel-frequency spectrogram. Five-layer RNNs with Gated Recurrent Unit (GRU) and sigmoid output layer are used as the second phase of our algorithm, which are extremely powerful machine learning tools capable of making full use of data fed to them. To evaluate the performance of the architecture, we experiment on Magnatagatune dataset using the measurement of the area under the ROC-curve (AUC-ROC). Experimental results show that the tagging performance can be boosted by the proposed method compared with the state-of-the-art models. Additionally, our architecture results in faster training speed and less memory usage. (c) 2018 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available