4.5 Article

Speech emotion recognition based on transfer learning from the FaceNet frameworka)

期刊

JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA
卷 149, 期 2, 页码 1338-1345

出版社

ACOUSTICAL SOC AMER AMER INST PHYSICS
DOI: 10.1121/10.0003530

关键词

-

资金

  1. Jilin Provincial Science and Technology Department [20180201003GX]
  2. Jilin Province Development and Reform Commission [2019C053-4]

向作者/读者索取更多资源

Speech plays a crucial role in human-computer emotional interaction, and this study utilizes the FaceNet model to improve speech emotion recognition. By pretraining on the CASIA dataset and fine-tuning on the IEMOCAP dataset, the proposed approach achieves high accuracy due to clean signals. Experimental results demonstrate that the method outperforms state-of-the-art approaches on the IEMOCAP dataset among single modal methods.
Speech plays an important role in human-computer emotional interaction. FaceNet used in face recognition achieves great success due to its excellent feature extraction. In this study, we adopt the FaceNet model and improve it for speech emotion recognition. To apply this model for our work, speech signals are divided into segments at a given time interval, and the signal segments are transformed into a discrete waveform diagram and spectrogram. Subsequently, the waveform and spectrogram are separately fed into FaceNet for end-to-end training. Our empirical study shows that the pretraining is effective on the spectrogram for FaceNet. Hence, we pretrain the network on the CASIA dataset and then fine-tune it on the IEMOCAP dataset with waveforms. It will derive the maximum transfer learning knowledge from the CASIA dataset due to its high accuracy. This high accuracy may be due to its clean signals. Our preliminary experimental results show an accuracy of 68.96% and 90% on the emotion benchmark datasets IEMOCAP and CASIA, respectively. The cross-training is then conducted on the dataset, and comprehensive experiments are performed. Experimental results indicate that the proposed approach outperforms state-of-the-art methods on the IEMOCAP dataset among single modal approaches.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据