3.8 Proceedings Paper

The VoiceMOS Challenge 2022

期刊

INTERSPEECH 2022
卷 -, 期 -, 页码 4536-4540

出版社

ISCA-INT SPEECH COMMUNICATION ASSOC
DOI: 10.21437/Interspeech.2022-970

关键词

VoiceMOS Challenge; synthetic speech evaluation; mean opinion score; automatic speech quality prediction

资金

  1. JSPS KAKENHI [21J20920]
  2. JST CREST [JPMJCR19A3, JPMJCR18A6, JPMJCR20D3]
  3. MEXT KAKENHI [21K11951, 21K19808]
  4. MOST [107-2221-E-001-008-MY3]

向作者/读者索取更多资源

The VoiceMOS Challenge aims to promote the study of automatic prediction of the mean opinion score (MOS) of synthetic speech. Through this challenge, 22 participating teams from academia and industry tested various approaches to predict human ratings of synthesized speech. The results highlight the effectiveness of fine-tuning self-supervised speech models for MOS prediction, as well as the challenges in predicting MOS ratings for unseen speakers, listeners, and systems in the out-of-domain setting.
We present the first edition of the VoiceMOS Challenge, a scientific event that aims to promote the study of automatic prediction of the mean opinion score (MOS) of synthetic speech. This challenge drew 22 participating teams from academia and industry who tried a variety of approaches to tackle the problem of predicting human ratings of synthesized speech. The listening test data for the main track of the challenge consisted of samples from 187 different text-to-speech and voice conversion systems spanning over a decade of research, and the out-of-domain track consisted of data from more recent systems rated in a separate listening test. Results of the challenge show the effectiveness of fine-tuning self-supervised speech models for the MOS prediction task, as well as the difficulty of predicting MOS ratings for unseen speakers and listeners, and for unseen systems in the out-of-domain setting.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据