3.8 Proceedings Paper

A Comparative Study on Recent Neural Spoofing Countermeasures for Synthetic Speech Detection

Journal

INTERSPEECH 2021
Volume -, Issue -, Pages 4259-4263

Publisher

ISCA-INT SPEECH COMMUNICATION ASSOC
DOI: 10.21437/Interspeech.2021-702

Keywords

anti-spoofing; countermeasure; ASVspoof 2019; logical access; deep learning; significance test

Funding

  1. JST CREST Grants, Japan [JPMJCR18A6, JPMJCR20D3]
  2. MEXT KAKENHI Grants, Japan [16H06302, 18H04120, 18H04112, 18KT0051]
  3. Google AI for Japan program

Ask authors/readers for more resources

Recent research has focused on back-end neural networks and training criteria for speech spoofing countermeasures. This study offers a comparative perspective on various models and recognizes the potential impact of random initial seed on model performance. Promising techniques, including average pooling and a new hyper-parameter-free loss function, led to the best single model with significantly different statistical performance compared to others.
A great deal of recent research effort on speech spoofing countermeasures has been invested into back-end neural networks and training criteria. We contribute to this effort with a comparative perspective in this study. Our comparison of countermeasure models on the ASVspoof 2019 logical access scenario takes into account common strategies to deal with input trials of varied length, recently proposed margin-based training criteria, and widely used front ends. We also measured intra-model differences through multiple training-evaluation rounds with random initialization. Our statistical analysis demonstrates that the performance of the same model may be statistically significantly different when just changing the random initial seed. We thus recommend similar statistical analysis or reporting results of multiple runs for further research on the database. Despite the intra-model differences, we observed a few promising techniques, including average pooling, to efficiently process varied-length inputs and a new hyper-parameter-free loss function. The two techniques led to the best single model in our experiment, which achieved an equal error rate of 1.92% and was significantly different in statistical sense from most of the other experimental models.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available