4.5 Article

Analysis of acoustic and voice quality features for the classification of infant and mother vocalizations

Journal

SPEECH COMMUNICATION
Volume 133, Issue -, Pages 41-61

Publisher

ELSEVIER
DOI: 10.1016/j.specom.2021.07.010

Keywords

Emotion classifier; Infant vocalizations; Infant-directed speech; Convolutional neural networks; Self-attention; Global feature; Feature selection

Funding

  1. National Institute on Drug Abuse, United States [R34DA050256]
  2. National Institute of Mental Health, United States [R21MH112578]
  3. National Institute of Food and Agriculture, U.S. Department of Agriculture [ILLU-793-339]

Ask authors/readers for more resources

This study focused on classifying infant and parent vocalizations to understand how infants learn to regulate emotions in social interactions. The results showed that neural networks outperformed linear discriminant analysis in terms of classification accuracy, with the convolutional neural network slightly outperforming the fully-connected network.
Classification of infant and parent vocalizations, particularly emotional vocalizations, is critical to understand-ing how infants learn to regulate emotions in social dyadic processes. This work is an experimental study of classifiers, features, and data augmentation strategies applied to the task of classifying infant and parent vocalization types. Our data were recorded both in the home and in the laboratory. Infant vocalizations were manually labeled as CRY, FUS (fuss), LAU (laugh), BAB (babble) or SCR (screech), while parent (mostly mother) vocalizations were labeled as IDS (infant-directed speech), ADS (adult-directed speech), PLA (playful), RHY (rhythmic speech or singing), LAU (laugh) or WHI (whisper). Linear discriminant analysis (LDA) was selected as a baseline classifier, because it gave the highest accuracy in a previously published study covering part of this corpus. LDA was compared to two neural network architectures: a two-layer fully-connected network (FCN), and a convolutional neural network with self-attention (CNSA). Baseline features extracted using the OpenSMILE toolkit were augmented by extra voice quality, phonetic, and prosodic features, each targeting perceptual features of one or more of the vocalization types. Three web data augmentation and transfer learning methods were tested: pre-training of network weights for a related task (adult emotion classification), augmentation of under-represented classes using data uniformly sampled from other corpora, and augmentation of under-represented classes using data selected by a minimum cross-corpus information difference criterion. Feature selection using Fisher scores and experiments of using weighted and unweighted samplers were also tested. Two datasets were evaluated: a benchmark dataset (CRIED) and our own corpus. In terms of unweighted-average recall of CRIED dataset, the CNSA achieved the best UAR compared with previous studies. In terms of classification accuracy, weighted F1, and macro F1 of our own dataset, the neural networks both significantly outperformed LDA; the FCN slightly (but not significantly) outperformed the CNSA. Cross-examining features selected by different feature selection algorithms permits a type of post-hoc feature analysis, in which the most important acoustic features for each binary type discrimination are listed. Examples of each vocalization type of overlapped features were selected, and their spectrograms are presented, and discussed with respect to the type-discriminative acoustic features selected by various algorithms. MFCC, log Mel Frequency Band Energy, LSP frequency, and F1 are found to be the most important spectral envelope features; F0 is found to be the most important prosodic feature.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available