TY - JOUR
T1 - Analysis of acoustic and voice quality features for the classification of infant and mother vocalizations
AU - Li, Jialu
AU - Hasegawa-Johnson, Mark
AU - McElwain, Nancy L.
N1 - Funding Information:
This work was supported by funding from the National Institute on Drug Abuse, United States ( R34DA050256 ), the National Institute of Mental Health, United States ( R21MH112578 ) and the National Institute of Food and Agriculture, U.S. Department of Agriculture ( ILLU-793-339 ).
Funding Information:
This work was supported by funding from the National Institute on Drug Abuse, United States (R34DA050256), the National Institute of Mental Health, United States (R21MH112578) and the National Institute of Food and Agriculture, U.S. Department of Agriculture (ILLU-793-339). We also thank Dr.in Katrin D. Bartl-Pokorny, Univ. Prof.in Dr.in Christa Einspieler, Assoc. Prof. DDr. Peter B. Marschik, Dr. Florian B. Pokorny, Dr.in Dajie Zhang at the interdisciplinary Developmental Neuroscience, Division of Phoniatrics, Medical University of Graz, Austria, and any other contributors for their input of study design and data development of the CRIED database.
Publisher Copyright:
© 2021 The Author(s)
PY - 2021/10
Y1 - 2021/10
N2 - Classification of infant and parent vocalizations, particularly emotional vocalizations, is critical to understanding how infants learn to regulate emotions in social dyadic processes. This work is an experimental study of classifiers, features, and data augmentation strategies applied to the task of classifying infant and parent vocalization types. Our data were recorded both in the home and in the laboratory. Infant vocalizations were manually labeled as CRY, FUS (fuss), LAU (laugh), BAB (babble) or SCR (screech), while parent (mostly mother) vocalizations were labeled as IDS (infant-directed speech), ADS (adult-directed speech), PLA (playful), RHY (rhythmic speech or singing), LAU (laugh) or WHI (whisper). Linear discriminant analysis (LDA) was selected as a baseline classifier, because it gave the highest accuracy in a previously published study covering part of this corpus. LDA was compared to two neural network architectures: a two-layer fully-connected network (FCN), and a convolutional neural network with self-attention (CNSA). Baseline features extracted using the OPENSMILE toolkit were augmented by extra voice quality, phonetic, and prosodic features, each targeting perceptual features of one or more of the vocalization types. Three web data augmentation and transfer learning methods were tested: pre-training of network weights for a related task (adult emotion classification), augmentation of under-represented classes using data uniformly sampled from other corpora, and augmentation of under-represented classes using data selected by a minimum cross-corpus information difference criterion. Feature selection using Fisher scores and experiments of using weighted and unweighted samplers were also tested. Two datasets were evaluated: a benchmark dataset (CRIED) and our own corpus. In terms of unweighted-average recall of CRIED dataset, the CNSA achieved the best UAR compared with previous studies. In terms of classification accuracy, weighted F1, and macro F1 of our own dataset, the neural networks both significantly outperformed LDA; the FCN slightly (but not significantly) outperformed the CNSA. Cross-examining features selected by different feature selection algorithms permits a type of post-hoc feature analysis, in which the most important acoustic features for each binary type discrimination are listed. Examples of each vocalization type of overlapped features were selected, and their spectrograms are presented, and discussed with respect to the type-discriminative acoustic features selected by various algorithms. MFCC, log Mel Frequency Band Energy, LSP frequency, and F1 are found to be the most important spectral envelope features; F0 is found to be the most important prosodic feature.
AB - Classification of infant and parent vocalizations, particularly emotional vocalizations, is critical to understanding how infants learn to regulate emotions in social dyadic processes. This work is an experimental study of classifiers, features, and data augmentation strategies applied to the task of classifying infant and parent vocalization types. Our data were recorded both in the home and in the laboratory. Infant vocalizations were manually labeled as CRY, FUS (fuss), LAU (laugh), BAB (babble) or SCR (screech), while parent (mostly mother) vocalizations were labeled as IDS (infant-directed speech), ADS (adult-directed speech), PLA (playful), RHY (rhythmic speech or singing), LAU (laugh) or WHI (whisper). Linear discriminant analysis (LDA) was selected as a baseline classifier, because it gave the highest accuracy in a previously published study covering part of this corpus. LDA was compared to two neural network architectures: a two-layer fully-connected network (FCN), and a convolutional neural network with self-attention (CNSA). Baseline features extracted using the OPENSMILE toolkit were augmented by extra voice quality, phonetic, and prosodic features, each targeting perceptual features of one or more of the vocalization types. Three web data augmentation and transfer learning methods were tested: pre-training of network weights for a related task (adult emotion classification), augmentation of under-represented classes using data uniformly sampled from other corpora, and augmentation of under-represented classes using data selected by a minimum cross-corpus information difference criterion. Feature selection using Fisher scores and experiments of using weighted and unweighted samplers were also tested. Two datasets were evaluated: a benchmark dataset (CRIED) and our own corpus. In terms of unweighted-average recall of CRIED dataset, the CNSA achieved the best UAR compared with previous studies. In terms of classification accuracy, weighted F1, and macro F1 of our own dataset, the neural networks both significantly outperformed LDA; the FCN slightly (but not significantly) outperformed the CNSA. Cross-examining features selected by different feature selection algorithms permits a type of post-hoc feature analysis, in which the most important acoustic features for each binary type discrimination are listed. Examples of each vocalization type of overlapped features were selected, and their spectrograms are presented, and discussed with respect to the type-discriminative acoustic features selected by various algorithms. MFCC, log Mel Frequency Band Energy, LSP frequency, and F1 are found to be the most important spectral envelope features; F0 is found to be the most important prosodic feature.
KW - Convolutional neural networks
KW - Emotion classifier
KW - Feature selection
KW - Global feature
KW - Infant vocalizations
KW - Infant-directed speech
KW - Self-attention
UR - http://www.scopus.com/inward/record.url?scp=85113414034&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85113414034&partnerID=8YFLogxK
U2 - 10.1016/j.specom.2021.07.010
DO - 10.1016/j.specom.2021.07.010
M3 - Article
C2 - 36062214
AN - SCOPUS:85113414034
SN - 0167-6393
VL - 133
SP - 41
EP - 61
JO - Speech Communication
JF - Speech Communication
ER -