TY - GEN
T1 - EmoChildRu
T2 - 17th International Conference on Speech and Computer, SPECOM 2015
AU - Lyakso, Elena
AU - Frolova, Olga
AU - Dmitrieva, Evgeniya
AU - Grigorev, Aleksey
AU - Kaya, Heysem
AU - Salah, Albert Ali
AU - Karpov, Alexey
PY - 2015/1/1
Y1 - 2015/1/1
N2 - We present the first child emotional speech corpus in Russian, called “EmoChildRu”, which contains audio materials of 3–7 year old kids. The database includes over 20K recordings (approx. 30 h), collected from 100 children. Recordings were carried out in three controlled settings by creating different emotional states for children: playing with a standard set of toys; repetition of words from a toy-parrot in a game store setting; watching a cartoon and retelling of the story, respectively. This corpus is designed to study the reflection of the emotional state in the characteristics of voice and speech and for studies of the formation of emotional states in ontogenesis. A portion of the corpus is annotated for three emotional states (discomfort, neutral, comfort). Additional data include brain activity measurements (original EEG, evoked potentials records), the results of the adult listeners analysis of child speech, questionnaires, and description of dialogues. The paper reports two child emotional speech analysis experiments on the corpus: by adult listeners (humans) and by an automatic classifier (machine), respectively. Automatic classification results are very similar to human perception, although the accuracy is below 55% for both, showing the difficulty of child emotion recognition from speech under naturalistic conditions.
AB - We present the first child emotional speech corpus in Russian, called “EmoChildRu”, which contains audio materials of 3–7 year old kids. The database includes over 20K recordings (approx. 30 h), collected from 100 children. Recordings were carried out in three controlled settings by creating different emotional states for children: playing with a standard set of toys; repetition of words from a toy-parrot in a game store setting; watching a cartoon and retelling of the story, respectively. This corpus is designed to study the reflection of the emotional state in the characteristics of voice and speech and for studies of the formation of emotional states in ontogenesis. A portion of the corpus is annotated for three emotional states (discomfort, neutral, comfort). Additional data include brain activity measurements (original EEG, evoked potentials records), the results of the adult listeners analysis of child speech, questionnaires, and description of dialogues. The paper reports two child emotional speech analysis experiments on the corpus: by adult listeners (humans) and by an automatic classifier (machine), respectively. Automatic classification results are very similar to human perception, although the accuracy is below 55% for both, showing the difficulty of child emotion recognition from speech under naturalistic conditions.
KW - Computational paralinguistics
KW - Emotional child speech
KW - Emotional states
KW - Perceptual analysis
KW - Spectrographic analysis
UR - http://www.scopus.com/inward/record.url?scp=84945966891&partnerID=8YFLogxK
U2 - 10.1007/978-3-319-23132-7_18
DO - 10.1007/978-3-319-23132-7_18
M3 - Conference contribution
AN - SCOPUS:84945966891
SN - 9783319231310
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 144
EP - 152
BT - Speech and Computer - 17th International Conference, SPECOM 2015, Proceedings
A2 - Ronzhin, Andrey
A2 - Potapova, Rodmonga
A2 - Fakotakis, Nikos
PB - Springer
Y2 - 20 September 2015 through 24 September 2015
ER -