Abstract
In this paper, we present a hierarchical framework for complex paralinguistic analysis of speech including gender, emotions and deception recognition. The main idea of the framework is built upon the research on interrelation between various paralinguistic phenomena. It uses gender information to predict emotional states, and the outcome of the emotion recognition to predict the truthfulness of the speech. We use multiple datasets (aGender, Ruslana, EmoDB and DSD) to perform within-corpus and cross-corpus experiments using various performance measures. The experimental results reveal that gender-specific models improve the effectiveness of automatic speech emotion recognition in terms of Unweighted Average Recall up to an absolute 5.7%, and the integration of emotion predictions improves the F-score of automatic deception detection compared to our baseline by an absolute 4.7%. The obtained cross-validation results of 88.4 +/- 1.5% for deception detection beat the existing state-of-the-art by an absolute 2.8%.
Original language | English |
---|---|
Pages | 4735-4739 |
Number of pages | 5 |
DOIs | |
Publication status | Published - 18 Sept 2022 |
Event | INTERSPEECH 2022 - Incheon, Korea, Incheon, Korea, Republic of Duration: 18 Sept 2022 → 22 Sept 2022 https://www.interspeech2022.org/ |
Conference
Conference | INTERSPEECH 2022 |
---|---|
Abbreviated title | INTERSPEECH 2022 |
Country/Territory | Korea, Republic of |
City | Incheon |
Period | 18/09/22 → 22/09/22 |
Internet address |
Bibliographical note
Funding Information:This work was supported by the Analytical Center for the Government of the Russian Federation (IGK 000000D730321P5Q0002), agreement No. 70-2021-00141.
Publisher Copyright:
Copyright © 2022 ISCA.
Keywords
- Computational Paralinguistics
- Gender Recognition
- Emotion Recognition
- Deception Detection