Video-based emotion recognition in the wild

Albert Ali Salah, Heysem Kaya, Furkan Gürpınar

Research output: Chapter in Book/Report/Conference proceedingChapterAcademicpeer-review

Abstract

In-the-wild emotion recognition requires dealing with large variances in input signals, multiple sources of noise that will distract the learners, as well as difficult annotation and ground truth acquisition conditions. In this chapter, we briefly survey the latest developments in multimodal approaches for video-based emotion recognition in the wild, and describe our approach to the problem. For the visual modality, we propose using summarizing functionals of complementary visual descriptors. For the audio modality, we propose a standard computational pipeline for paralinguistics. We combine audio and visual features with least squares regression-based classifiers and weighted score-level fusion. We report state-of-the-art results on the EmotiW Challenge for “in-the-wild” facial-expression recognition. Our approach scales to other problems, and ranked top in two challenges; the ChaLearn-LAP First Impressions Challenge (ICPR'2016) and ChaLearn-LAP Job Interview Candidate Screening Challenge (CVPR'2017), respectively.
Original languageEnglish
Title of host publicationMultimodal behavior analysis in the wild
Subtitle of host publicationadvances and challenges
EditorsXavier Alameda-Pineda, Elisa Ricci, Nicu Sebe
Place of PublicationLondon
PublisherAcademic Press
Pages369-386
Number of pages18
ISBN (Electronic)9780128146026
ISBN (Print)9780128146019
DOIs
Publication statusPublished - 2 Jan 2019

Publication series

NameComputer vision and pattern recognition series

Fingerprint

Dive into the research topics of 'Video-based emotion recognition in the wild'. Together they form a unique fingerprint.

Cite this