Eyes whisper depression: A CCA based multimodal approach

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

This paper presents our work on ACM MM Audio Visual Emotion Corpus 2013 (AVEC 2013) depression recognition sub-challenge using the baseline features in accordance with the challenge protocol. We use Canonical Correlation Analysis for audio-visual fusion as well as covariate extraction for the target task. The video baseline provides histograms of local phase quantization features extracted from 4×4=16 regions of the detected face. We summarize the video features over segments of length 20 seconds using mode and range functionals. We observe that features of range functional that measure the variance tendency provides statistically significantly higher canonical correlation than mode functional features that measure the mean tendency. Moreover, when audio-visual features are used with varying number of covariates per region, the regions that were consistently found the best are the ones corresponding to two eyes and the right part of the mouth. We reach 9.44 Root Mean Square Error on the challenge test set using audio-visual decision fusion, improving the video baseline 30% relative.

Original languageEnglish
Title of host publicationMM 2014 - Proceedings of the 2014 ACM Conference on Multimedia
PublisherAssociation for Computing Machinery
Pages961-964
Number of pages4
ISBN (Electronic)9781450330633
DOIs
Publication statusPublished - 1 Jan 2014
Event2014 ACM Conference on Multimedia, MM 2014 - Orlando, United States
Duration: 3 Nov 20147 Nov 2014

Conference

Conference2014 ACM Conference on Multimedia, MM 2014
Country/TerritoryUnited States
CityOrlando
Period3/11/147/11/14

Keywords

  • Audio-visual emotion corpus
  • Audio-visual fusion
  • Canonical Correlation Analysis
  • Depression recogni
  • Feature extraction

Fingerprint

Dive into the research topics of 'Eyes whisper depression: A CCA based multimodal approach'. Together they form a unique fingerprint.

Cite this