Abstract
In this paper, we present our approach for the Emotional Impact of Movies task of Mediaeval 2017 Challenge, involving multimodal fusion for predicting arousal and valence for movie clips. In our system, we have two pipelines. In the first one, we extracted audio/visual features, and used a combination of PCA, Fisher vector encoding, feature selection, and extreme learning machine classifiers. In the second one, we focused on the classifiers, rather than on feature selection.
| Original language | English |
|---|---|
| Publication status | Published - 12 Sept 2017 |
Publication series
| Name | CEUR Workshop Proceedings |
|---|---|
| Publisher | CEUR WS |
| ISSN (Print) | 1613-0073 |
Funding
This work is supported by Bogazici University Project BAP 16A01P4 and by the BAGEP Award of the Science Academy.