Canonical Correlation Analysis for Data Fusion in Multimodal Emotion Recognition
Multimodal emotion recognition systems aim at classifying emotion data, usually from different natures, into discrete affective categories. These systems fuse different modalities as each modality classifies the data from its own viewpoint and can compensate the limitations of others when combining...
Saved in:
| Published in: | 2018 9th International Symposium on Telecommunications (IST) pp. 676 - 681 |
|---|---|
| Main Author: | |
| Format: | Conference Proceeding |
| Language: | English |
| Published: |
IEEE
01.12.2018
|
| Subjects: | |
| Online Access: | Get full text |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Summary: | Multimodal emotion recognition systems aim at classifying emotion data, usually from different natures, into discrete affective categories. These systems fuse different modalities as each modality classifies the data from its own viewpoint and can compensate the limitations of others when combining with them. Existing approaches for multimodal data fusion either use feature-level or decision-level fusion. The former needs different modalities to be synchronized while the latter has not this limitation. In the current study, audio, visual, and users' comments are used as modalities for video emotion recognition. Although the first two modalities are synchronized, users' comments are not synchronized with them and this makes the use of pure feature-level data fusion impossible. In order to exploit the benefits of the feature-level approach, a hybrid method is proposed in the current study which first applies feature-level canonical correlation analysis (CCA) to audio and visual modalities and then combines the outputs with users' comments using a decision-level fusion. The results of applying this proposed method to DEAP dataset shows that using feature-level CCA not only outperforms the baseline feature-level method but also achieves a higher performance than the decision-level fusion. |
|---|---|
| DOI: | 10.1109/ISTEL.2018.8661140 |