Deep Learning Based Data Fusion Methods for Multimodal Emotion Recognition

Judith Nkechinyere Njoku, Angela C. Caliwag, Wansu Lim, Sangho Kim, Han Jeong Hwang, Jin Woo Jeong

Research output: Contribution to journalArticlepeer-review

10 Scopus citations

Abstract

Multimodal emotion recognition is a robust and reliable method as it utilizes multimodal data for more comprehensive representation of emotions. Data fusion is a key step in multimodal emotion recognition, because the accuracy of the recognition model mostly depends on how the different modalities are combined. The goal of this paper is to compare the performances of deep learning (DL) based models for the task of data fusion and multimodal emotion recognition. The contributions of this paper are two folds: 1) We introduce three DL models for multimodal fusion and classification: early fusion, hybrid fusion, and multi-task learning. 2) We systematically compare the performance of these models on three multimodal datasets. Our experimental results demonstrate that multi-task learning achieves the best results across all modalities; 75.41%, 68.33%, and 78.75% for classification of three emotional states using the combinations of audio-visual, EEG-audio, and EEG-visual data, respectively.

Original languageEnglish
Pages (from-to)79-87
Number of pages9
JournalJournal of Korean Institute of Communications and Information Sciences
Volume47
Issue number1
DOIs
StatePublished - Jan 2022

Keywords

  • Data-fusion
  • deep learning
  • EEG
  • emotion recognition
  • multimodal

Fingerprint

Dive into the research topics of 'Deep Learning Based Data Fusion Methods for Multimodal Emotion Recognition'. Together they form a unique fingerprint.

Cite this