IEEE Access (Jan 2019)

The Fusion of Electroencephalography and Facial Expression for Continuous Emotion Recognition

  • Dahua Li,
  • Zhe Wang,
  • Chuhan Wang,
  • Shuang Liu,
  • Wenhao Chi,
  • Enzeng Dong,
  • Xiaolin Song,
  • Qiang Gao,
  • Yu Song

DOI
https://doi.org/10.1109/ACCESS.2019.2949707
Journal volume & issue
Vol. 7
pp. 155724 – 155736

Abstract

Read online

Recently, the study of emotion recognition has received increasing attentions by the rapid development of noninvasive sensor technologies, machine learning algorithms and compute capability of computers. Compared with single modal emotion recognition, the multimodal paradigm introduces complementary information for emotion recognition. Hence, in this work, we presented a decision level fusion framework for detecting emotions continuously by fusing the Electroencephalography (EEG) and facial expressions. Three types of movie clips (positive, negative, and neutral) were utilized to elicit specific emotions of subjects, the EEG and facial expression signals were recorded simultaneously. The power spectrum density (PSD) features of EEG were extracted by time-frequency analysis, and then EEG features were selected for regression. For the facial expression, the facial geometric features were calculated by facial landmark localization. Long short-term memory networks (LSTM) were utilized to accomplish the decision level fusion and captured temporal dynamics of emotions. The results have shown that the proposed method achieved outstanding performance for continuous emotion recognition, and it yields 0.625± 0.029 of concordance correlation coefficient (CCC). From the results, the fusion of two modalities outperformed EEG and facial expression separately. Furthermore, different numbers of time-steps of LSTM was applied to analyze the temporal dynamic capturing.

Keywords