Electronic Research Archive (Feb 2022)

Emotion recognition in talking-face videos using persistent entropy and neural networks

  • Eduardo Paluzo-Hidalgo ,
  • Rocio Gonzalez-Diaz ,
  • Guillermo Aguirre-Carrazana

DOI
https://doi.org/10.3934/era.2022034
Journal volume & issue
Vol. 30, no. 2
pp. 644 – 660

Abstract

Read online

The automatic recognition of a person's emotional state has become a very active research field that involves scientists specialized in different areas such as artificial intelligence, computer vision, or psychology, among others. Our main objective in this work is to develop a novel approach, using persistent entropy and neural networks as main tools, to recognise and classify emotions from talking-face videos. Specifically, we combine audio-signal and image-sequence information to compute a topology signature (a 9-dimensional vector) for each video. We prove that small changes in the video produce small changes in the signature, ensuring the stability of the method. These topological signatures are used to feed a neural network to distinguish between the following emotions: calm, happy, sad, angry, fearful, disgust, and surprised. The results reached are promising and competitive, beating the performances achieved in other state-of-the-art works found in the literature.

Keywords