Proceedings of the XXth Conference of Open Innovations Association FRUCT (May 2023)

Enhancing Human-Computer Interaction through Emotion Recognition in Real-Life Speech

  • Eshwari Rao K P,
  • Anjan S,
  • Nagabhushan T N,
  • Kirill Krinkin,
  • Yulia Schichkina

DOI
https://doi.org/10.5281/zenodo.8005393
Journal volume & issue
Vol. 33, no. 2
pp. 415 – 417

Abstract

Read online

Extracting the data from real-life speech and recognizing emotions from them is one of the challenging tasks. This has gained popularity over the past few years. The goal of this study is to have a direct human-computer interaction (HCI) to determine the person’s condition or emotion through analyzing voice. The main purpose of this study is to extract emotion from the recorded audio as well as from the analysis of text extracted from the audio which is integrated together in a precise manner. The multimodal integration of speech and text gives good results by observing the emotional state of a person. Machine learning and deep learning algorithms are used to determine the emotional state of a person. The result of this study shows the accuracy around 62%.

Keywords