Applied Sciences (May 2024)

Synthesis of Non-Linguistic Utterances for Sound Design Support Using a Genetic Algorithm

  • Ahmed Khota,
  • Eric W. Cooper,
  • Yu Yan

DOI
https://doi.org/10.3390/app14114572
Journal volume & issue
Vol. 14, no. 11
p. 4572

Abstract

Read online

As social robots become more prevalent, they often employ non-speech sounds, in addition to other modes of communication, to communicate emotion and intention in an increasingly complex visual and audio environment. These non-speech sounds are usually tailor-made, and research into the generation of non-speech sounds that can convey emotions has been limited. To enable social robots to use a large amount of non-speech sounds in a natural and dynamic way, while expressing a wide range of emotions effectively, this work proposes an automatic method of sound generation using a genetic algorithm, coupled with a random forest model trained on representative non-speech sounds to validate each produced sound’s ability to express emotion. The sounds were tested in an experiment wherein subjects rated the perceived valence and arousal. Statistically significant clusters of sounds in the valence arousal space corresponded to different emotions, showing that the proposed method generates sounds that can readily be used in social robots.

Keywords