Applied Sciences (Jul 2020)

Gated Recurrent Attention for Multi-Style Speech Synthesis

  • Sung Jun Cheon,
  • Joun Yeop Lee,
  • Byoung Jin Choi,
  • Hyeonseung Lee,
  • Nam Soo Kim

DOI
https://doi.org/10.3390/app10155325
Journal volume & issue
Vol. 10, no. 15
p. 5325

Abstract

Read online

End-to-end neural network-based speech synthesis techniques have been developed to represent and synthesize speech in various prosodic style. Although the end-to-end techniques enable the transfer of a style with a single vector of style representation, it has been reported that the speaker similarity observed from the speech synthesized with unseen speaker-style is low. One of the reasons for this problem is that the attention mechanism in the end-to-end model is overfitted to the training data. To learn and synthesize voices of various styles, an attention mechanism that can preserve longer-term context and control the context is required. In this paper, we propose a novel attention model which employs gates to control the recurrences in the attention. To verify the proposed attention’s style modeling capability, perceptual listening tests were conducted. The experiments show that the proposed attention outperforms the location-sensitive attention in both similarity and naturalness.

Keywords