首页> 外文会议>2018 IEEE Spoken Language Technology Workshop >Parameter Generation Algorithms for Text-To-Speech Synthesis with Recurrent Neural Networks
【24h】

Parameter Generation Algorithms for Text-To-Speech Synthesis with Recurrent Neural Networks

机译:递归神经网络的文本到语音合成的参数生成算法

获取原文
获取原文并翻译 | 示例

摘要

Recurrent Neural Networks (RNN) have recently proved to be effective in acoustic modeling for TTS. Various techniques such as the Maximum Likelihood Parameter Generation (MLPG) algorithm have been naturally inherited from the HMM-based speech synthesis framework. This paper investigates in which situations parameter generation and variance restoration approaches help for RNN-based TTS. We explore how their performance is affected by various factors such as the choice of the loss function, the application of regularization methods and the amount of training data. We propose an efficient way to calculate MLPG using a convolutional kernel. Our results show that the use of the L1 loss with proper regularization outperforms any system built with the conventional L2 loss and does not require to apply MLPG (which is necessary otherwise). We did not observe perceptual improvements when embedding MLPG into the acoustic model. Finally, we show that variance restoration approaches are important for cepstral features but only yield minor perceptual gains for the prediction of F0.
机译:最近证明,递归神经网络(RNN)在TTS的声学建模中很有效。自然地从基于HMM的语音合成框架继承了各种技术,例如最大似然参数生成(MLPG)算法。本文研究了在哪些情况下参数生成和方差恢复方法有助于基于RNN的TTS。我们探讨了其性能如何受到各种因素的影响,例如损失函数的选择,正则化方法的应用和训练数据的数量。我们提出了一种使用卷积核来计算MLPG的有效方法。我们的结果表明,通过适当的正则化使用L1损失优于任何使用常规L2损失构建的系统,并且不需要应用MLPG(否则是必需的)。将MLPG嵌入声学模型时,我们没有观察到感知上的改进。最后,我们表明方差恢复方法对于倒谱特征很重要,但对于F0的预测仅产生较小的感知增益。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号