首页> 外文会议>International Conference on Speech and Computer >Deep Recurrent Neural Networks in Speech Synthesis Using a Continuous Vocoder
【24h】

Deep Recurrent Neural Networks in Speech Synthesis Using a Continuous Vocoder

机译:使用连续声探剂的语音合成中深入复发性神经网络

获取原文

摘要

In our earlier work in statistical parametric speech synthesis, we proposed a vocoder using continuous FO in combination with Maximum Voiced Frequency (MVF), which was successfully used with a feed-forward deep neural network (DNN). The advantage of a continuous vocoder in this scenario is that vocoder parameters are simpler to model than traditional vocoders with discontinuous FO. However, DNNs have a lack of sequence modeling which might degrade the quality of synthesized speech. In order to avoid this problem, we propose the use of sequence-to-sequence modeling with recurrent neural networks (RNNs). In this paper, four neural network architectures (long short-term memory (LSTM), bidirectional LSTM (BLSTM), gated recurrent network (GRU), and standard RNN) are investigated and applied using this continuous vocoder to model FO, MVF, and Mel-Generalized Cepstrum (MGC) for more natural sounding speech synthesis. Experimental results from objective and subjective evaluations have shown that the proposed framework converges faster and gives state-of-the-art speech synthesis performance while outperforming the conventional feed-forward DNN.
机译:在我们前面的统计参数语音合成工作中,我们提出了组合使用连续FO与最大浊音频率(MVF),这是成功地与前馈深层神经网络(DNN)使用的声码器。在这种情况下连续声码器的优点是,声码器的参数是简单的模型比连续FO传统的声码器。然而,DNNs都缺少序列建模这可能会降低合成语音的质量。为了避免这个问题,我们提出用递归神经网络(RNNs)使用序列到序列建模。在本文中,四层神经网络结构(长短期存储器(LSTM),双向LSTM(BLSTM),门控复发性网络(GRU),和标准RNN)进行调查,并使用这个连续声码器FO,MVF建模应用,并梅尔倒谱广义(MGC)更自然的发声语音合成。从客观和主观评价实验结果表明,所提出的框架收敛速度更快,并给出国家的最先进的语音合成的性能,同时表现好于常规的前馈DNN。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号