首页> 外文会议>International Conference on speech and computer >Deep Recurrent Neural Networks in Speech Synthesis Using a Continuous Vocoder
【24h】

Deep Recurrent Neural Networks in Speech Synthesis Using a Continuous Vocoder

机译:使用连续声码器的语音合成中的深度递归神经网络

获取原文

摘要

In our earlier work in statistical parametric speech synthesis, we proposed a vocoder using continuous FO in combination with Maximum Voiced Frequency (MVF), which was successfully used with a feed-forward deep neural network (DNN). The advantage of a continuous vocoder in this scenario is that vocoder parameters are simpler to model than traditional vocoders with discontinuous FO. However, DNNs have a lack of sequence modeling which might degrade the quality of synthesized speech. In order to avoid this problem, we propose the use of sequence-to-sequence modeling with recurrent neural networks (RNNs). In this paper, four neural network architectures (long short-term memory (LSTM), bidirectional LSTM (BLSTM), gated recurrent network (GRU), and standard RNN) are investigated and applied using this continuous vocoder to model FO, MVF, and Mel-Generalized Cepstrum (MGC) for more natural sounding speech synthesis. Experimental results from objective and subjective evaluations have shown that the proposed framework converges faster and gives state-of-the-art speech synthesis performance while outperforming the conventional feed-forward DNN.
机译:在统计参数语音合成的早期工作中,我们提出了一种将连续FO与最大语音频率(MVF)结合使用的声码器,该声码器已成功与前馈深度神经网络(DNN)结合使用。在这种情况下,连续声码器的优势在于,与具有不连续FO的传统声码器相比,声码器参数更易于建模。但是,DNN缺少序列建模,这可能会降低合成语音的质量。为了避免这个问题,我们建议使用带有递归神经网络(RNN)的序列到序列建模。本文研究了四种神经网络架构(长短期记忆(LSTM),双向LSTM(BLSTM),门控递归网络(GRU)和标准RNN),并使用此连续声码器对FO,MVF和梅尔通用倒谱(MGC),实现更自然的语音合成。来自主观和主观评估的实验结果表明,所提出的框架收敛速度更快,并提供了最先进的语音合成性能,同时胜过了传统的前馈DNN。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号