首页> 外文会议>Annual conference of the International Speech Communication Association;INTERSPEECH 2011 >Speech Synthesis based on Articulatory-Movement HMMs with Voice-source Codebooks
【24h】

Speech Synthesis based on Articulatory-Movement HMMs with Voice-source Codebooks

机译:基于发音运动码本的关节运动HMM的语音合成

获取原文

摘要

Speech synthesis based on one-model of articulatory movement HMMs, that are commonly applied to both speech recognition (SR) and speech synthesis (SS), is described. In an SS module, speaker-invariant HMMs are applied to generate an articulatory feature (AF) sequence, and then, after converting AFs into vocal tract parameters by using a multilayer neural network (MLN), a speech signal is synthesized through an LSP digital filter. The CELP coding technique is applied to improve voice-sources when generating these sources from embedded codes in the corresponding state of HMMs. The proposed SS module separates phonetic information and the individuality of a speaker. Therefore, the targeted speaker's voice can be synthesized with a small amount of speech data. In the experiments, we carried out listening tests for ten subjects and evaluated both of sound quality and individuality of synthesized speech. As a result, we confirmed that the proposed SS module could produce good quality speech of the targeted speaker even when the training was done with the data set of two-sentences.
机译:描述了基于关节运动HMM的一种模型的语音合成,该模型通常应用于语音识别(SR)和语音合成(SS)。在SS模块中,应用说话者不变的HMM生成发音特征(AF)序列,然后,使用多层神经网络(MLN)将AF转换为声道参数后,通过LSP数字信号合成语音信号筛选。当从HMM对应状态下的嵌入代码生成语音源时,将CELP编码技术应用于改善语音源。提议的SS模块将语音信息和说话者的个性分开。因此,目标讲话者的语音可以与少量语音数据合成。在实验中,我们对十个主题进行了听力测试,并评估了声音质量和合成语音的个性。结果,我们确认,即使使用两个句子的数据集进行训练,所提出的SS模块也可以产生目标讲话者的高质量语音。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号