首页> 外文会议>2012 IEEE Workshop on Spoken Language Technology. >Unsupervised cross-lingual knowledge transfer in DNN-based LVCSR
【24h】

Unsupervised cross-lingual knowledge transfer in DNN-based LVCSR

机译:基于DNN的LVCSR中无监督的跨语言知识转移

获取原文
获取原文并翻译 | 示例

摘要

We investigate the use of cross-lingual acoustic data to initialise deep neural network (DNN) acoustic models by means of unsupervised restricted Boltzmann machine (RBM) pre-training. DNNs for German are pretrained using one or all of German, Portuguese, Spanish and Swedish. The DNNs are used in a tandem configuration, where the network outputs are used as features for a hidden Markov model (HMM) whose emission densities are modeled by Gaussian mixture models (GMMs), as well as in a hybrid configuration, where the network outputs are used as the HMM state likelihoods. The experiments show that unsupervised pretraining is more crucial for the hybrid setups, particularly with limited amounts of transcribed training data. More importantly, unsupervised pretraining is shown to be language-independent.
机译:我们调查使用跨语言声学数据通过无监督的受限玻尔兹曼机(RBM)预训练来初始化深度神经网络(DNN)声学模型。使用德语,葡萄牙语,西班牙语和瑞典语中的一种或全部训练德语的DNN。 DNN用于串联配置,其中网络输出用作隐马尔可夫模型(HMM)的特征,隐马尔可夫模型的发射密度由高斯混合模型(GMM)建模,也用于混合配置中,其中网络输出被用作HMM状态可能性。实验表明,无监督的预训练对于混合设置更为关键,尤其是在转录的训练数据数量有限的情况下。更重要的是,无监督的预训练被证明与语言无关。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号