首页> 外文会议>IEEE Workshop on Spoken Language Technology >Unsupervised cross-lingual knowledge transfer in DNN-based LVCSR
【24h】

Unsupervised cross-lingual knowledge transfer in DNN-based LVCSR

机译:无监督的基于DNN的LVCSR的交叉语言知识转移

获取原文

摘要

We investigate the use of cross-lingual acoustic data to initialise deep neural network (DNN) acoustic models by means of unsupervised restricted Boltzmann machine (RBM) pre-training. DNNs for German are pretrained using one or all of German, Portuguese, Spanish and Swedish. The DNNs are used in a tandem configuration, where the network outputs are used as features for a hidden Markov model (HMM) whose emission densities are modeled by Gaussian mixture models (GMMs), as well as in a hybrid configuration, where the network outputs are used as the HMM state likelihoods. The experiments show that unsupervised pretraining is more crucial for the hybrid setups, particularly with limited amounts of transcribed training data. More importantly, unsupervised pretraining is shown to be language-independent.
机译:我们调查使用交叉语言声学数据通过无监督的限制Boltzmann机(RBM)预训练来初始化深神经网络(DNN)声学模型。 德国人的DNN是使用德语,葡萄牙语,西班牙语和瑞典的一个或全部掠夺。 DNN用于串联配置,其中网络输出用作隐藏的马尔可夫模型(HMM)的特征,其发射密度由高斯混合模型(GMMS)以及网络输出的混合配置。 被用作嗯状态可能性。 实验表明,无监督的预测对混合机构来说更关心,特别是有限的转录训练数据。 更重要的是,无监督的预先预测被认为是独立的语言。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号