首页> 外国专利> MULTI-TASK TRAINING ARCHITECTURE AND STRATEGY FOR ATTENTION-BASED SPEECH RECOGNITION SYSTEM

MULTI-TASK TRAINING ARCHITECTURE AND STRATEGY FOR ATTENTION-BASED SPEECH RECOGNITION SYSTEM

机译:基于注意的语音识别系统的多任务训练体系结构和策略

摘要

Methods and apparatuses are provided for performing sequence to sequence (Seq2Seq) speech recognition training performed by at least one processor. The method includes acquiring a training set comprising a plurality of pairs of input data and target data corresponding to the input data, encoding the input data into a sequence of hidden states, performing a connectionist temporal classification (CTC) model training based on the sequence of hidden states, performing an attention model training based on the sequence of hidden states, and decoding the sequence of hidden states to generate target labels by independently performing the CTC model training and the attention model training.
机译:提供了用于执行由至少一个处理器执行的序列到序列(Seq2Seq)语音识别训练的方法和装置。该方法包括获取训练集,该训练集包括多对输入数据和与该输入数据相对应的目标数据,将输入数据编码为一系列隐藏状态,基于该序列执行连接主义时间分类(CTC)模型训练。隐藏状态,基于隐藏状态序列进行注意力模型训练,并通过独立执行CTC模型训练和注意模型训练对隐藏状态序列进行解码,生成目标标签。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号