首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >High-Accuracy and Low-Latency Speech Recognition with Two-Head Contextual Layer Trajectory LSTM Model
【24h】

High-Accuracy and Low-Latency Speech Recognition with Two-Head Contextual Layer Trajectory LSTM Model

机译:两头上下文层轨迹LSTM模型的高精度和低延迟语音识别

获取原文

摘要

While the community keeps promoting end-to-end models over conventional hybrid models, which usually are long short-term memory (LSTM) models trained with a cross entropy criterion followed by a sequence discriminative training criterion, we argue that such conventional hybrid models can still be significantly improved. In this paper, we detail our recent efforts to improve conventional hybrid LSTM acoustic models for high-accuracy and low-latency automatic speech recognition. To achieve high accuracy, we use a contextual layer trajectory LSTM (cltLSTM), which decouples the temporal modeling and target classification tasks, and incorporates future context frames to get more information for accurate acoustic modeling. We further improve the training strategy with sequence-level teacher-student learning. To obtain low latency, we design a two-head cltLSTM, in which one head has zero latency and the other head has a small latency, compared to an LSTM. When trained with Microsoft’s 65 thousand hours of anonymized training data and evaluated with test sets with 1.8 million words, the proposed two-head cltLSTM model with the proposed training strategy yields a 28.2% relative WER reduction over the conventional LSTM acoustic model, with a similar perceived latency.
机译:尽管社区一直在推广端到端模型,而不是传统的混合模型,而传统的混合模型通常是使用交叉熵准则和序列判别训练准则训练的长短期记忆(LSTM)模型,但我们认为此类常规混合模型可以仍需明显改善。在本文中,我们详细介绍了我们最近为改进传统的混合LSTM声学模型以实现高精度和低延迟自动语音识别所做的努力。为了实现高精度,我们使用上下文层轨迹LSTM(cltLSTM),它将时间建模与目标分类任务分离开,并结合了将来的上下文框架以获取更多信息,以进行准确的声学建模。我们通过序列级师生学习进一步改善培训策略。为了获得低延迟,我们设计了一个两头cltLSTM,与LSTM相比,其中一个头的延迟为零,而另一个头的延迟则较小。当使用Microsoft的6.5万小时的匿名培训数据进行培训并使用180万个单词的测试集进行评估时,具有建议的培训策略的拟议的两头cltLSTM模型与常规的LSTM声学模型相比,相对WER降低了28.2%感知延迟。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号