首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >DO as I Mean, Not as I Say: Sequence Loss Training for Spoken Language Understanding
【24h】

DO as I Mean, Not as I Say: Sequence Loss Training for Spoken Language Understanding

机译:我的意思是,不是我所说的:用于口语语言的序列损失培训

获取原文

摘要

Spoken language understanding (SLU) systems extract transcriptions, as well as semantics of intent or named entities from speech, and are essential components of voice activated systems. SLU models, which either directly extract semantics from audio or are composed of pipelined automatic speech recognition (ASR) and natural language understanding (NLU) models, are typically trained via differentiable cross-entropy losses, even when the relevant performance metrics of interest are word or semantic error rates. In this work, we propose non-differentiable sequence losses based on SLU metrics as a proxy for semantic error and use the REINFORCE trick to train ASR and SLU models with this loss. We show that custom sequence loss training is the state-of-the-art on open SLU datasets and leads to 6% relative improvement in both ASR and NLU performance metrics on large proprietary datasets. We also demonstrate how the semantic sequence loss training paradigm can be used to update ASR and SLU models without transcripts, using semantic feedback alone.
机译:口语语言理解(SLU)系统提取转录,以及来自语音的意图或命名实体的语义,是语音激活系统的基本组件。 SLU模型,其直接从音频提取语义或由流水线自动语音识别(ASR)和自然语言理解(NLU)模型,通常通过可差的跨熵损失培训,即使感兴趣的相关性能指标是字或语义错误率。在这项工作中,我们提出了基于SLU指标的非微弱序列损失作为语义误差的代理,并使用这种损失训练ASR和SLU模型的增强技巧。我们表明,定制序列损失培训是开放式SLU数据集的最先进,并在大专有数据集中导致ASR和NLU性能指标的相对改进。我们还展示了语义序列丢失训练范式如何使用单独使用语义反馈来更新无需成绩单的ASR和SLU模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号