首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Improved End-To-End Spoken Utterance Classification with a Self-Attention Acoustic Classifier
【24h】

Improved End-To-End Spoken Utterance Classification with a Self-Attention Acoustic Classifier

机译:带有自注意声学分类器的改进的端到端口语话语分类

获取原文

摘要

While human language provides a natural interface for humanmachine communication, there are several challenges concerning extracting the intents of a speaker when interacting with a virtual agent, especially when the speaker is in a noisy acoustic environment, that still remains to be solved. In this paper, we propose a new architecture for end-to-end spoken utterance classification (SUC) and also explore the impact of leveraging lexical information in conjunction with acoustic information obtained from the end-to-end model for SUC. We demonstrate that strong performance can be obtained by the model with acoustic features alone compared to a text classifier on ASR outputs. Furthermore, when acoustic and lexical embeddings from these classifiers are combined, accuracy that is on par with human agents can be achieved.
机译:尽管人类语言为人机通信提供了自然的界面,但在与虚拟代理进行交互时,尤其是在说话者处于嘈杂的声学环境中时,如何提取说话者的意图仍存在一些挑战,这仍然有待解决。在本文中,我们提出了一种用于端到端话语分类(SUC)的新体系结构,并且还探讨了利用词法信息以及从SUC端到端模型获得的声学信息的影响。我们证明,与ASR输出上的文本分类器相比,仅具有声学特征的模型即可获得强大的性能。此外,将来自这些分类器的声音和词汇嵌入进行组合时,可以实现与人工代理相当的准确性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号