首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >End-to-End Spoken Language Understanding Using Transformer Networks and Self-Supervised Pre-Trained Features
【24h】

End-to-End Spoken Language Understanding Using Transformer Networks and Self-Supervised Pre-Trained Features

机译:使用变压器网络和自我监督的预训练功能的端到端口语理解

获取原文

摘要

Transformer networks and self-supervised pre-training have consistently delivered state-of-art results in the field of natural language processing (NLP); however, their merits in the field of spoken language understanding (SLU) still need further investigation. In this paper we introduce a modular End-to-End (E2E) SLU transformer network based architecture which allows the use of self-supervised pre- trained acoustic features, pre-trained model initialization and multi-task training. Several SLU experiments for predicting intent and entity labels/values using the ATIS dataset are performed. These experiments investigate the interaction of pre-trained model initialization and multi-task training with either traditional filterbank or self-supervised pre-trained acoustic features. Results show not only that self-supervised pre-trained acoustic features outperform filterbank features in almost all the experiments, but also that when these features are used in combination with multi-task training, they almost eliminate the necessity of pre-trained model initialization.
机译:变压器网络和自我监督的预培训在自然语言处理领域(NLP)始终提供最先进的结果;但是,他们在语言理解领域的优点(SLU)仍需要进一步调查。在本文中,我们介绍了一种模块化端到端(E2E)SLU变压器网络的架构,允许使用自我监督的预训练的声学功能,预先训练的模型初始化和多任务培训。执行用于预测使用ATI数据集的意图和实体标签/值的几个SLU实验。这些实验研究了预先训练的模型初始化和多任务培训与传统滤波器或自我监督的预训练的声学特征的相互作用。结果表明,不仅是自我监督的预训练的声学特征几乎所有实验中的溢出扫描器功能差错,而且当这些功能与多任务培训结合使用时,它们几乎消除了预先训练的模型初始化的必要性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号