首页> 外文会议>Workhshop on NLP for Conversational AI >Accelerating Natural Language Understanding in Task-Oriented Dialog
【24h】

Accelerating Natural Language Understanding in Task-Oriented Dialog

机译:在面向任务的对话中加速自然语言理解

获取原文

摘要

Task-oriented dialog models typically leverage complex neural architectures and large-scale, pre-trained Transformers to achieve state-of-the-art performance on popular natural language understanding benchmarks. However, these models frequently have in excess of tens of millions of parameters, making them impossible to deploy on-device where resource-efficiency is a major concern. In this work, we show that a simple convolutional model compressed with structured pruning achieves largely comparable results to BERT (Devlin et al., 2019) on ATIS and Snips, with under 100K parameters. Moreover, we perform acceleration experiments on CPUs, where we observe our multi-task model predicts intents and slots nearly 63 x faster than even DistilBERT (Sanh et al., 2019).
机译:面向任务导向的对话模型通常利用复杂的神经架构和大规模预先训练的变压器,以实现对流行的自然语言理解基准的最先进的性能。然而,这些模型经常超过数百万个参数,使得它们无法部署在设备上,其中资源效率是一个主要问题。在这项工作中,我们表明,用结构化修剪压缩的简单卷积模型实现了伯特(Devlin等,2019)在ATIS和Snips上大量比较的结果,并具有100k参数。此外,我们对CPU进行加速实验,在那里观察我们的多任务模型预测意图,比甚至甚至甚至甚至甚至更快的插槽(Sanh等,2019)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号