【24h】

DiSAN: Directional Self-Attention Network for RNN/CNN-Free Language Understanding

机译:isan:定向自我关注网络用于RNN /无CNN的语言理解

获取原文

摘要

Recurrent neural nets (RNN) and convolutional neural nets (CNN) are widely used on NLP tasks to capture the long-term and local dependencies, respectively. Attention mechanisms have recently attracted enormous interest due to their highly parallelizable computation, significantly less training time, and flexibility in modeling dependencies. We propose a novel attention mechanism in which the attention between elements from input sequence(s) is directional and multi-dimensional (i.e., feature-wise). A light-weight neural net, "Directional Self-Attention Network (DiSAN)", is then proposed to learn sentence embedding, based solely on the proposed attention without any RNN/CNN structure. DiSAN is only composed of a directional self-attention with temporal order encoded, followed by a multi-dimensional attention that compresses the sequence into a vector representation. Despite its simple form, DiSAN outperforms complicated RNN models on both prediction quality and time efficiency, It achieves the best test accuracy among all sentence encoding methods and improves the most recent best result by 1.02% on the Stanford Natural Language Inference (SNLI) dataset, and shows state-of-the-art test accuracy on the Stanford Sentiment Treebank (SST), Multi-Genre natural language inference (MultiNLI), Sentences Involving Compositional Knowledge (SICK), Customer Review, MPQA, TREC question-type classification and Subjectivity (SUBJ) datasets.
机译:经常性的神经网络(RNN)和卷积神经网(CNN)被广泛用于NLP任务,分别捕获长期和局部依赖性。由于它们的高度平行化计算,显着较低的培训时间和建模依赖性的灵活性,最近引起了巨大的兴趣。我们提出了一种新的注意机制,其中来自输入序列的元件之间的注意是方向和多维(即,特征 - 明智)。然后提出了一种轻质神经网络,“定向自我关注网络(浅滩)”,以仅仅基于没有任何RNN / CNN结构的提出的注意力。浅滩仅由具有编码时间阶的定向自我关注,然后将序列压缩成矢量表示的多维注意力。尽管表单简单,但易于表现出复杂的RNN模型在预测质量和时间效率上,它实现了所有句子编码方法中的最佳测试准确性,并在斯坦福自然语言推理(SNLI)数据集上提高了最新的最佳结果1.02%,并显示斯坦福州情绪树木银行(SST),多类型自然语言推理(Multinli),涉及组成知识(病情),客户评论,MPQA,TREC问题类型分类和主体性的句子上的最先进的测试准确性(subj)数据集。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号