...
首页> 外文期刊>ACM transactions on Asian and low-resource language information processing >SACNN: Self-attentive Convolutional Neural Network Model for Natural Language Inference
【24h】

SACNN: Self-attentive Convolutional Neural Network Model for Natural Language Inference

机译:SACNN:自我临床卷积神经网络模型,用于自然语言推断

获取原文
获取原文并翻译 | 示例
           

摘要

Inference has been central problem for understanding and reasoning in artificial intelligence. Especially, Natural Language Inference is an interesting problem that has attracted the attention of many researchers. Natural language inference intends to predict whether a hypothesis sentence can be inferred from the premise sentence. Most prior works rely on a simplistic association between the premise and hypothesis sentence pairs, which is not sufficient for learning complex relationships between them. The strategy also fails to exploit local context information fully. Long Short Term Memory (LSTM) or gated recurrent units networks (GRU) are not effective in modeling long-term dependencies, and their schemes are far more complex as compared to Convolutional Neural Networks (CNN). To address this problem of long-term dependency, and to involve context for modeling better representation of a sentence, in this article, a general Self-Attentive Convolution Neural Network (SACNN) is presented for natural language inference and sentence pair modeling tasks. The proposed model uses CNNs to integrate mutual interactions between sentences, and each sentence with their counterparts is taken into consideration for the formulation of their representation. Moreover, the selfattention mechanism helps fully exploit the context semantics and long-term dependencies within a sentence. Experimental results proved that SACNN was able to outperform strong baselines and achieved an accuracy of 89.7% on the stanford natural language inference (SNLI) dataset.
机译:推理是人工智能理解和推理的核心问题。特别是,自然语言推论是一种有趣的问题,它引起了许多研究人员的注意。自然语言推断打算预测可以从前提句子中推断出假设判决。大多数事先作品依赖于前提和假设句子对之间的简单关联,这不足以学习它们之间的复杂关系。该策略还会无法充分利用本地上下文信息。长短期内存(LSTM)或门控复发单位网络(GRU)在建模长期依赖性方面无效,与卷积神经网络(CNN)相比,它们的方案更复杂。为了解决长期依赖性的这个问题,并且涉及用于建模更好地表达句子的背景,在本文中,呈现了一种用于自然语言推理和句子对建模任务的一般自我细心卷积神经网络(SACNN)。所提出的模型使用CNN来集成句子之间的相互相互作用,并且考虑到其代表的制定的每个句子。此外,自助活动机制有助于充分利用句子中的上下文语义和长期依赖性。实验结果证明,SACNN能够优于强大的基线,并在斯坦福自然语言推理(SNLI)数据集中实现了89.7%的准确性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号