...
首页> 外文期刊>Neurocomputing >Bidirectional LSTM with attention mechanism and convolutional layer for text classification
【24h】

Bidirectional LSTM with attention mechanism and convolutional layer for text classification

机译:Bidirectional LSTM与文本分类的注意机制和卷积层

获取原文
获取原文并翻译 | 示例

摘要

Neural network models have been widely used in the field of natural language processing (NLP). Recurrent neural networks (RNNs), which have the ability to process sequences of arbitrary length, are common methods for sequence modeling tasks. Long short-term memory (LSTM) is one kind of RNNs and has achieved remarkable performance in text classification. However, due to the high dimensionality and sparsity of text data, and to the complex semantics of the natural language, text classification presents difficult challenges. In order to solve the above problems, a novel and unified architecture which contains a bidirectional LSTM (BiLSTM), attention mechanism and the convolutional layer is proposed in this paper. The proposed architecture is called attention-based bidirectional long short-term memory with convolution layer (AC-BiLSTM). In AC-BiLSTM, the convolutional layer extracts the higher-level phrase representations from the word embedding vectors and BiLSTM is used to access both the preceding and succeeding context representations. Attention mechanism is employed to give different focus to the information out-putted from the hidden layers of BiLSTM. Finally, the softmax classifier is used to classify the processed context information. AC-BiLSTM is able to capture both the local feature of phrases as well as global sentence semantics. Experimental verifications are conducted on six sentiment classification datasets and a question classification dataset, including detailed analysis for AC-BiLSTM. The results clearly show that AC-BiLSTM outperforms other state-of-the-art text classification methods in terms of the classification accuracy. (C) 2019 Elsevier B.V. All rights reserved.
机译:神经网络模型已广泛应用于自然语言处理领域(NLP)。具有处理任意长度序列的经常性神经网络(RNN)是序列建模任务的常用方法。长短期内存(LSTM)是一种RNN,在文本分类中取得了显着性能。然而,由于文本数据的高度和稀疏性,以及自然语言的复杂语义,文本分类呈现出困难的挑战。为了解决上述问题,本文提出了一种包含双向LSTM(BILSTM),注意机构和卷积层的新颖和统一的架构。该建议的架构被称为基于关注的双向短期内记忆,卷积层(AC-BILSTM)。在AC-BILSTM中,卷积层提取来自嵌入向量的更高级别短语表示,并且BILSTM用于访问前面和后续上下文表示。注意机制被用来向从Bilstm的隐藏层提供不同的重点。最后,Softmax分类器用于对处理后的上下文信息进行分类。 AC-Bilstm能够捕获短语的本地特征以及全局句子语义。实验验证在六种情绪分类数据集和问题分类数据集中进行,包括对AC-Bilstm的详细分析。结果清楚地表明,在分类准确性方面,AC-Bilstm优于其他最先进的文本分类方法。 (c)2019 Elsevier B.v.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号