首页> 外文期刊>Computer speech and language >Attention-based BiLSTM fused CNN with gating mechanism model for Chinese long text classification
【24h】

Attention-based BiLSTM fused CNN with gating mechanism model for Chinese long text classification

机译:基于注意力的Bilstm融合CNN与中国长文本分类的门控机制模型

获取原文
获取原文并翻译 | 示例
           

摘要

Neural networks have been widely used in the field of text classification, and have achieved good results on various Chinese datasets. However, for long text classification, there are a lot of redundant information in text data, and some of the redundant information may involve other topic information, which makes long text classification challenging. To solve the above problems, this paper proposes a new text classification model, called attention-based BiLSTM fused CNN with gating mechanism(ABLG-CNN). In ABLG-CNN, word2vec is used to train word vector representation. The attention mechanism is used to calculate context vector of words to derive keyword information. Bidirectional long short-term memory network (BiLSTM) captures context features. Based on this, convolutional neural network(CNN) captures topic salient features. In view of the possible existence of sentences involving other topic information in long texts, a gating mechanism is introduced to assign weights to BiLSTM and CNN output features to obtain text fusion features that are favorable for classification. ABLG-CNN can capture text context semantics and local phrase features, and perform experimental verification on two long text news datasets. The experimental results show that ABLG-CNN's classification performance is better than other latest text classification methods.
机译:神经网络已广泛用于文本分类领域,并在各种中文数据集中实现了良好的结果。然而,对于长篇文本分类,文本数据中存在许多冗余信息,并且一些冗余信息可以涉及其他主题信息,这使得长文本分类具有挑战性。为了解决上述问题,本文提出了一种新的文本分类模型,称为基于关注BiLSTM融合CNN与门控机制(ABLG-CNN)。在ABLG-CNN中,Word2VEC用于培训Word矢量表示。注意机制用于计算单词的上下文向量以派生关键字信息。双向长期内存网络(BILSTM)捕获上下文功能。基于此,卷积神经网络(CNN)捕获主题突出特征。鉴于在长文本中存在涉及其他主题信息的可能存在的句子,引入了一个门控机制以将权重与BILSTM和CNN输出特征分配以获得有利于分类的文本融合功能。 ABLG-CNN可以捕获文本上下文语义和本地短语功能,并在两个长文本新闻数据集上执行实验验证。实验结果表明,ABLG-CNN的分类性能优于其他最新文本分类方法。

著录项

  • 来源
    《Computer speech and language》 |2021年第7期|101182.1-101182.12|共12页
  • 作者单位

    School of Automation Guangdong University of Technology Guangzhou 510006 China;

    School of Automation Guangdong University of Technology Guangzhou 510006 China School of Computer Guangdong University of Technology Guangzhou 510006 China;

    School of Computer Guangdong University of Technology Guangzhou 510006 China;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Attention mechanism; BiLSTM; CNN; Gating mechanism; Text classification;

    机译:注意机制;Bilstm;CNN;门控机制;文本分类;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号