首页> 外文期刊>Neurocomputing >Keywords extraction with deep neural network model
【24h】

Keywords extraction with deep neural network model

机译:深度神经网络模型的关键词提取

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

Keywords can express the main content of an article or a sentence. Keywords extraction is a critical issue in many Natural Language Processing (NLP) applications and can improve the performance of many NLP systems. The traditional methods of keywords extraction are based on machine learning or graph model. The performance of these methods is influenced by the feature selection and the manually defined rules. In recent years, with the emergence of deep learning technology, learning features automatically with the deep learning algorithm can improve the performance of many tasks. In this paper, we propose a deep neural network model for the task of keywords extraction. We make two extensions on the basis of traditional LSTM model. First, to better utilize both the historic and following contextual information of the given target word, we propose a target center-based LSTM model (TC-LSTM), which learns to encode the target word by considering its contextual information. Second, on the basis of TC-LSTM model, we apply the self-attention mechanism, which enables our model has an ability to focus on informative parts of the associated text. In addition, we also introduce a two-stage training method, which takes advantage of large-scale pseudo training data. Experimental results show the advantage of our method, our model beats all the baseline systems all across the board. And also, the two-stage training method is of great significance for improving the effectiveness of the model. (C) 2019 Elsevier B.V. All rights reserved.
机译:关键字可以表达文章或句子的主要内容。关键字提取是许多自然语言处理(NLP)应用程序中的关键问题,它可以提高许多NLP系统的性能。关键字提取的传统方法基于机器学习或图模型。这些方法的性能受功能选择和手动定义的规则影响。近年来,随着深度学习技术的出现,具有深度学习算法的自动学习功能可以提高许多任务的性能。在本文中,我们提出了一种用于关键字提取任务的深度神经网络模型。我们在传统LSTM模型的基础上进行了两个扩展。首先,为了更好地利用给定目标词的历史和后续上下文信息,我们提出了一个基于目标中心的LSTM模型(TC-LSTM),该模型学习通过考虑目标词的上下文信息来对其进行编码。其次,在TC-LSTM模型的基础上,我们应用了自我注意机制,这使我们的模型能够专注于相关文本的信息部分。此外,我们还介绍了一种两阶段训练方法,该方法利用了大规模伪训练数据。实验结果表明了我们方法的优势,我们的模型击败了所有基准系统。而且,两阶段训练方法对于提高模型的有效性也具有重要意义。 (C)2019 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号