【24h】

Chinese Triple Extraction Based on BERT Model

机译:基于BERT模型的中国三倍提取

获取原文

摘要

Information extraction (IE) plays a crucial role in natural language processing, which extracts structured facts like entities, attributes, relations and events from unstructured text. The results of information extraction can be applied in many fields including information retrieval, intelligent QA system, to name a few. We define a pair of entities and their relation from a sentence as a triple. Different from most relation extraction tasks, which only extract one relation from a sentence of known entities, we achieved that extracting both relation and entities(a triple, as defined above), from a plain sentence. Until now, there are so many methods proposed to solve information extraction problem and deep learning has made great progress last several years. Among the field of deep learning, the pre-trained model BERT has achieved greatly successful results in a lot of NLP tasks. So we divide our triple extraction task into two sub-tasks, relation classification and entity tagging, and design two models based on BERT for these two sub-tasks, including a CNN-BERT and a Simple BERT. We experimented our models on DuIE Chinese dataset and achieved excellent results.
机译:信息提取(IE)在自然语言处理中发挥着至关重要的作用,其从非结构化文本中提取了实体,属性,关系和事件等结构化事实。信息提取的结果可以应用于许多领域,包括信息检索,智能QA系统,命名几个。我们将一对实体和他们的关系定义为三倍。与大多数关系的提取任务不同,只能从已知实体的句子中提取一个关系,我们实现了从普通句中提取关系和实体(如上所定义的三倍)。到目前为止,有很多方法建议解决信息提取问题,深度学习在过去几年中取得了很大进展。在深度学习领域中,预先训练的模型BERT在很多NLP任务中取得了极大的成功结果。因此,我们将三重提取任务划分为两个子任务,关系分类和实体标记,并根据这两个子任务的伯爵设计了两种模型,包括CNN-BERT和简单的伯特。我们在Duie中文数据集上尝试了我们的模型,并取得了优异的成果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号