首页> 外文期刊>Knowledge-Based Systems >Enhancing Transformer-based language models with commonsense representations for knowledge-driven machine comprehension
【24h】

Enhancing Transformer-based language models with commonsense representations for knowledge-driven machine comprehension

机译:通过用于知识驱动的机器理解的致致通知表示,增强基于变压器的语言模型

获取原文
获取原文并翻译 | 示例
           

摘要

Compared to the traditional machine reading comprehension (MRC) with limitation to the information in a passage, knowledge-driven MRC tasks aim to enable models to answer the question according to text and related commonsense knowledge. Although pre-trained Transformer-based language models (TrLMs) such as BERT and Roberta, have shown powerful performance in MRC, external knowledge such as unspoken commonsense and world knowledge still cannot be used and explained explicitly. In this work, we present three simple yet effective injection methods integrated into the structure of TrLMs to fine-tune downstream knowledge-driven MRC tasks with off-the-shelf commonsense representations. Moreover, we introduce a mask mechanism for a token-level multi-hop relationship searching to filter external knowledge. We have conducted extensive experiments on DREAM and CosmosQA, two prevalent knowledge-driven datasets. Experimental results indicate that the incremental TrLMs have outperformed the baseline systems by 1%-4.1% with a fewer computational cost. Further analysis shows the effectiveness of the proposed methods and the robustness of the incremental model in the case of an incomplete training set. (c) 2021 Elsevier B.V. All rights reserved.commentSuperscript/Subscript Available/comment
机译:与传统的机器阅读理解(MRC)相比,对段落中的信息的限制,知识驱动的MRC任务旨在根据文本和相关的致辞知识启用模型来回答问题。虽然BERT和ROBERTA等预先接受的基于变压器的语言模型(TRLMS),但在MRC中表现出强大的性能,诸如未说出售的偶然和世界知识等外部知识仍然无法明确地使用和解释。在这项工作中,我们提出了三种简单但有效的注射方法,集成到TRLMS结构中,以微调下游知识驱动的MRC任务与现成的致辞表示。此外,我们向令牌级多跳关系引入掩模机制,搜索过滤外部知识。我们对梦想和COSMOSQA进行了广泛的实验,两个普遍的知识驱动的数据集。实验结果表明,增量TRLMS从基线系统优于1%-4.1%,计算成本较少。进一步的分析显示了在不完全训练集的情况下提出的方法和增量模型的鲁棒性的有效性。 (c)2021 elestvier b.v.保留所有权利。&注释&可用的上标/下标& /评论

著录项

  • 来源
    《Knowledge-Based Systems》 |2021年第23期|106936.1-106936.12|共12页
  • 作者单位

    Northwestern Polytech Univ Sch Comp Sci & Engn Xian 710072 Peoples R China;

    Northwestern Polytech Univ Sch Comp Sci & Engn Xian 710072 Peoples R China;

    Northwestern Polytech Univ Sch Comp Sci & Engn Xian 710072 Peoples R China|Henan Key Lab Big Data Proc & Analyt Elect Commer Luoyang 471000 Peoples R China;

    Northwestern Polytech Univ Sch Comp Sci & Engn Xian 710072 Peoples R China;

    Northwestern Polytech Univ Sch Comp Sci & Engn Xian 710072 Peoples R China;

    London South Bank Univ Sch Engn Div Comp Sci & Informat London England;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Machine Reading Comprehension; Transformer; Commonsense knowledge; Pretrained language model;

    机译:机器阅读理解;变压器;致商知识;普里德兰语模型;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号