首页> 外文会议>Third workshop on structured prediction for NLP 2019 >Lightly-supervised Representation Learning with Global Interpretability
【24h】

Lightly-supervised Representation Learning with Global Interpretability

机译:具有全球可解释性的轻度监督表示学习

获取原文
获取原文并翻译 | 示例

摘要

We propose a lightly-supervised approach for information extraction, in particular named entity classification, which combines the benefits of traditional bootstrapping, i.e., use of limited annotations and interpretability of extraction patterns, with the robust learning approaches proposed in representation learning. Our algorithm iteratively learns custom em-beddings for both the multi-word entities to be extracted and the patterns that match them from a few example entities per category. We demonstrate that this representation-based approach outperforms three other state-of-the-art bootstrapping approaches on two datasets: CoNLL-2003 and OntoNotes. Additionally, using these embeddings, our approach outputs a globally-interpretable model consisting of a decision list, by ranking patterns based on their proximity to the average entity embedding in a given class. We show that this interpretable model performs close to our complete bootstrapping model, proving that representation learning can be used to produce interpretable models with small loss in performance. This decision list can be edited by human experts to mitigate some of that loss and in some cases outperform the original model.
机译:我们提出了一种用于信息提取的轻监督方法,特别是命名实体分类,该方法结合了传统自举的好处(即使用有限的注释和提取模式的可解释性)以及在表示学习中提出的强大的学习方法。我们的算法从每个类别的几个示例实体中迭代学习要提取的多词实体和与之匹配的模式的自定义嵌入。我们证明了这种基于表示的方法在两个数据集:CoNLL-2003和OntoNotes方面优于其他三个最新的自举方法。另外,使用这些嵌入,我们的方法通过基于与给定类中嵌入的平均实体的接近程度对模式进行排序,输出由决策列表组成的全局可解释模型。我们证明了该可解释模型的性能接近于我们完整的自举模型,证明了表示学习可用于产生性能损失很小的可解释模型。该决策列表可以由人类专家编辑,以减轻某些损失,并在某些情况下优于原始模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号