首页> 外文会议>Annual meeting of the Association for Computational Linguistics >End-to-end Deep Reinforcement Learning Based Coreference Resolution
【24h】

End-to-end Deep Reinforcement Learning Based Coreference Resolution

机译:基于端到端深度强化学习的共指解析

获取原文

摘要

Recent neural network models have significantly advanced the task of coreference resolution. However, current neural coreference models are typically trained with heuristic loss functions that are computed over a sequence of local decisions. In this paper, we introduce an end-to-end reinforcement learning based coreference resolution model to directly optimize coreference evaluation metrics. Specifically, we modify the state-of-the-art higher-order mention ranking approach in Lee et al. (2018) to a reinforced policy gradient model by incorporating the reward associated with a sequence of coreference linking actions. Furthermore, we introduce maximum entropy reg-ularization for adequate exploration to prevent the model from prematurely converging to a bad local optimum. Our proposed model achieves new state-of-the-art performance on the English OntoNotes v5.0 benchmark.
机译:最近的神经网络模型已经大大提高了共指解析的任务。但是,当前的神经共指模型通常使用启发式损失函数进行训练,这些函数是根据一系列局部决策计算得出的。在本文中,我们介绍了一种基于端到端强化学习的共指分解模型,以直接优化共指评估指标。具体来说,我们在Lee等人中修改了最新的高阶提及排名方法。 (2018)通过结合与一系列共指链接动作相关的奖励,形成了强化的政策梯度模型。此外,我们引入最大熵正则化以进行充分探索,以防止模型过早收敛到不良的局部最优值。我们提出的模型在英语OntoNotes v5.0基准上实现了最新的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号