【24h】

Doing Good or Doing Right? Exploring the Weakness of Commonsense Causal Reasoning Models

机译:做得好或做对的? 探索致辞因果推理模型的弱点

获取原文

摘要

Pretrained language models (PLM) achieve surprising performance on the Choice of Plausible Alternatives (COPA) task. However, whether PLMs have truly acquired the ability of causal reasoning remains a question. In this paper, we investigate the problem of semantic similarity bias and reveal the vulnerability of current COPA models by certain attacks. Previous solutions that tackle the superficial cues of unbalanced token distribution still encounter the same problem of semantic bias, even more seriously due to the utilization of more training data. We mitigate this problem by simply adding a regularization loss and experimental results show that this solution not only improves the model's generalization ability, but also assists the models to perform more robustly on a challenging dataset, BCOPA-CE. which has unbiased token distribution and is more difficult for models to distinguish cause and effect.
机译:预先训练的语言模型(PLM)在选择合理的替代品(COPA)任务时达到令人惊讶的表现。 但是,PLMS是否真正获得了因果推理能力仍然是一个问题。 在本文中,我们调查了语义相似性偏差的问题,并通过某些攻击揭示了当前COPA模型的脆弱性。 以前解决了不平衡令牌分发的浅表线索的解决方案仍然遇到了语义偏见的相同问题,甚至更严重是由于利用更多培训数据。 我们通过简单地添加正规化损失和实验结果表明,这种解决方案不仅提高了模型的泛化能力,而且还可以帮助模型在挑战数据集,BCOPA-CE上更加强大地执行模型。 这具有无偏见的令牌分布,对于模型来说更难以区分原因和效果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号