首页> 外文会议>Annual Meeting of the Association for Computational Linguistics >Learning to Faithfully Rationalize by Construction
【24h】

Learning to Faithfully Rationalize by Construction

机译:学会通过构建忠实地合理化

获取原文

摘要

In many settings it is important for one to be able to understand why a model made a particular prediction. In NLP this often entails extracting snippets of an input text 'responsible for' corresponding model output; when such a snippet comprises tokens that indeed informed the model's prediction, it is a faithful explanation. In some settings, faithfulness may be critical to ensure transparency. Lei et al. (2016) proposed a model to produce faithful rationales for neural text classification by defining independent snippet extraction and prediction modules. However, the discrete selection over input tokens performed by this method complicates training, leading to high variance and requiring careful hyperparameter tuning. We propose a simpler variant of this approach that provides faithful explanations by construction. In our scheme, named FRESH, arbitrary feature importance scores (e.g., gradients from a trained model) are used to induce binary labels over token inputs, which an extractor can be trained to predict. An independent classifier module is then trained exclusively on snippets provided by the extractor; these snippets thus constitute faithful explanations, even if the classifier is arbitrarily complex. In both automatic and manual evaluations we find that variants of this simple framework yield predictive performance superior to 'end-to-end' approaches, while being more general and easier to train.
机译:在许多情况下,理解模型为何做出特定预测是很重要的。在NLP中,这通常需要提取“负责”相应模型输出的输入文本片段;当这样一个片段包含确实通知了模型预测的标记时,它是一个忠实的解释。在某些情况下,忠诚可能是确保透明度的关键。Lei等人(2016)提出了一个模型,通过定义独立的片段提取和预测模块,为神经文本分类提供可靠的理论依据。然而,这种方法对输入标记进行的离散选择使训练复杂化,导致高方差,需要仔细调整超参数。我们提出了这种方法的一个更简单的变体,它通过构造提供了可靠的解释。在我们的方案中,命名为FRESH的任意特征重要性分数(例如,来自训练模型的梯度)用于在标记输入上诱导二进制标签,提取器可以训练预测。然后,独立的分类器模块专门根据提取器提供的片段进行训练;因此,这些片段构成了忠实的解释,即使分类器任意复杂。在自动和手动评估中,我们发现这种简单框架的变体产生的预测性能优于“端到端”方法,同时更通用、更易于培训。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号