首页> 外文会议>Conference on empirical methods in natural language processing >Rule induction for global explanation of trained models
【24h】

Rule induction for global explanation of trained models

机译:训练模型的全球解释规则诱导

获取原文

摘要

Understanding the behavior of a trained network and finding explanations for its outputs is important for improving the network's performance and generalization ability, and for ensuring trust in automated systems. Several approaches have previously been proposed to identify and visualize the most important features by analyzing a trained network. However, the relations between different features and classes are lost in most cases. We propose a technique to induce sets of if-then-else rules that capture these relations to globally explain the predictions of a network. We first calculate the importance of the features in the trained network. We then weigh the original inputs with these feature importance scores, simplify the transformed input space, and finally fit a rule induction model to explain the model predictions. We find that the output rule-sets can explain the predictions of a neural network trained for 4-class text classification from the 20 newsgroups dataset to a macro-averaged F-score of 0.80.
机译:了解训练有素的网络的行为并为其输出找到解释对于提高网络的性能和泛化能力以及确保在自动化系统中的信任来说是重要的。先前已经提出了几种方法来通过分析训练有素的网络来识别和可视化最重要的特征。然而,在大多数情况下,不同特征和类之间的关系丢失了。我们提出了一种技术来诱导捕获这些关系的IF-Oll的规则,以全局解释网络的预测。我们首先计算培训网络中的功能的重要性。然后,我们使用这些特征重要性分数来称量原始输入,简化了变换的输入空间,最后拟合了规则感应模型来解释模型预测。我们发现输出规则集可以解释从20个新闻组数据集到4级文本分类的神经网络的预测到宏平均f分数为0.80。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号