【24h】

Rule induction for global explanation of trained models

机译:规则归纳,对训练模型进行全局解释

获取原文
获取原文并翻译 | 示例

摘要

Understanding the behavior of a trained network and finding explanations for its outputs is important for improving the network's performance and generalization ability, and for ensuring trust in automated systems. Several approaches have previously been proposed to identify and visualize the most important features by analyzing a trained network. However, the relations between different features and classes are lost in most cases. We propose a technique to induce sets of if-then-else rules that capture these relations to globally explain the predictions of a network. We first calculate the importance of the features in the trained network. We then weigh the original inputs with these feature importance scores, simplify the transformed input space, and finally fit a rule induction model to explain the model predictions. We find that the output rule-sets can explain the predictions of a neural network trained for 4-class text classification from the 20 newsgroups dataset to a macro-averaged F-score of 0.80.
机译:了解训练有素的网络的行为并找到其输出的解释对于提高网络的性能和泛化能力以及确保对自动化系统的信任至关重要。先前已经提出了几种方法来通过分析训练有素的网络来识别和可视化最重要的功能。但是,大多数情况下会丢失不同功能和类之间的关系。我们提出一种技术,以诱导捕获这些关系的if-then-else规则集,以全局性地解释网络的预测。我们首先计算训练网络中功能的重要性。然后,我们使用这些特征重要性得分权衡原始输入,简化转换后的输入空间,最后拟合规则归纳模型来解释模型预测。我们发现输出规则集可以解释为4类文本分类训练的神经网络的预测,该神经网络从20个新闻组数据集到0.80的宏平均F评分。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号