首页> 外文会议>IEEE Winter Conference on Applications of Computer Vision >Towards Visually Explaining Video Understanding Networks with Perturbation
【24h】

Towards Visually Explaining Video Understanding Networks with Perturbation

机译:在视觉上解释扰动的视频理解网络

获取原文

摘要

"Making black box models explainable " is a vital problem that accompanies the development of deep learning networks. For networks taking visual information as input, one basic but challenging explanation method is to identify and visualize the input pixels/regions that dominate the network’s prediction. However, most existing works focus on explaining networks taking a single image as input and do not consider the temporal relationship that exists in videos. Providing an easy-to-use visual explanation method that is applicable to diversified structures of video understanding networks still remains an open challenge. In this paper, we investigate a generic perturbation-based method for visually explaining video understanding networks. Besides, we propose a novel loss function to enhance the method by constraining the smoothness of its results in both spatial and temporal dimensions. The method enables the comparison of explanation results between different network structures to become possible and can also avoid generating the pathological adversarial explanations for video inputs. Experimental comparison results verified the effectiveness of our method.
机译:“制作黑匣子型号可解释”是一个重要的问题,伴随着深入学习网络的发展。对于将视觉信息视为输入的网络,一个基本但具有挑战性的解释方法是识别和可视化主导网络预测的输入像素/区域。然而,大多数现有的作品侧重于解释拍摄单个图像的网络作为输入,并且不考虑视频中存在的时间关系。提供易于使用的视觉解释方法,适用于视频理解网络的多元化结构仍然是一个开放的挑战。在本文中,我们研究了在视觉上解释视频理解网络的基于通用扰动的方法。此外,我们提出了一种新颖的损失功能来增强该方法,通过约束其在空间和时间尺寸的结果的平滑度。该方法使得可以比较不同网络结构之间的解释结果,并且还可以避免为视频输入产生病理对抗解释。实验比较结果验证了我们方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号