...
首页> 外文期刊>JMLR: Workshop and Conference Proceedings >KD-MRI A knowledge distillation framework for image reconstruction and image restoration in MRI workflow
【24h】

KD-MRI A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

机译:KD-MRI在MRI工作流程中的图像重建和图像恢复的知识蒸馏框架

获取原文
           

摘要

Deep learning networks are being developed in every stage of the MRI workflow and have provided state-of-the-art results. However, this has come at the cost of increased computation requirement and storage. Hence, replacing the networks with compact models at various stages in the MRI workflow can significantly reduce the required storage space and provide considerable speedup. In computer vision, knowledge distillation is a commonly used method for model compression. In our work, we propose a knowledge distillation (KD) framework for the image to image problems in the MRI workflow in order to develop compact, low-parameter models without a significant drop in performance. We propose a combination of the attention-based feature distillation method and imitation loss and demonstrate its effectiveness on the popular MRI reconstruction architecture, DC-CNN. We conduct extensive experiments using Cardiac, Brain, and Knee MRI datasets for 4x, 5x and 8x accelerations. We observed that the student network trained with the assistance of the teacher using our proposed KD framework provided significant improvement over the student network trained without assistance across all the datasets and acceleration factors. Specifically, for the Knee dataset, the student network achieves $65%$ parameter reduction, 2x faster CPU running time, and 1.5x faster GPU running time compared to the teacher. Furthermore, we compare our attention-based feature distillation method with other feature distillation methods. We also conduct an ablative study to understand the significance of attention-based distillation and imitation loss. We also extend our KD framework for MRI super-resolution and show encouraging results.
机译:深度学习网络正在MRI工作流程的每个阶段开发,并提供了最先进的结果。但是,这已按增加的计算需求和存储成本。因此,在MRI工作流程中的各个阶段用紧凑型模型替换网络可以显着减少所需的存储空间并提供相当大的加速。在计算机视觉中,知识蒸馏是模型压缩的常用方法。在我们的工作中,我们提出了一种知识蒸馏(KD)框架,用于图像中的图像问题,以便在没有显着下降的情况下开发紧凑的低参数模型。我们提出了基于关注的特征蒸馏方法和模仿损失的组合,并展示其对流行的MRI重建架构DC-CNN的有效性。我们使用心脏,脑和膝关节MRI数据集进行广泛的实验,用于4倍,5倍和8x加速度。我们观察到,使用我们提出的KD框架的教师协助培训的学生网络提供了对学生网络培训的显着改进,而不提供所有数据集和加速因素的帮助。具体而言,对于膝盖数据集,学生网络达到65%的$参数减少,2倍的CPU运行时间,与老师相比,GPU运行时间更快。此外,我们将引人注目的特征蒸馏方法与其他特征蒸馏方法进行比较。我们还开展了一个消融的研究,了解注意力蒸馏和模仿损失的重要性。我们还扩展了MRI超级分辨率的KD框架,并显示了令人鼓舞的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号