...
首页> 外文期刊>PLoS One >Compressing deep graph convolution network with multi-staged knowledge distillation
【24h】

Compressing deep graph convolution network with multi-staged knowledge distillation

机译:用多阶段知识蒸馏压缩深图卷积网络

获取原文
           

摘要

Given a trained deep graph convolution network (GCN), how can we effectively compress it into a compact network without significant loss of accuracy? Compressing a trained deep GCN into a compact GCN is of great importance for implementing the model to environments such as mobile or embedded systems, which have limited computing resources. However, previous works for compressing deep GCNs do not consider the multi-hop aggregation of the deep GCNs, though it is the main purpose for their multiple GCN layers. In this work, we propose MustaD ( Mu lti- sta ged knowledge D istillation), a novel approach for compressing deep GCNs to single-layered GCNs through multi-staged knowledge distillation (KD). MustaD distills the knowledge of 1) the aggregation from multiple GCN layers as well as 2) task prediction while preserving the multi-hop feature aggregation of deep GCNs by a single effective layer. Extensive experiments on four real-world datasets show that MustaD provides the state-of-the-art performance compared to other KD based methods. Specifically, MustaD presents up to 4.21%p improvement of accuracy compared to the second-best KD models.
机译:鉴于训练有素的深图卷积网络(GCN),我们如何将它与紧凑型网络压缩到紧凑的网络中,而无需显着损失准确性?将训练的深GCN压缩成Compact GCN,对于将模型实施到诸如移动或嵌入式系统(如计算资源有限的环境)中,这是非常重要的。然而,以前用于压缩深GCN的工作不考虑深度GCN的多跳聚合,尽管它是它们多个GCN层的主要目的。在这项工作中,我们提出了芥末(MU LTI-STA GED知识D Istillation),通过多分阶段知识蒸馏(KD)将深GCNS压缩到单层GCN的新方法。 Mustad蒸馏1)从多个GCN层的聚合以及2)任务预测,同时通过单个有效层保留深GCN的多跳特征聚合。四个现实数据集的广泛实验显示,与基于KD的方法相比,Mustad提供了最先进的性能。具体而言,与第二款最佳KD模型相比,MustAd呈现高达4.21%的准确度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号