首页> 外文会议>IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops >Do All MobileNets Quantize Poorly? Gaining Insights into the Effect of Quantization on Depthwise Separable Convolutional Networks Through the Eyes of Multi-scale Distributional Dynamics
【24h】

Do All MobileNets Quantize Poorly? Gaining Insights into the Effect of Quantization on Depthwise Separable Convolutional Networks Through the Eyes of Multi-scale Distributional Dynamics

机译:所有移动器都能量化不佳吗? 通过多尺度分布动态的眼睛,获得量化对深度可分离卷积网络的效果的洞察

获取原文

摘要

As the “Mobile AI” revolution continues to grow, so does the need to understand the behaviour of edge-deployed deep neural networks. In particular, MobileNets [9], [22] are the go-to family of deep convolutional neural networks (CNN) for mobile. However, they often have significant accuracy degradation under post-training quantization. While studies have introduced quantization-aware training and other methods to tackle this challenge, there is limited understanding into why MobileNets (and potentially depthwise-separable CNNs (DWSCNN) in general) quantize so poorly compared to other CNN architectures. Motivated to gain deeper insights into this phenomenon, we take a different strategy and study the multi-scale distributional dynamics of MobileNet-V1, a set of smaller DWSCNNs, and regular CNNs. Specifically, we investigate the impact of quantization on the weight and activation distributional dynamics as information propagates from layer to layer, as well as overall changes in distributional dynamics at the network level. This fine-grained analysis revealed significant dynamic range fluctuations and a “distributional mismatch” between channelwise and layerwise distributions in DWSCNNs that lead to increasing quantized degradation and distributional shift during information propagation. Furthermore, analysis of the activation quantization errors show that there is greater quantization error accumulation in DWSCNN compared to regular CNNs. The hope is that such insights can lead to innovative strategies for reducing such distributional dynamics changes and improve post-training quantization for mobile.
机译:随着“移动AI”革命继续增长,需要了解边缘部署的深神经网络的行为。特别地,Mobilenets [9],[22]是移动的深卷积神经网络(CNN)的转向家庭。然而,它们通常在训练后量化下具有显着的准确性降解。虽然研究引入了量化感知的培训和其他方法来解决这一挑战,但有限地了解为什么移动电动机(以及潜在深度可分离的CNNS(DWSCNN)一般)与其他CNN架构相比如此差。有动力提高对这种现象的深刻洞察力,我们采取了不同的策略,研究MobileNet-V1,一组较小的DWSCNN和常规CNN的多尺度分配动态。具体地,我们调查量化对重量和激活分布动态的影响,因为信息从层到层传播,以及网络级别的分布动态的总体变化。该细粒度分析显示了在DWSCNN中的通道和层状分布之间的显着动态范围波动和“分布不匹配”,其导致信息传播期间增加量化的劣化和分布偏移。此外,对激活量化误差的分析表明,与常规CNN相比,DWSCNN中存在更大的量化误差累积。希望如此,这种见解可能导致减少这种分布动力学的创新策略,并改善移动后训练后量化。

著录项

相似文献

  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号