首页> 外文会议>International conference on artificial neural networks >IQNN: Training Quantized Neural Networks with Iterative Optimizations
【24h】

IQNN: Training Quantized Neural Networks with Iterative Optimizations

机译:IQNN:通过迭代优化来训练量化神经网络

获取原文
获取外文期刊封面目录资料

摘要

Quantized Neural Networks (QNNs) use low bitwidth numbers for representing parameters and intermediate results. The lowering of bitwidths saves storage space and allows for exploiting bitwise operar tions to speed up computations. However, QNNs often have lower prediction accuracies than their floating point counterpaxts, due to the extra quantization errors. In this paper, we propose a quantization algorithm that iteratively solves for the optimal scaling factor during every forward pass, which significantly reduces quantization errors. Moreover, we propose a novel initialization method for the iterative quantization, which speeds up convergence and further reduces quantization errors. Overall, our method improves prediction accuracies of QNNs at no extra costs for the inference. Experiments confirm the efficacy of our method in the quantization of AlexNet, GoogLeNet and ResNet. In particular, we are able to train a GoogLeNet having 4-bit weights and activations to reach 11.4% in top-5 single-crop error on ImageNet dataset, outperforming state-of-the-art QNNs. The code will be available online.
机译:量化神经网络(QNN)使用低位宽数字表示参数和中间结果。降低位宽可节省存储空间,并允许利用按位运算来加快计算速度。然而,由于额外的量化误差,QNN的预测精度通常比其浮点数低。在本文中,我们提出了一种量化算法,该算法迭代地求解每次正向传递过程中的最佳缩放比例,从而显着减少了量化误差。此外,我们提出了一种新颖的迭代量化初始化方法,该方法可以加快收敛速度​​并进一步减少量化误差。总体而言,我们的方法可以提高QNN的预测准确性,而无需为推理进行任何额外的操作。实验证实了我们的方法在AlexNet,GoogLeNet和ResNet量化中的有效性。特别是,我们能够训练一个具有4位权重和激活的GoogLeNet,使其在ImageNet数据集的前5个单裁剪错误中达到11.4%,胜过最新的QNN。该代码将在线提供。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号