首页> 外文会议>ACM/IEEE Design Automation Conference >SparseTrain: Exploiting Dataflow Sparsity for Efficient Convolutional Neural Networks Training
【24h】

SparseTrain: Exploiting Dataflow Sparsity for Efficient Convolutional Neural Networks Training

机译:SparseTrain:有效的卷积神经网络训练利用数据流稀疏性

获取原文

摘要

Training Convolutional Neural Networks (CNNs) usually requires a large number of computational resources. In this paper, SparseTrain is proposed to accelerate CNN training by fully exploiting the sparsity. It mainly involves three levels of innovations: activation gradients pruning algorithm, sparse training dataflow, and accelerator architecture. By applying a stochastic pruning algorithm on each layer, the sparsity of back-propagation gradients can be increased dramatically without degrading training accuracy and convergence rate. Moreover, to utilize both natural sparsity (resulted from ReLU or Pooling layers) and artificial sparsity (brought by pruning algorithm), a sparse-aware architecture is proposed for training acceleration. This architecture supports forward and back-propagation of CNN by adopting 1-Dimensional convolution dataflow. We have built a cycle-accurate architecture simulator to evaluate the performance and efficiency based on the synthesized design with 14nm FinFET technologies. Evaluation results on AlexNet/ResNet show that SparseTrain could achieve about 2.7× speedup and 2.2× energy efficiency improvement on average compared with the original training process.
机译:训练卷积神经网络(CNN)通常需要大量的计算资源。在本文中,提出了SparseTrain,通过充分利用稀疏性来加速CNN训练。它主要涉及三个创新级别:激活梯度修剪算法,稀疏训练数据流和加速器体系结构。通过在每一层上应用随机修剪算法,可以在不降低训练精度和收敛速度的情况下,大幅提高反向传播梯度的稀疏性。此外,为了利用自然稀疏性(来自ReLU或Pooling层)和人工稀疏性(通过修剪算法带来),提出了一种稀疏感知架构来训练加速。该架构通过采用一维卷积数据流来支持CNN的正向和反向传播。我们构建了一个周期精确的架构模拟器,以基于14nm FinFET技术的综合设计评估性能和效率。 AlexNet / ResNet上的评估结果表明,与原始训练过程相比,SparseTrain可以平均实现2.7倍的加速和2.2倍的能源效率提高。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号