首页> 外文会议>IEEE International Conference on Information and Automation >Distilling deep neural networks with reinforcement learning
【24h】

Distilling deep neural networks with reinforcement learning

机译:通过强化学习提炼深度神经网络

获取原文

摘要

Deep architecture can improve performance of neural networks whereas it increases the computational complexity. Compressing networks is the key to solve this problem. The framework Knowledge Distilling (KD) compresses cumbersome networks well. It improved mimic learning, enabling knowledge to be transferred from cumbersome networks to compressed networks without constraint of architectures. Inspired by AlphaGo Zero, this paper proposed an algorithm combining KD with reinforcement learning to compress networks on changing datasets. In this algorithm, the compressed networks interact with the environment made by KD to produce datasets that are appropriate w.r.t the model. Monte Carlo Tree Search (MCTS) of AlphaGo Zero is used to produce the datasets by making a trade-off between the prediction of compressed networks and the knowledge. In experiments, the algorithm proved to be effective in compressing networks by training ResNet on CIFAR datasets, with mean squared error as the object function.
机译:深度建筑可以提高神经网络的性能,而它会增加计算复杂性。压缩网络是解决此问题的关键。框架知识蒸馏(KD)很好地压缩了繁琐的网络。它改进了模仿学习,使知识能够从繁琐的网络转移到压缩网络,而不会限制架构。这篇论文提出了一种将KD与加强学习的算法组合以压缩在改变数据集中的网络。在该算法中,压缩网络与KD制造的环境交互以产生适当的数据集W.R.T模型。 alphago零的蒙特卡罗树搜索(MCTS)用于通过在压缩网络的预测和知识预测之间进行权衡来生产数据集。在实验中,该算法证明通过在CIFAR数据集上训练Reset来对网络进行有效,具有平均平方误差作为对象功能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号