【24h】

Lottery Ticket Hypothesis: Placing the k-orrect Bets

机译:彩票假设:放置正确的赌注

获取原文

摘要

Neural Network pruning has been one of the widely used techniques for reducing parameter count from an over-parametrized network. In the paper titled Deconstructing Lottery Tickets, the authors showed that pruning is a way of training, which we extend to show that pruning a well-trained network at initialization does not exhibit significant gains in accuracy. Stabilizing the Lottery Ticket Hypothesis motivates us to explore pruning after k~(th) epoch. We show that there exists a minimum value of k above which there is insignificant gain in accuracy, and the network enjoys a maximum level of pruning at this value of k while maintaining or increasing the accuracy of the original network. We test our claims on MNIST, CIFAR10, and CIFAR100 with small architectures such as lenet-300-100, Conv-2,4,6, and more extensive networks such as Resnet-20. We then discuss why pruning at initialization does not result in considerable benefits compared to pruning at k.
机译:神经网络修剪是从过度参数化网络中减少参数计数的广泛使用的技术之一。 在符合解构彩票票据的论文中,作者认为修剪是一种培训方式,我们延伸到显示在初始化时修剪训练有素的网络并没有表现出显着的提升。 稳定彩票假设激励我们在K〜(Th)时期之后探索修剪。 我们表明,上述k的最小值是准确性微不足道的增益,并且网络在维持或增加原始网络的准确性时,网络在该值下享有最大修剪水平。 我们在MNIST,CIFAR10和CIFAR100上测试我们的索赔,其中小型架构,如LENET-300-100,CONV-2,4,6和更广泛的网络,如RESET-20。 然后,我们讨论为什么修剪初始化的原因不会导致与k修剪相比的相当大的益处。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号