首页> 外文会议>IEEE Annual Ubiquitous Computing, Electronics Mobile Communication Conference >Derivative-Free Optimization of Neural Networks using Local Search
【24h】

Derivative-Free Optimization of Neural Networks using Local Search

机译:使用局部搜索的神经网络无导数优化

获取原文

摘要

Deep Neural Networks have received a great deal of attention in the past few years. Applications of Deep Learning broached areas of different domains such as Reinforcement Learning and Computer Vision. Despite their popularity and success, training neural networks can be a challenging process. This paper presents a study on derivative-free, single-candidate optimization of neural networks using Local Search (LS). LS is an algorithm where constrained noise is iteratively applied to subsets of the search space. It is coupled with a Score Decay mechanism to enhance performance. LS is a subsidiary of the Random Search family. Experiments were conducted using a setup that is both suitable for an introduction of the algorithm and representative of modern deep learning tasks, based on the FashionMNIST dataset. Training of a 5-Million parameter CNN was done in several scenarios, including Stochastic Gradient Descent (SGD) coupled with Backpropagation (BP) for comparison. Results reveal that although LS was not competitive in terms of convergence speed, it was actually able to converge to a lower loss than SGD. In addition, LS trained the CNN using Accuracy rather than Loss as a learning signal, though to a lower performance. In conclusion, LS presents a viable alternative in cases where SGD fails or is not suitable. The simplicity of LS can make it attractive to non-experts who would want to try neural nets for the first-time or on novel, non-differentiable tasks.
机译:在过去的几年中,深度神经网络受到了广泛的关注。深度学习的应用涉及不同领域,例如强化学习和计算机视觉。尽管获得了广泛的欢迎和成功,但是训练神经网络仍然是一个具有挑战性的过程。本文提出了使用局部搜索(LS)进行神经网络的无导数,单候选优化的研究。 LS是一种算法,其中将受约束的噪声迭代地应用于搜索空间的子集。它与“得分衰减”机制结合在一起以提高性能。 LS是Random Search家族的子公司。基于FashionMNIST数据集,使用适合于引入算法并代表现代深度学习任务的设置进行了实验。在几种情况下对500万个参数的CNN进行了训练,包括随机梯度下降(SGD)和反向传播(BP)进行比较。结果表明,尽管LS在收敛速度方面没有竞争力,但实际上它能够收敛到比SGD更低的损失。此外,尽管性能较低,但LS还是使用“准确性”而不是“损失”作为学习信号来训练CNN。总之,在SGD失败或不合适的情况下,LS提供了可行的替代方案。 LS的简单性使其对于那些想首次尝试神经网络或在新颖,不可微的任务上尝试神经网络的非专家有吸引力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号