首页> 外文会议>IEEE International Workshop on Signal Processing Systems >Adaptive Learning Rate Adjustment with Short-Term Pre-Training in Data-Parallel Deep Learning
【24h】

Adaptive Learning Rate Adjustment with Short-Term Pre-Training in Data-Parallel Deep Learning

机译:数据并行深度学习中具有短期预训练的自适应学习速率调整

获取原文

摘要

This paper introduces a method to adaptively choose a learning rate (LR) with short-term pre-training (STPT). This is useful for quick model prototyping in data-parallel deep learning. For unknown models, it is necessary to tune numerous hyperparameters. The proposed method reduces computational time and increases efficiency in finding an appropriate LR; multiple LRs are evaluated by STPT in data-parallel deep learning. STPT means training only with the beginning iterations in an epoch. When eight LRs are evaluated using eight parallel workers, the proposed method can easily reduce the computational time by 87.5% in comparison with the conventional method. The accuracy is also improved by 4.8% in comparison with the conventional method with a reference LR of 0.1; thus, no deterioration in accuracy is observed. For an unknown model, this method shows a better training curve trend than other cases with fixed LRs.
机译:本文介绍了一种通过短期预训练(STPT)自适应选择学习率(LR)的方法。这对于数据并行深度学习中的快速模型原型很有用。对于未知模型,有必要调整许多超参数。所提出的方法减少了计算时间并提高了寻找合适的LR的效率。 STPT在数据并行深度学习中评估多个LR。 STPT意味着仅在一个时期的开始迭代中进行训练。当使用八个并行工作者评估八个LR时,与传统方法相比,该方法可以轻松地将计算时间减少87.5%。与参考LR为0.1的传统方法相比,精度也提高了4.8%。因此,没有观察到精度的降低。对于未知模型,此方法显示出比其他具有固定LR的情况更好的训练曲线趋势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号