首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Curriculum Learning by Transfer Learning: Theory and Experiments with Deep Networks
【24h】

Curriculum Learning by Transfer Learning: Theory and Experiments with Deep Networks

机译:通过转移学习进行课程学习:深度网络的理论和实验

获取原文
           

摘要

We provide theoretical investigation of curriculum learning in the context of stochastic gradient descent when optimizing the convex linear regression loss. We prove that the rate of convergence of an ideal curriculum learning method is monotonically increasing with the difficulty of the examples. Moreover, among all equally difficult points, convergence is faster when using points which incur higher loss with respect to the current hypothesis. We then analyze curriculum learning in the context of training a CNN. We describe a method which infers the curriculum by way of transfer learning from another network, pre-trained on a different task. While this approach can only approximate the ideal curriculum, we observe empirically similar behavior to the one predicted by the theory, namely, a significant boost in convergence speed at the beginning of training. When the task is made more difficult, improvement in generalization performance is also observed. Finally, curriculum learning exhibits robustness against unfavorable conditions such as excessive regularization.
机译:当优化凸线性回归损失时,我们提供了在随机梯度下降的情况下课程学习的理论研究。我们证明理想的课程学习方法的收敛速度随着示例的难度而单调增加。此外,在所有同样困难的点中,使用相对于当前假设而言会导致更高损失的点时,收敛会更快。然后,我们在训练CNN的背景下分析课程学习。我们描述了一种方法,该方法通过从另一个网络进行转移学习来推断课程,该网络已预先训练有别的任务。虽然这种方法只能近似理想的课程,但我们在经验上观察到的行为与该理论所预测的行为相似,即在训练开始时收敛速度显着提高。当任务变得更加困难时,还可以观察到泛化性能的提高。最后,课程学习表现出针对不利条件(例如过度正规化)的鲁棒性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号