首页> 外文会议>International Conference on Machine Learning >Generative Teaching Networks: Accelerating Neural Architecture Search by Learning to Generate Synthetic Training Data
【24h】

Generative Teaching Networks: Accelerating Neural Architecture Search by Learning to Generate Synthetic Training Data

机译:生成教学网络:通过学习生成合成训练数据来加速神经结构搜索

获取原文

摘要

This paper investigates the intriguing question of whether we can create learning algorithms that automatically generate training data, learning environments, and curricula in order to help AI agents rapidly learn. We show that such algorithms are possible via Generative Teaching Networks (GTNs), a general approach that is, in theory, applicable to supervised, unsupervised, and reinforcement learning, although our experiments only focus on the supervised case. GTNs are deep neural networks that generate data and/or training environments that a learner (e.g. a freshly initialized neural network) trains on for a few SGD steps before being tested on a target task. We then differentiate through the entire learning process via meta-gradients to update the GTN parameters to improve performance on the target task. This paper introduces GTNs, discusses their potential, and showcases that they can substantially accelerate learning. We also demonstrate a practical and exciting application of GTNs: accelerating the evaluation of candidate architectures for neural architecture search (NAS). GTN-NAS improves the NAS state of the art, finding higher performing architectures when controlling for the search proposal mechanism. GTN-NAS also is competitive with the overall state of the art approaches, which achieve top performance while using orders of magnitude less computation than typical NAS methods. Speculating forward, GTNs may represent a first step toward the ambitious goal of algorithms that generate their own training data and, in doing so, open a variety of interesting new research questions and directions.
机译:本文研究了我们是否可以创建自动生成培训数据,学习环境和课程的学习算法的有趣问题,以便快速学习AI代理。我们表明,这种算法是通过生成教学网络(GTN)的一般方法,即理论上,适用于监督,无监督和加强学习,尽管我们的实验仅关注受监督案例。 GTN是深度神经网络,用于生成学习者(例如,新初始化的神经网络)列车的数据和/或培训环境,在测试到目标任务上之前几个SGD步骤。然后,我们通过元梯度通过整个学习过程来分辨,以更新GTN参数以提高目标任务的性能。本文介绍了GTN,讨论了他们的潜力,并展示他们可以大大加速学习。我们还证明了GTN的实际和激动人心的应用:加速对神经结构搜索的候选架构的评估(NAS)。 GTN-NAS改进了本领域的NAS状态,在控制搜索提议机制时找到更高的执行架构。 GTN-NAS也具有竞争性的现有技术方法,这在使用比典型的NAS方法的计算令人越来越少的计算时实现了最佳性能。向前推测,GTN可以代表迈向生成自己的培训数据的雄心勃勃的算法的第一步,并在这样做,打开各种有趣的新研究问题和方向。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号