首页> 外文会议>Pacific Rim international conference on artificial intelligence >Towards Meta-learning of Deep Architectures for Efficient Domain Adaptation
【24h】

Towards Meta-learning of Deep Architectures for Efficient Domain Adaptation

机译:对高效域适应的深度架构的元学习

获取原文

摘要

This paper proposes an efficient domain adaption approach using deep learning along with transfer and meta-level learning. The objective is to identify how many blocks (i.e. groups of consecutive layers) of a pre-trained image classification network need to be fine-tuned based on the characteristics of the new task. In order to investigate it, a number of experiments have been conducted using different pre-trained networks and image datasets. The networks were fine-tuned, starting from the blocks containing the output layers and progressively moving towards the input layer, on various tasks with characteristics different from the original task. The amount of fine-tuning of a pre-trained network (i.e. the number of top layers requiring adaptation) is usually dependent on the complexity, size, and domain similarity of the original and new tasks. Considering these characteristics, a question arises of how many blocks of the network need to be fine-tuned to get maximum possible accuracy? Which of a number of available pre-trained networks require fine-tuning of the minimum number of blocks to achieve this accuracy? The experiments, that involve three network architectures each divided into 10 blocks on average and five datasets, empirically confirm the intuition that there exists a relationship between the similarity of the original and new tasks and the depth of network needed to fine-tune in order to achieve accuracy comparable with that of a model trained from scratch. Further analysis shows that the fine-tuning of the final top blocks of the network, which represent the high-level features, is sufficient in most of the cases. Moreover, we have empirically verified that less similar tasks require fine-tuning of deeper portions of the network, which however is still better than training a network from scratch.
机译:本文提出了一种利用深度学习以及转移和元级学习的有效域适应方法。目标是根据新任务的特征来识别预先训练的图像分类网络的预先调整的图像分类网络的块(即连续层组)。为了调查它,已经使用不同的预先训练的网络和图像数据集进行了许多实验。从包含输出层的块开始,从包含输出层的块开始,从包含输出层的块开始微调,从具有与原始任务不同的各种任务。预先训练的网络的微调量(即,需要适应的顶层数)通常取决于原始任务和新任务的复杂性,大小和域相似度。考虑到这些特征,问题出现了许多网络需要微调以获得最大可能的准确性?可用预先训练的网络中的哪一个需要微调最小块数以实现这种准确性?涉及三个网络架构的实验,每个网络架构平均分为10个块,并经验证实了原始和新任务的相似性与微调所需的网络的相似性之间存在关系的直觉,以便实现与从头训练培训的模型相当的准确性。进一步的分析表明,在大多数情况下,该网络的最终顶级块的微调是足够的。此外,我们经验经验验证了更有类似的任务需要微调网络的更深部分,然而仍然比从头划伤训练网络更好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号