首页> 外文会议>Pacific Rim international conference on artificial intelligence >Towards Meta-learning of Deep Architectures for Efficient Domain Adaptation
【24h】

Towards Meta-learning of Deep Architectures for Efficient Domain Adaptation

机译:面向深度架构的元学习,以实现高效的领域适应

获取原文

摘要

This paper proposes an efficient domain adaption approach using deep learning along with transfer and meta-level learning. The objective is to identify how many blocks (i.e. groups of consecutive layers) of a pre-trained image classification network need to be fine-tuned based on the characteristics of the new task. In order to investigate it, a number of experiments have been conducted using different pre-trained networks and image datasets. The networks were fine-tuned, starting from the blocks containing the output layers and progressively moving towards the input layer, on various tasks with characteristics different from the original task. The amount of fine-tuning of a pre-trained network (i.e. the number of top layers requiring adaptation) is usually dependent on the complexity, size, and domain similarity of the original and new tasks. Considering these characteristics, a question arises of how many blocks of the network need to be fine-tuned to get maximum possible accuracy? Which of a number of available pre-trained networks require fine-tuning of the minimum number of blocks to achieve this accuracy? The experiments, that involve three network architectures each divided into 10 blocks on average and five datasets, empirically confirm the intuition that there exists a relationship between the similarity of the original and new tasks and the depth of network needed to fine-tune in order to achieve accuracy comparable with that of a model trained from scratch. Further analysis shows that the fine-tuning of the final top blocks of the network, which represent the high-level features, is sufficient in most of the cases. Moreover, we have empirically verified that less similar tasks require fine-tuning of deeper portions of the network, which however is still better than training a network from scratch.
机译:本文提出了一种有效的领域自适应方法,该方法将深度学习与转移和元级学习一起使用。目的是基于新任务的特征来确定需要对预训练图像分类网络的多少块(即,连续层的组)进行微调。为了对其进行研究,已经使用不同的预训练网络和图像数据集进行了许多实验。在包含不同于原始任务的各种任务上,对网络进行了微调,从包含输出层的块开始,逐渐向输入层移动。预训练网络的微调数量(即需要调整的顶层数量)通常取决于原始任务和新任务的复杂性,大小和域相似性。考虑到这些特性,就产生了一个问题,即需要微调网络的多少块以获得最大可能的精度?哪些可用的预训练网络中的哪一个需要微调最小块数才能达到此精度?实验涉及三个网络体系结构,每个网络体系结构平均分为10个块和五个数据集,从经验上证实了直觉,即原始任务和新任务的相似性与微调所需的网络深度之间存在关系。达到与从头开始训练的模型可比的精度。进一步的分析表明,在大多数情况下,微调代表高级功能的网络最后几个顶级块就足够了。此外,我们已经通过经验证明,较少的相似任务需要对网络的较深部分进行微调,但是,这仍然比从头开始训练网络要好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号