首页> 外文会议>IEEE Winter Conference on Applications of Computer Vision >GradMix: Multi-source Transfer across Domains and Tasks
【24h】

GradMix: Multi-source Transfer across Domains and Tasks

机译:GradMix:跨域和任务的多源传输

获取原文

摘要

The computer vision community is witnessing an unprecedented rate of new tasks being proposed and addressed, thanks to the deep convolutional networks’ capability to find complex mappings from $mathcal{X}$ to $mathcal{Y}$. The advent of each task often accompanies the release of a large-scale annotated dataset, for supervised training of deep network. However, it is expensive and time-consuming to manually label sufficient amount of training data. Therefore, it is important to develop algorithms that can leverage off-the-shelf labeled dataset to learn useful knowledge for the target task. While previous works mostly focus on transfer learning from a single source, we study multi-source transfer across domains and tasks (MS-DTT), in a semi-supervised setting. We propose GradMix, a model-agnostic method applicable to any model trained with gradient-based learning rule, to transfer knowledge via gradient descent by weighting and mixing the gradients from all sources during training. GradMix follows a meta-learning objective, which assigns layer-wise weights to the source gradients, such that the combined gradient follows the direction that minimize the loss for a small set of samples from the target dataset. In addition, we propose to adaptively adjust the learning rate for each mini-batch based on its importance to the target task, and a pseudo-labeling method to leverage the unlabeled samples in the target domain. We conduct MS-DTT experiments on two tasks: digit recognition and action recognition, and demonstrate the advantageous performance of the proposed method against multiple baselines.
机译:由于深层卷积网络能够找到从$ \ mathcal {X} $到$ \ mathcal {Y} $的复杂映射,计算机视觉社区正在见证提出和解决新任务的空前高的速度。每个任务的出现通常伴随着大规模带注释的数据集的发布,用于监督深度网络。但是,手动标记足够数量的训练数据既昂贵又耗时。因此,开发可利用现成标签的数据集来学习目标任务的有用知识的算法非常重要。尽管以前的工作主要集中在从单一来源进行转移学习,但我们在半监督的情况下研究跨域和任务(MS-DTT)的多来源转移。我们提出GradMix,一种与模型无关的方法,适用于使用基于梯度的学习规则训练的任何模型,以通过在训练过程中对所有来源的梯度进行加权和混合来通过梯度下降传递知识。 GradMix遵循元学习目标,该目标将分层权重分配给源梯度,以使组合梯度遵循的方向可以最大程度地减少目标数据集中一小部分样本的损失。此外,我们建议根据每个微型批处理对目标任务的重要性来自适应地调整其学习率,并提出一种伪标记方法,以利用目标域中未标记的样本。我们在两个任务上进行了MS-DTT实验:数字识别和动作识别,并证明了该方法在多个基准上的优越性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号