首页> 外文会议>International Joint Conference on Natural Language Processing;Annual Meeting of the Association for Computational Linguistics >Matching Distributions between Model and Data: Cross-domain Knowledge Distillation for Unsupervised Domain Adaptation
【24h】

Matching Distributions between Model and Data: Cross-domain Knowledge Distillation for Unsupervised Domain Adaptation

机译:模型和数据之间的匹配分布:无监督域适应的跨域知识蒸馏

获取原文
获取外文期刊封面目录资料

摘要

Unsupervised Domain Adaptation (UDA) aims to transfer the knowledge of source domain to the unlabeled target domain. Existing methods typically require to learn to adapt the target model by exploiting the source data and sharing the network architecture across domains. However, this pipeline makes the source data risky and is inflexible for deploying the target model. This paper tackles a novel setting where only a trained source model is available and different network architectures can be adapted for target domain in terms of deployment environments. We propose a generic framework named Cross-domain Knowledge Distillation (CdKD) without needing any source data. CdKD matches the joint distributions between a trained source model and a set of target data during distilling the knowledge from the source model to the target domain. As a type of important knowledge in the source domain, for the first time, the gradient information is exploited to boost the transfer performance. Experiments on cross-domain text classification demonstrate that CdKD achieves superior performance, which verifies the effectiveness in this novel setting.
机译:无监督域适应(UDA)旨在将源域的知识转移到未标记的目标域。现有方法通常需要学习通过利用源数据并在域中共享网络架构来调整目标模型。但是,该管道使源数据有风险,并且对于部署目标模型是不灵活的。本文解决了一个新颖的设置,其中只有训练有素的源模型可用,并且可以在部署环境方面适用于目标域的不同网络架构。我们提出了一个名为跨域知识蒸馏(CDKD)的通用框架,而无需任何源数据。 CDKD匹配训练源模型与一组目标数据在蒸馏到目标域的知识期间的联合分布。作为源域中的重要知识,首次利用梯度信息来提高传输性能。跨域文本分类的实验表明CDKD实现了卓越的性能,这验证了这部小型环境中的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号