首页> 外文会议>International Conference on Pattern Recognition >Rethinking deep active learning: Using unlabeled data at model training
【24h】

Rethinking deep active learning: Using unlabeled data at model training

机译:重新思考深度主动学习:在模型培训时使用未标记的数据

获取原文

摘要

Active learning typically focuses on training a model on few labeled examples alone, while unlabeled ones are only used for acquisition. In this work we depart from this setting by using both labeled and unlabeled data during model training across active learning cycles. We do so by using unsupervised feature learning at the beginning of the active learning pipeline and semi-supervised learning at every active learning cycle, on all available data. The former has not been investigated before in active learning, while the study of latter in the context of deep learning is scarce and recent findings are not conclusive with respect to its benefit. Our idea is orthogonal to acquisition strategies by using more data, much like ensemble methods use more models. By systematically evaluating on a number of popular acquisition strategies and datasets, we find that the use of unlabeled data during model training brings a spectacular accuracy improvement in image classification, compared to the differences between acquisition strategies. We thus explore smaller label budgets, even one label per class.
机译:主动学习通常侧重于单独绘制少数标记的示例的模型,而未标记的则仅用于采集。在这项工作中,我们通过在主动学习周期的模型培训期间使用标记和未标记的数据来离开此设置。我们通过在所有活动学习循环开始在主动学习管道的开始和半监督学习时使用无监督的功能学习,在所有可用数据上。前者尚未在积极学习之前进行调查,而后者在深度学习的背景下的研究是稀缺的,而最近的发现对于它的利益而不是决定性的。我们的想法是通过使用更多数据来获取策略的正交,很像集合方法使用更多型号。通过系统地评估许多流行的采购策略和数据集,我们发现在模型培训期间使用未标记的数据带来了图像分类中的壮观准确性,而采集策略之间的差异相比。因此,我们探索了较小的标签预算,甚至每班一个标签。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号