首页> 外文期刊>Neural processing letters >Improve Semi-supervised Learning with Metric Learning Clusters and Auxiliary Fake Samples
【24h】

Improve Semi-supervised Learning with Metric Learning Clusters and Auxiliary Fake Samples

机译:通过公制学习群集和辅助假样本改善半监督学习

获取原文
获取原文并翻译 | 示例

摘要

Because it is very expensive to collect a large number of labeled samples to train deep neural networks in certain fields, semi-supervised learning (SSL) researcher has become increasingly important in recent years. There are many consistency regularization-based methods for solving SSL tasks, such as the Pi model and mean teacher. In this paper, we first show through an experiment that the traditional consistency-based methods exist the following two problems: (1) as the size of unlabeled samples increases, the accuracy of these methods increases very slowly, which means they cannot make full use of unlabeled samples. (2) When the number of labeled samples is vary small, the performance of these methods will be very low. Based on these two findings, we propose two methods, metric learning clustering (MLC) and auxiliary fake samples, to alleviate these problems. The proposed methods achieve state-of-the-art results on SSL benchmarks. The error rates are 10.20%, 38.44% and 4.24% for CIFAR-10 with 4000 labels, CIFAR-100 with 10,000 labels and SVHN with 1000 labels by using MLC. For MNIST, the auxiliary fake samples method shows great results in cases with the very few labels.
机译:因为收集大量标记的样本是非常昂贵的,以培训某些领域的深度神经网络,半监督学习(SSL)研究人员近年来越来越重要。解决SSL任务的基于基于正常化的方法,例如PI模型和均值教师。在本文中,我们首先通过实验展示了传统的基于一致性的方法存在以下两个问题:(1)随着未标记样本的尺寸增加,这些方法的准确性速度非常缓慢地增加,这意味着它们不能充分利用未标记的样本。 (2)当标记样品的数量变化时,这些方法的性能将非常低。基于这两个调查结果,我们提出了两种方法,度量学习聚类(MLC)和辅助假样本,以减轻这些问题。所提出的方法在SSL基准测试中实现最先进的结果。使用4000个标签,带有100,000个标签和SVHN的CIFAR-10,误差率为10.20%,38.44%和4.24%,使用MLC,带有100,000个标签和1000个标签的SVHN。对于Mnist,辅助假样品方法在少数标签中显示出很大的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号