...
首页> 外文期刊>Knowledge-Based Systems >Combining active learning and semi-supervised learning to construct SVM classifier
【24h】

Combining active learning and semi-supervised learning to construct SVM classifier

机译:结合主动学习和半监督学习构建支持向量机分类器

获取原文
获取原文并翻译 | 示例
           

摘要

One key issue for most classification algorithms is that they need large amounts of labeled samples to train the classifier. Since manual labeling is time consuming, researchers have proposed technologies of active learning and semi-supervised learning to reduce manual labeling workload. There is a certain degree of complementarity between active learning and semi-supervised learning, and therefore some researches combine them to further reduce manual labeling workload. However, researches on combining active learning and semi-supervised learning for SVM classifier are rare. Of numerous SVM active learning algorithms, the most popular is the one that queries the sample closest to the current classification hyperplane in each iteration, which is denoted as SVMAL in this paper. Realizing that SVM_(AL) is only interested in samples that are more likely to be on the class boundary, while ignoring the usage of the rest large amounts of unlabeled samples, this paper designs a semi-supervised learning algorithm to make full use of the rest non-queried samples, and further forms a new active semi-supervised SVM algorithm. The proposed active semi-supervised SVM algorithm uses active learning to select class boundary samples, and semi-supervised learning to select class central samples, for class central samples are believed to better describe the class distribution, and to help SVMAl finding the boundary samples more precisely. In order not to introduce too many labeling errors when exploring class central samples, the label changing rate is used to ensure the reliability of the predicted labels. Experimental results show that the proposed active semi-supervised SVM algorithm performs much better than the pure SVM active learning algorithm, and thus can further reduce manual labeling workload.
机译:大多数分类算法的一个关键问题是它们需要大量的标记样本来训练分类器。由于手动标记非常耗时,因此研究人员提出了主动学习和半监督学习的技术,以减少手动标记的工作量。主动学习和半监督学习之间存在一定程度的互补性,因此一些研究将它们结合起来以进一步减少手动标注工作量。但是,很少有将主动学习和半监督学习相结合的支持向量机分类器的研究。在众多的SVM主动学习算法中,最流行的是在每次迭代中查询最接近当前分类超平面的样本,在本文中称为SVMAL。意识到SVM_(AL)只对更可能在类边界上的样本感兴趣,而忽略了其余大量未标记样本的使用,本文设计了一种半监督学习算法来充分利用SVM_(AL)休息非查询样本,并进一步形成一种新的主​​动半监督SVM算法。提出的主动半监督SVM算法使用主动学习来选择类边界样本,并使用半监督学习来选择类中心样本,因为相信类中心样本可以更好地描述类分布,并帮助SVMA1更多地找到边界样本。恰好。为了在探索班级中心样本时不会引入太多标签错误,使用标签变化率来确保预测标签的可靠性。实验结果表明,提出的主动半监督支持向量机算法比纯支持向量机主动学习算法具有更好的性能,可以进一步减少人工标注的工作量。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号