首页> 外文会议>Conference on uncertainty in artificial intelligence >Active Learning with Distributional Estimates
【24h】

Active Learning with Distributional Estimates

机译:主动学习与分布估计

获取原文

摘要

Active Learning (AL) is increasingly important in a broad range of applications. Two main AL principles to obtain accurate classification with few labeled data are refinement of the current decision boundary and exploration of poorly sampled regions. In this paper we derive a novel AL scheme that balances these two principles in a natural way. In contrast to many AL strategies, which are based on an estimated class conditional probability p(yx), a key component of our approach is to view this quantity as a random variable, hence explicitly considering the uncertainty in its estimated value. Our main contribution is a novel mathematical framework for uncertainty-based AL, and a corresponding AL scheme, where the uncertainty in p(yx) is modeled by a second-order distribution. On the practical side, we show how to approximate such second-order distributions for kernel density classification. Finally, we find that over a large number of UCI, USPS and Caltech-4 datasets, our AL scheme achieves significantly better learning curves than popular AL methods such as uncertainty sampling and error reduction sampling, when all use the same kernel density classifier.
机译:主动学习(AL)在广泛的应用中越来越重要。获得带有少量标记数据的准确分类的两个主要AL原则是对当前决策边界的细化和对采样较差区域的探索。在本文中,我们得出了一种新颖的AL方案,可以自然地平衡这两个原理。与许多基于估计的类别条件概率p(y \ x)的AL策略相比,我们方法的关键组成部分是将该数量视为随机变量,因此明确考虑了其估计值的不确定性。我们的主要贡献是基于不确定性的AL的新颖数学框架,以及相应的AL方案,其中p(y \ x)中的不确定性是通过二阶分布建模的。在实践方面,我们展示了如何对内核密度分类进行近似的二阶分布。最后,我们发现在所有UCI,USPS和Caltech-4数据集上,当全部使用相同的内核密度分类器时,我们的AL方案比诸如不确定性采样和误差减少采样之类的流行AL方法获得了更好的学习曲线。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号