首页> 外文会议>Algorithmic learning theory >Learning a Classifier when the Labeling Is Known
【24h】

Learning a Classifier when the Labeling Is Known

机译:知道标签时学习分类器

获取原文
获取原文并翻译 | 示例

摘要

We introduce a new model of learning, Known-Labeling-Classifier-Learning (KLCL). The goal of such learning is to find a low-error classifier from some given target-class of predictors, when the correct labeling is known to the learner. This learning problem can be viewed as measuring the information conveyed by the identity of input examples, rather than by their labels. Given some class of predictors H., a labeling function, and an i.i.d. unlabeled sample generated by some unknown data distribution, the goal of our learner is to find a classifier in H that has as low as possible error with respect to the sample-generating distribution and the given labeling function. When the labeling function does not belong to the target class, the error of members of the class (and thus their relative quality as label predictors) varies with the marginal of the underlying data distribution. We prove a trichotomy with respect to the KLCL sample complexity. Namely, we show that for any learnable concept class H, its KLCL sample complexity is either 0 or Θ(l/ε) or Ω(1/ε~2). Furthermore, we give a simple combinatorial property of concept classes that characterizes this trichotomy. Our results imply new sample-size lower bounds for the common agnostic PAC model - a lower bound of Ω(l/ε~2) on the sample complexity of learning deterministic classifiers, as well as novel results about the utility of unlabeled examples in a semi-supervised learning setup.
机译:我们介绍了一种新的学习模型,即已知标签分类器学习(KLCL)。这种学习的目的是在学习者知道正确的标签时,从某些给定的预测变量目标类中找到低错误的分类器。可以将这种学习问题视为衡量输入示例的身份而不是其标签所传达的信息。给定一类预测变量H.,标记函数和i.d.由于某些未知数据分布生成的未标记样本,我们的学习者的目标是找到H的分类器,该分类器相对于样本生成分布和给定的标记函数具有尽可能低的误差。当标签功能不属于目标类别时,类别成员的错误(以及它们作为标签预测变量的相对质量)会随着基础数据分布的边际而变化。我们证明了关于KLCL样本复杂性的三分法。即,我们表明,对于任何可学习的概念类H,其KLCL样本复杂度为0或Θ(l /ε)或Ω(1 /ε〜2)。此外,我们给出了概念类的简单组合属性,它描述了这种三分法。我们的结果暗示了通用不可知PAC模型的新样本大小下限-学习确定性分类器的样本复杂度的Ω(l /ε〜2)下界,以及关于未标记示例在a中的实用性的新结果半监督学习设置。

著录项

  • 来源
    《Algorithmic learning theory》|2011年|p.440-451|共12页
  • 会议地点 Espoo(FI);Espoo(FI)
  • 作者单位

    Faculty of Mathematics, University of Waterloo, Waterloo, ON N2L 3G1, Canada;

    David R. Cheriton School of Computer Science, University of Waterloo, Waterloo, ON N2L 3G1, Canada;

  • 会议组织
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 人工智能理论;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号