...
首页> 外文期刊>Artificial intelligence >On the consistency of multi-label learning
【24h】

On the consistency of multi-label learning

机译:关于多标签学习的一致性

获取原文
获取原文并翻译 | 示例

摘要

Multi-label learning has attracted much attention during the past few years. Many multi-label approaches have been developed, mostly working with surrogate loss functions because multi-label loss functions are usually difficult to optimize directly owing to their non-convexity and discontinuity. These approaches are effective empirically, however, little effort has been devoted to the understanding of their consistency, i.e., the convergence of the risk of learned functions to the Bayes risk. In this paper, we present a theoretical analysis on this important issue. We first prove a necessary and sufficient condition for the consistency of multi-label learning based on surrogate loss functions. Then, we study the consistency of two well-known multi-label loss functions, i.e., ranking loss and hamming loss. For ranking loss, our results disclose that, surprisingly, none of convex surrogate loss is consistent; we present the partial ranking loss, with which some surrogate losses are proven to be consistent. We also discuss on the consistency of univariate surrogate losses. For hamming loss, we show that two multi-label learning methods, i.e., one-vs-all and pairwise comparison, which can be regarded as direct extensions from multi-class learning, are inconsistent in general cases yet consistent under the dominating setting, and similar results also hold for some recent multi-label approaches that are variations of one-vs-all. In addition, we discuss on the consistency of learning approaches that address multi-label learning by decomposing into a set of binary classification problems.
机译:在过去的几年中,多标签学习吸引了很多关注。已经开发了许多多标签方法,主要使用代理损失函数,因为多标签损失函数通常由于其非凸性和不连续性而通常难以直接优化。这些方法在经验上是有效的,但是,很少有努力用于理解它们的一致性,即,学习功能的风险与贝叶斯风险的收敛。在本文中,我们对这一重要问题进行了理论分析。我们首先证明了基于替代损失函数的多标签学习一致性的充要条件。然后,我们研究了两个众所周知的多标签损失函数的一致性,即秩损失和汉明损失。对于等级损失,我们的结果表明,令人惊讶的是,没有一个凸替代代理损失是一致的。我们提出了部分排名损失,与之相关的一些替代损失被证明是一致的。我们还讨论了单变量代理损失的一致性。对于汉明损失,我们证明了两种多标签学习方法,即一对多比较和成对比较,可以看作是多类学习的直接扩展,在一般情况下是不一致的,但在主要情况下是一致的,类似的结果也适用于一些最新的多标签方法,即一对多的变体。此外,我们讨论了通过分解为一组二进制分类问题来解决多标签学习的学习方法的一致性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号