首页> 外文会议>International conference on Machine learning >How boosting the margin can also boost classifier complexity
【24h】

How boosting the margin can also boost classifier complexity

机译:如何提高利润率还可以提高分类器的复杂性

获取原文

摘要

Boosting methods are known not to usually overfit training data even as the size of the generated classifiers becomes large. Schapire et al. attempted to explain this phenomenon in terms of the margins the classifier achieves on training examples. Later, however, Breiman cast serious doubt on this explanation by introducing a boosting algorithm, arc-gv, that can generate a higher margins distribution than AdaBoost and yet performs worse. In this paper, we take a close look at Breiman's compelling but puzzling results. Although we can reproduce his main finding, we find that the poorer performance of arc-gv can be explained by the increased complexity of the base classifiers it uses, an explanation supported by our experiments and entirely consistent with the margins theory. Thus, we find maximizing the margins is desirable, but not necessarily at the expense of other factors, especially base-classifier complexity.
机译:众所周知,即使生成的分类器的大小变大,增强方法也通常不会过度拟合训练数据。 Schapire等。试图用分类器在训练实例上获得的余量来解释这种现象。然而,后来,布雷曼通过引入增强算法arc-gv对该解释提出了严重怀疑,该算法可以产生比AdaBoost更高的边距分布,但性能却更差。在本文中,我们仔细观察了布雷曼引人入胜但令人困惑的结果。尽管我们可以重现他的主要发现,但我们发现arc-gv较差的性能可以由其使用的基本分类器的复杂性增加来解释,该分类器得到了我们实验的支持,并且与边距理论完全一致。因此,我们发现最大化边距是可取的,但不一定以其他因素为代价,尤其是基本分类器的复杂性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号