【24h】

Contribution of Boosting in Wrapper Models

机译:促进包装模型的贡献

获取原文

摘要

We describe a new way to deal with feature selection when boosting is used to assess the relevancy of feature subsets. In the context of wrapper models, the accuracy is here replaced as a performance function by a particular exponential criterion, usually optimized in boosting algorithms. A first experimental study brings to the fore the relevance of our approach. However, this new "boosted" strategy needs the construction at each step of many learners, leading to high computational costs. We focus then, in a second part, on how to speed-up boosting convergence to reduce this complexity. We propose a new update of the instance distribution, which is the core of a boosting algorithm. We exploit these results to implement a new forward selection algorithm which converges much faster using overbiased distributions over learning instances. Speed-up is achieved by reducing the number of weak hypothesis when many identical observations are shared by different classes. A second experimental study on the UCI repository shows significantly speeding improvements with our new update without altering the feature subset selection.
机译:我们描述了一种在升压时处理特征选择的新方法来评估特征子集的相关性。在包装模型的背景下,这里的准确度被特定指数标准替换为性能函数,通常在升压算法中优化。第一个实验研究带来了我们方法的相关性。然而,这种新的“提升”策略需要在许多学习者的每个步骤中建造,导致高计算成本。然后,我们专注于第二部分,如何加速升高趋同以降低这种复杂性。我们提出了一个新的实例分布的更新,这是升压算法的核心。我们利用这些结果实现了一种新的前向选择算法,它使用过度偏航的分布在学习实例上收敛得多。当不同类别共享许多相同的观察时,通过减少弱假设的数量来实现加速。对UCI存储库的第二个实验研究显示,通过我们的新更新,显着超速改进,而无需更改特征子集选择。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号