首页> 外文会议>International Joint Conference on Neural Networks >Learning optimization for decision tree classification of non-categorical data with information gain impurity criterion
【24h】

Learning optimization for decision tree classification of non-categorical data with information gain impurity criterion

机译:具有信息增益杂质准则的非分类数据决策树分类的学习优化

获取原文

摘要

We consider the problem of construction of decision trees in cases when data is non-categorical and is inherently high-dimensional. Using conventional tree growing algorithms that either rely on univariate splits or employ direct search methods for determining multivariate splitting conditions is computationally prohibitive. On the other hand application of standard optimization methods for finding locally optimal splitting conditions is obstructed by abundance of local minima and discontinuities of classical goodness functions such as e.g. information gain or Gini impurity. In order to avoid this limitation a method to generate smoothed replacement for measuring impurity of splits is proposed. This enables to use vast number of efficient optimization techniques for finding locally optimal splits and, at the same time, decreases the number of local minima. The approach is illustrated with examples.
机译:当数据是非分类的并且本质上是高维的情况下,我们考虑构造决策树的问题。在计算上,使用依赖于单变量拆分或采用直接搜索方法来确定多元拆分条件的常规树木生长算法是无法实现的。另一方面,标准优化方法在寻找局部最优分裂条件上的应用由于局部极小值的丰富和经典善函数的不连续性而受到阻碍,例如信息增益或基尼杂质。为了避免这种限制,提出了一种生成平滑替换以测量分割的杂质的方法。这使得可以使用大量有效的优化技术来查找局部最优分割,同时减少局部最小值的数量。通过示例说明了该方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号