首页> 外文期刊>International Journal of Performability Engineering >Decision Tree Incremental Learning Algorithm Oriented Intelligence Data
【24h】

Decision Tree Incremental Learning Algorithm Oriented Intelligence Data

机译:决策树增量学习算法面向智能数据

获取原文
获取原文并翻译 | 示例
           

摘要

Decision tree is one of the most popular classification methods because of its advantages of easy comprehension. However, the decision tree constructed by existed methods is usually too large and complicated. So, in some applications, the practicability is limited. In this paper, combining NOLCDT with IID5R algorithm, an improved hybrid classifier algorithm, HCS, is proposed. HCS algorithm consists of two phases: building initial decision tree and incremental learning. The initial decision tree is constructed according to the NOLCDT algorithm, and then the incremental learning is performed with IID5R. The NOLCDT algorithm selects the candidate attribute with the largest information gain and divides the node into two branches, which avoids generating too many branches. Thus, this prevents the decision tree is too complex. The NOLCDT algorithm also improves on the selection of the next node to be split, which computes the corresponding nodal splitting measure for all candidate splits, and always selects the node which has largest information gain from all candidate split nodes as the next split node, so that each split has the greatest information gain. In addition, based on ID5R, an improved algorithm IID5R is proposed to evaluate the quality of classification attributes and estimates a minimum number of steps for which these attributes are guaranteed such a selection. HCS takes advantage of the decision tree and the incremental learning method, which is easy to understand and suitable for incremental learning. The contrast experiment between the traditional decision tree algorithm and HCS algorithm with UCI data set is proposed; the experimental results show that HCS can solve the increment problem very well. The decision tree is simpler so that it is easy to understand, and so the incremental phase consumes less time.
机译:决策树是最受欢迎的分类方法之一,因为它很容易理解。然而,由存在的方法构建的决策树通常太大并且复杂。因此,在某些应用中,实用性有限。本文将NOLCDT与IID5R算法组合,提出了一种改进的混合分类器算法,HCS。 HCS算法由两个阶段组成:构建初始决策树和增量学习。初始决策树根据NOLCDT算法构建,然后使用IID5R执行增量学习。 NOLCDT算法选择具有最大信息增益的候选属性,并将节点划分为两个分支,避免生成太多分支。因此,这可以防止决策树太复杂。 NOLCDT算法还提高了要拆分的下一个节点的选择,这计算了所有候选分割的相应的Nodal拆分度量,并且始终选择具有从所有候选拆分节点具有最大信息增益的节点作为下一个分割节点,因此每个拆分都有最大的信息增益。另外,基于ID5R,提出了一种改进的算法IID5R来评估分类属性的质量,并估计这些属性保证这种选择的最小步骤数。 HCS利用决策树和增量学习方法,这易于理解,适合增量学习。提出了具有UCI数据集的传统决策树算法和HCS算法之间的对比试验;实验结果表明,HCS可以很好地解决增量问题。决策树更简单,以便易于理解,因此增量阶段消耗更少的时间。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号