首页> 外文会议> >Error-based pruning of decision trees grown on very large data sets can work!
【24h】

Error-based pruning of decision trees grown on very large data sets can work!

机译:对基于非常大的数据集的决策树进行基于错误的修剪是可行的!

获取原文

摘要

It has been asserted that, using traditional pruning methods, growing decision trees with increasingly larger amounts of training data will result in larger tree sizes even when accuracy does not increase. With regard to error-based pruning, the experimental data used to illustrate this assertion have apparently been obtained using the default setting for pruning strength; in particular, using the default certainty factor of 25 in the C4.5 decision tree implementation. We show that, in general, an appropriate setting of the certainty factor for error-based pruning will cause decision tree size to plateau when accuracy is not increasing with more training data.
机译:有人断言,使用传统的修剪方法,使用越来越多的训练数据来增长决策树将导致树的大小更大,即使准确性没有增加。关于基于错误的修剪,用于说明此断言的实验数据显然已使用修剪强度的默认设置获得;请参见图5。特别是在C4.5决策树实现中使用默认确定性因子25。我们显示,通常,当准确性不随更多训练数据而增加时,对基于错误的修剪进行确定性因子的适当设置将导致决策树大小达到平稳状态。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号