首页> 外文会议>Third International Conference on Data Mining, 2002 >Selective Bayesian classifier: feature selection for the Naieve Bayesian classifier using decision trees
【24h】

Selective Bayesian classifier: feature selection for the Naieve Bayesian classifier using decision trees

机译:选择性贝叶斯分类器:使用决策树为Naieve贝叶斯分类器进行特征选择

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

It is known that Nai've Bayesian classifier (NB) works very well on some domains, and poorly on some. The performance of NB suffers in domains that involve correlated features. C4.5 decision trees, on the other hand, typically perform better than the Nai've Bayesian algorithm on such domains. This paper describes a Selective Bayesian classifier (SBC) that simply uses only those features that C4.5 would use in its decision tree when learning a small example of a training set, a combination of the two different natures of classifiers. Experiments conducted on eleven datasets indicate that SBC performs reliably better than NB on all domains, and SBC outperforms C4.5 on many datasets of which C4.5 outperform NB. SBC also can eliminate, on most cases, more than half of the original attributes, which can greatly reduce the size of the training and test data, as well as the running tune. Further, the SBC algorithm typically learns faster than both C4.5 and NB, needing fewer training examples to reach high accuracy of classification.
机译:众所周知,Nai've贝叶斯分类器(NB)在某些域上效果很好,而在某些域上效果不佳。 NB的性能在涉及相关功能的域中会受到影响。另一方面,C4.5决策树在此类域上的性能通常优于Nai've Bayesian算法。本文介绍了一种选择性贝叶斯分类器(SBC),当学习训练集的一个小例子(分类器的两种不同性质的组合)时,仅使用C4.5在决策树中使用的那些功能。在11个数据集上进行的实验表明,在所有域上SBC的性能都比NB可靠,并且在许多C4.5优于NB的数据集上,SBC的性能均优于C4.5。在大多数情况下,SBC还可以消除一半以上的原始属性,这可以极大地减少训练和测试数据以及运行曲调的大小。此外,SBC算法通常比C4.5和NB都学习速度更快,需要更少的训练示例即可达到较高的分类精度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号