首页> 外文会议>International conference on artificial neural networks >A CFS-Based Feature Weighting Approach to Naive Bayes Text Classifiers
【24h】

A CFS-Based Feature Weighting Approach to Naive Bayes Text Classifiers

机译:朴素贝叶斯文本分类器的基于CFS的特征加权方法

获取原文

摘要

Recent work in supervised learning has shown that naive Bayes text classifiers with strong assumptions of independence among features, such as multinomial naive Bayes (MNB), complement naive Bayes (CNB) and the one-versus-all-but-one model (OVA), have achieved remarkable classification performance. This fact raises the question of whether a naive Bayes text classifier with less restrictive assumptions can perform even better. Responding to this question, we firstly evaluate the correlation-based feature selection (CFS) approach in this paper and find that it performs even worse than the original versions. Then, we propose a CFS-based feature weighting approach to these naive Bayes text classifiers. We call our feature weighted versions FWMNB, FWCNB and FWOVA respectively. Our proposed approach weakens the strong assumptions of independence among features by weighting the correlated features. The experimental results on a large suite of benchmark datasets show that our feature weighted versions significantly outperform the original versions in terms of classification accuracy.
机译:最近在监督学习中的工作表明,朴素贝叶斯文本分类器在功能之间具有独立性很强的假设,例如多项朴素贝叶斯(MNB),互补朴素贝叶斯(CNB)和“一对多”模型(OVA) ,取得了卓越的分类性能。这个事实提出了一个问题,即具有较少限制假设的朴素贝叶斯文本分类器是否可以表现得更好。针对这个问题,我们首先在本文中评估了基于相关的特征选择(CFS)方法,发现它的性能甚至比原始版本差。然后,我们为这些朴素的贝叶斯文本分类器提出了一种基于CFS的特征加权方法。我们分别将功能加权版本称为FWMNB,FWCNB和FWOVA。我们提出的方法通过加权相关特征来削弱特征之间独立性的强大假设。在大量基准数据集上的实验结果表明,在分类准确性方面,我们的特征加权版本明显优于原始版本。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号