首页> 外文期刊>International Journal of Artificial Intelligence Tools: Architectures, Languages, Algorithms >AN L_1-REGULARIZED NAIVE BAYES-INSPIRED CLASSIFIER FOR DISCARDING REDUNDANT AND IRRELEVANT PREDICTORS
【24h】

AN L_1-REGULARIZED NAIVE BAYES-INSPIRED CLASSIFIER FOR DISCARDING REDUNDANT AND IRRELEVANT PREDICTORS

机译:L_1调节朴素贝叶斯启发分类器,用于消除冗余和无关紧要的预测

获取原文
获取原文并翻译 | 示例
           

摘要

The naive Bayes model is a simple but often satisfactory supervised classification method. The original naive Bayes scheme, does, however, have a serious weakness, namely, the harmful effect of redundant predictors. In this paper, we study how to apply a regularization technique to learn a computationally efficient classifier that is inspired by naive Bayes. The proposed formulation, combined with an L_1-penalty, is capable of discarding harmful, redundant predictors. A modification of the LARS algorithm is devised to solve this problem. We tackle both real-valued and discrete predictors, assuring that our method is applicable to a wide range of data. In the experimental section, we empirically study the effect of redundant and irrelevant predictors. We also test the method on a high-dimensional data set from the neuroscience field, where there are many more predictors than data cases. Finally, we run the method on a real data set than combines categorical with numeric predictors. Our approach is compared with several naive Bayes variants and other classification algorithms (SVM and kNN), and is shown to be competitive.
机译:朴素贝叶斯模型是一种简单但通常令人满意的监督分类方法。但是,原始的朴素贝叶斯方案确实存在严重的弱点,即冗余预测变量的有害影响。在本文中,我们研究如何应用正则化技术来学习受朴素贝叶斯启发的高效计算分类器。所提出的公式与L_1罚则相结合,能够丢弃有害的,多余的预测变量。为了解决这个问题,对LARS算法进行了修改。我们处理实值和离散预测变量,确保我们的方法适用于广泛的数据。在实验部分,我们根据经验研究冗余和不相关的预测因素的影响。我们还在神经科学领域的高维数据集上测试了该方法,在该数据集上,预测指标比数据案例要多得多。最后,我们在将实际分类与数字预测变量相结合的实际数据集上运行该方法。我们的方法与几种朴素的贝叶斯变体和其他分类算法(SVM和kNN)进行了比较,并显示出竞争优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号