首页> 外文会议>International Conference on Information, Intelligence, Systems and Applications >Why the Naive Bayes approximation is not as Naive as it appears
【24h】

Why the Naive Bayes approximation is not as Naive as it appears

机译:为什么天真贝叶斯近似并不像出现的那样天真

获取原文

摘要

The Naive Bayes approximation and associated classifier is widely used in machine learning and data mining and offers very robust performance across a large spectrum of problem domains. As it depends on a very strong assumption - independence among features - this has been somewhat puzzling. Various hypotheses have been put forward to explain its success and moreover many generalizations have been proposed. In this paper we propose a set of "local" error measures - associated with the likelihood functions for particular subsets of attributes and for each class - and show explicitly how these local errors combine to give a "global" error associated to the full attribute set. By so doing we formulate a framework within which the phenomenon of error cancelation, or augmentation, can be quantitatively evaluated and its impact on classifier performance estimated and predicted a priori. These diagnostics also allow us to develop a deeper and more quantitative understanding of why the Naive Bayes approximation is so robust and under what circumstances one expects it to break down.
机译:Naive Bayes近似和相关分类器广泛用于机器学习和数据挖掘,并在大频谱的问题域中提供非常强大的性能。由于它取决于特征之间的非常强烈的假设 - 这种功能 - 这一直有点令人费解。已经提出了各种假设来解释其成功,而且已经提出了许多概括。在本文中,我们提出了一组“本地”误差测量 - 与特定属性子集的似章函数和每个类的似章相关 - 并且显式显示这些本地错误如何组合以给出与完整属性集关联的“全局”错误。通过这样做,我们可以在其中制定一个框架,其中可以定量地评估错误取消或增强的现象,并且其对分类器性能的影响估计并预测了先验。这些诊断还允许我们制定更深层次,更加定量的理解,为什么天真贝叶斯近似是如此稳健,并且在什么情况下,人们希望它分解。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号