首页> 外文会议>International Conference on Knowledge Science, Engineering and Management >Cost-Sensitive Classification with k-Nearest Neighbors
【24h】

Cost-Sensitive Classification with k-Nearest Neighbors

机译:与k最近邻居的成本敏感分类

获取原文

摘要

Cost-sensitive learning algorithms are typically motivated by imbalance data in clinical diagnosis that contains skewed class distribution. While other popular classification methods have been improved against imbalance data, it is only unsolved to extend k-Nearest Neighbors (kNN) classification, one of top-10 datamining algorithms, to make it cost-sensitive to imbalance data. To fill in this gap, in this paper we study two simple yet effective cost-sensitive kNN classification approaches, called Direct-CS-kNN and Distance-CS-kNN. In addition, we utilize several strategies (i.e., smoothing, minimum-cost k value selection, feature selection and ensemble selection) to improve the performance of Direct-CS-kNN and Distance-CS-kNN. We conduct several groups of experiments to evaluate the efficiency with UCI datasets, and demonstrate that the proposed cost-sensitive kNN classification algorithms can significantly reduce misclassification cost, often by a large margin, as well as consistently outperform CS-4.5 with/without additional enhancements.
机译:成本敏感的学习算法通常由包含偏斜类分布的临床诊断中的不平衡数据而激励。虽然其他流行的分类方法已经针对不平衡数据得到改善,但它仅取消通过扩展K-Collect Neighbors(KNN)分类,以使其成为前10个Datamining算法之一,使其对不平衡数据成本敏感。为了填补这种差距,在本文中,我们研究了两个简单但有效的成本敏感的KNN分类方法,称为Direct-CS-KNN和距离-CS-KNN。此外,我们还利用了几种策略(即,平滑,最小成本k值选择,特征选择和集合选择)来提高Direct-CS-KNN和距离-CS-KNN的性能。我们进行几组实验来评估UCI数据集的效率,并证明所提出的成本敏感的KNN分类算法可以显着降低错误分类成本,通常是大幅度的余量,以及始终如一地优于CS-4.5 /无需额外的增强。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号