【24h】

B-kNN to Improve the Efficiency of kNN

机译:B-KNN提高KNN效率

获取原文

摘要

The kNN algorithm typically relies on the exhaustive use of training datasets, which aggravates efficiency on large datasets. In this paper, we present the B-kNN algorithm to improve the efficiency of kNN using a two-fold preprocess scheme built upon the notion of minimum and maximum points and boundary subsets. For a given training dataset, B-kNN first identifies classes and for each class, it further identifies the minimum and maximum points (MMP) of the class. A given testing object is evaluated to the MMP of each class. If the object belongs to the MMP, the object is predicted belonging to the class. If not, a boundary subset (BS) is defined for each class. Then, BSs are fed into kNN for determining the class of the object. As BSs are significantly smaller in size than their classes, the efficiency of kNN improves. We present two case studies to evaluate B-kNN. The results show an average of 97% improvement in efficiency over kNN using the entire training dataset, while making little sacrifice of the accuracy compared to kNN.
机译:KNN算法通常依赖于彻底使用训练数据集,这会加剧大型数据集的效率。在本文中,我们介绍了B-KNN算法通过基于最小和最大点和边界子集的概念构建的双折叠预处理方案来提高KNN的效率。对于给定的训练数据集,B-KNN首先识别类和每个类,它进一步标识了类的最小和最大点(MMP)。给定的测试对象被评估为每个类的MMP。如果对象属于MMP,则预先属于该类对象。如果不是,则为每个类定义边界子集(BS)。然后,BSS被送入KNN以确定对象的类别。随着BSS的尺寸明显小于其类,KNN的效率提高了。我们提出了两个案例研究来评估B-KNN。结果显示使用整个训练数据集的效率平均提高了97%,同时与knn相比,牺牲了一点牺牲。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号