首页> 外文会议>International Conference on Cloud Computing and Big Data >Rationalizing the Parameters of K-Nearest Neighbor Classification Algorithm
【24h】

Rationalizing the Parameters of K-Nearest Neighbor Classification Algorithm

机译:合理化k最近邻分类算法的参数

获取原文

摘要

With arrival of big-data era, data mining algorithm becomes more and more important. K nearest neighbor algorithm is a representative algorithm for data classification; it is a simple classification method which is widely used in many fields. But some unreasonable parameters of KNN limit its scope of application, such as sample feature values must be numeric types; Some unreasonable parameters limit its classification efficiency, such as the number of training samples is too much, too high feature dimension; Some unreasonable parameters limit the effect of classification, such as the selection of K value is not reasonable, such as distance calculating method is not reasonable, Class voting method is not reasonable. This paper proposed some methods to rationalize the unreasonable parameters above, such as feature value quantification, Dimension reduction, weighted distance and weighted voting function. This paper uses experimental results based on benchmark data to show the effect.
机译:随着大数据时代的到来,数据挖掘算法变得越来越重要。 K最近邻算法是数据分类的代表性算法;这是一种简单的分类方法,广泛用于许多领域。但是一些不合理的KNN参数限制了其应用范围,例如示例特征值必须是数字类型;一些不合理的参数限制其分类效率,例如训练样本的数量太大,特征尺寸太高;一些不合理的参数限制了分类的效果,例如k值的选择是不合理的,如距离计算方法是不合理的,类投票方法不合理。本文提出了一些方法来合理化上述不合理参数,例如特征值量化,尺寸减小,加权距离和加权投票函数。本文使用基于基准数据的实验结果来显示效果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号