首页> 外文会议>Computing science and statistics >Proximity Graphs for Nearest NeighborDecision Rules: Recent Progress
【24h】

Proximity Graphs for Nearest NeighborDecision Rules: Recent Progress

机译:最近邻居决策规则的接近图:最新进展

获取原文
获取原文并翻译 | 示例

摘要

In the typical nonparametric approach to pattern classifcation, randomrndata (the training set of patterns) are collected and used to design a decisionrnrule (classifer). One of the most well known such rules is the к-nearest-rnneighbor decision rule (also known as instance-based learning, and lazy learn-rning) in which an unknown pattern is classifed into the majority class amongrnits к nearest neighbors in the training set. Several questions related to thisrnrule have received considerable attention over the years. Such questions in-rnclude the following. How can the storage of the training set be reducedrnwithout degrading the performance of the decision rule? How should the re-rnduced training set be selected to represent the diu000berent classes? How largernshould к be? How should the value of к be chosen? Should all к neighborsrnbe equally weighted when used to decide the class of an unknown pattern? Ifrnnot, how should the weights be chosen? Should all the features (attributes)rnwe weighted equally and if not how should the feature weights be chosen?rnWhat distance metric should be used? How can the rule be made robust tornoverlapping classes or noise present in the training data? How can the rule bernmade invariant to scaling of the measurements? Geometric proximity graphsrnsuch as Voronoi diagrams and their many relatives provide elegant solutionsrnto most of these problems. After a brief and non-exhaustive review of somernof the classical canonical approaches to solving these problems, the methodsrnthat use proximity graphs are discussed, some new observations are made,rnand avenues for further research are proposed.
机译:在典型的模式分类非参数方法中,随机数据(模式的训练集)被收集并用于设计决策规则(分类器)。最著名的此类规则之一是“最近邻决策”规则(也称为基于实例的学习和“懒惰学习”规则),其中,未知模式被分类为训练中最接近的邻居中的多数类。组。多年来,与此规则有关的几个问题受到了相当多的关注。这些问题包括以下内容。如何在不降低决策规则性能的情况下减少训练集的存储?重新选择的训练集应如何选择以代表不同的班级? к应该多大? к的值应如何选择?当用于决定未知模式的类别时,是否应该对所有к邻居进行均等的加权?如果不是,应该如何选择权重?是否应该对所有特征(属性)均等地加权,否则应如何选择特征权重?应该使用哪种距离度量?如何使规则变得健壮的重叠类或训练数据中出现的噪声?规则如何使测量定标不变?诸如Voronoi图之类的几何接近图及其许多亲属为大多数这些问题提供了很好的解决方案。在对某些经典的经典方法进行简要简短的综述之后,讨论了使用邻近图的方法,提出了一些新的观察方法,并提出了进一步研究的途径。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号