首页> 外文期刊>Neural Network World >IMPROVING K-NEAREST NEIGHBOR EFFICIENCY FOR TEXT CATEGORIZATION
【24h】

IMPROVING K-NEAREST NEIGHBOR EFFICIENCY FOR TEXT CATEGORIZATION

机译:改进K-近邻文本分类的效率

获取原文
获取原文并翻译 | 示例
           

摘要

With the increasing use of the Internet and electronic documents, automatic text categorization becomes imperative. Many classification methods have been applied to text categorization. The k-nearest neighbors (k-NN) is known to be one of the best state of the art classifiers when used for text categorization. However, k-NN suffers from limitations such as high computation, low tolerance to noise, and its dependency to the parameter k and distance function. In this paper, we first survey some improvements algorithms proposed in the literature to face those shortcomings. And second, we discuss an approach to improve k-NN efficiency without degrading the performance of classification. Experimental results on the 20Newsgroup and Reuters corpora show that the proposed approach increases the performance of k-NN and reduces the time classification.
机译:随着Internet和电子文档的日益普及,自动文本分类变得势在必行。许多分类方法已应用于文本分类。当用于文本分类时,k近邻(k-NN)是最先进的分类器之一。但是,k-NN受诸如计算量大,对噪声的容忍度低以及其对参数k和距离函数的依赖性等限制。在本文中,我们首先调查一些文献中提出的针对这些缺点的改进算法。其次,我们讨论了一种在不降低分类性能的情况下提高k-NN效率的方法。在20Newsgroup和Reuters语料库上的实验结果表明,该方法提高了k-NN的性能并减少了时间分类。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号