首页> 外文会议>Adaptive and natural computing algorithms >Generalized Information-Theoretic Measures for Feature Selection
【24h】

Generalized Information-Theoretic Measures for Feature Selection

机译:特征选择的广义信息论方法

获取原文
获取原文并翻译 | 示例

摘要

Information-theoretic measures are frequently employed to select the most relevant subset of features from datasets. This paper focuses on the analysis of continuous-valued features. We compare the common approach with discretization of features prior the analysis, to the direct usage of exact values. Due to the overwhelming costs of computing continuous information-theoretic measures based on Shannon entropy the Renyi and Tsallis generalized measures are considered. To enable computation with continuous Tsallis measures a novel modification of the information potential is introduced. The quality of the analysed measures was assessed indirectly through the classification accuracy in conjuction with the greedy feature selection process. The experiments on datasets from UCI repository show considerable improvements of the results when using both generalized continuous measures.
机译:信息理论方法经常用于从数据集中选择特征最相关的子集。本文着重分析连续值特征。我们将分析之前将特征离散化的通用方法与直接使用精确值进行比较。由于基于Shannon熵计算连续信息理论测度的巨大成本,因此考虑了Renyi和Tsallis广义测度。为了能够使用连续的Tsallis度量进行计算,引入了一种对信息潜力的新颖修改。通过与贪婪特征选择过程结合的分类准确性,间接评估了分析措施的质量。使用UCI知识库中的数据集进行的实验表明,在同时使用两种广义连续度量时,结果均得到了显着改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号