首页> 外文会议>International Conference on Progress in Informatics and Computing >A Feature Metric Algorithm Combining the Wasserstein Distance and Mutual Information
【24h】

A Feature Metric Algorithm Combining the Wasserstein Distance and Mutual Information

机译:三架距离和相互信息的特征度量算法

获取原文

摘要

Mutual information based feature selection algorithms measure feature relevance by comparing mutual information between features and class labels, however those features do not necessarily lead to good feature selection and classification accuracy. For each feature dimension, the Wasserstein distance can be used to measure the difference of distribution of categories, which provides a more powerful feature relevance metric method than mutual information. Feature redundancy can still be measured by mutual information. Thus, an optimization objective function combing the measures of Wasserstein distance and mutual information is proposed, which can obtain a smaller feature set with strong feature relevance and less feature redundancy. The effectiveness of the proposed metric method is verified by conducting tests on the UCI datasets. Compared with other common feature selection method such as MRMR, CIFE, MIFS, and MIM, our method reduces the number of selected features by almost 50% to 80% but gets higher accuracy.
机译:基于互信息的特征选择算法测量功能相关性通过比较特征和类标签之间的相互信息,但这些功能不一定导致良好的特征选择和分类准确性。对于每个特征尺寸,Wassersein距离可用于测量类别分布的差异,这提供了比相互信息更强大的特征相关度量方法。功能冗余仍可通过相互信息来衡量。因此,提出了优化目标函数梳理Wassersein距离和互信息的测量,其可以获得具有强特征相关性和更少特征冗余的较小特征集。通过对UCI数据集进行测试来验证所提出的度量方法的有效性。与其他常见特征选择方法相比,如MRMR,CIFE,MIFS和MIM,我们的方法将所选特征的数量降低近50%至80%,但准确性更高。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号