首页> 外文期刊>IEEE transactions on systems, man, and cybernetics. Part B, Cybernetics >Scaling Genetic Programming to Large Datasets Using Hierarchical Dynamic Subset Selection
【24h】

Scaling Genetic Programming to Large Datasets Using Hierarchical Dynamic Subset Selection

机译:使用分层动态子集选择将遗传编程扩展到大型数据集

获取原文
获取原文并翻译 | 示例
           

摘要

The computational overhead of genetic programming (GP) may be directly addressed without recourse to hardware solutions using active learning algorithms based on the random or dynamic subset selection heuristics (RSS or DSS). This correspondence begins by presenting a family of hierarchical DSS algorithms: RSS-DSS, cascaded RSS-DSS, and the balanced block DSS algorithm, where the latter has not been previously introduced. Extensive benchmarking over four unbalanced real-world binary classification problems with 30 000–500 000 training exemplars demonstrates that both the cascade and balanced block algorithms are able to reduce the likelihood of degenerates while providing a significant improvement in classification accuracy relative to the original RSS-DSS algorithm. Moreover, comparison with GP trained without an active learning algorithm indicates that classification performance is not compromised, while training is completed in minutes as opposed to half a day.
机译:遗传编程(GP)的计算开销可以直接解决,而无需借助基于随机或动态子集选择启发式算法(RSS或DSS)的主动学习算法求助于硬件解决方案。这种对应关系从介绍一系列层次化DSS算法开始:RSS-DSS,级联RSS-DSS和平衡块DSS算法,而后者以前没有引入过。使用30 000–500 000训练样本对四个不平衡的现实世界二元分类问题进行了广泛的基准测试,结果表明,级联算法和平衡块算法均能够降低退化的可能性,同时相对于原始RSS- DSS算法。此外,与未经主动学习算法训练的GP的比较表明,分类性能没有受到影响,而训练是在几分钟内完成的,而不是半天。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号