首页> 外文会议>European Conference on the Applications of Evolutionary Computation >Neuro-Evolutionary Transfer Learning Through Structural Adaptation
【24h】

Neuro-Evolutionary Transfer Learning Through Structural Adaptation

机译:通过结构适应的神经进化转移学习

获取原文

摘要

Transfer learning involves taking an artificial neural network (ANN) trained on one dataset (the source) and adapting it to a new, second dataset (the target). While transfer learning has been shown to be quite powerful and is commonly used in most modern-day statistical learning setups, its use has generally been restricted by architecture, i.e., in order to facilitate the reuse of internal learned synaptic weights, the underlying topology of the ANN to be transferred across tasks must remain the same and a new output layer must be attached (entailing removing the old output layer's weights). This work removes this restriction by proposing a neuro-evolutionary approach that facilitates what we call adaptive structure transfer learning, which means that an ANN can be transferred across tasks that have different input and output dimensions while having the internal latent structure continuously optimized. We test the proposed optimizer on two challenging real-world time series prediction problems - our process adapts recurrent neural networks (RNNs) to (1) predict coal-fired power plant data before and after the addition of new sensors, and to (2) predict engine parameters where RNN estimators are trained on different airframes with different engines. Experiments show that not only does the proposed neuro-evolutionary transfer learning process result in RNNs that evolve and train faster on the target set than those trained from scratch but, in many cases, the RNNs generalize better even after a long training and evolution process. To our knowledge, this work represents the first use of neuro-evolution for transfer learning, especially for RNNs, and is the first methodological framework capable of adapting entire structures for arbitrary input/output spaces.
机译:转移学习涉及采用在一个数据集(源)上培训的人工神经网络(ANN)并将其调整为新的第二个数据集(目标)。虽然转移学习已被证明是相当强大的,并且通常用于大多数现代统计学习设置,但其使用通常受架构的限制,即,以便于重复使用内部学习的突触权重,基础拓扑要跨任务传输的ANN必须保持相同,并且必须附加新的输出层(因此删除旧输出层的权重)。这项工作通过提出一种神经进化方法来消除这种限制,这促进了我们所谓的自适应结构转移学习的内容,这意味着可以跨越具有不同输入和输出尺寸的任务传输ANN,同时具有内部潜在结构连续优化。我们在两个具有挑战性的真实世界时间序列预测问题上测试所提出的优化器 - 我们的过程适应经常性的神经网络(RNN)到(1)在添加新传感器之前和之后预测燃煤发电厂数据,以及(2)预测发动机参数,其中RNN估计器在具有不同发动机的不同机架上培训。实验表明,建议的神经进化转移学习过程导致RNN的结果不仅可以在目标集上发展和培训比从划痕训练的那些培训,但在许多情况下,即使在长期训练和演进过程之后,RNNS也会更好地推广。为了我们的知识,这项工作代表了对转移学习的第一次evolution的首次使用,特别是对于RNN,是一种能够适应任意输入/输出空间的整个结构的第一种方法框架。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号