首页> 外文会议>International conference on artificial neural networks >EvoCNN: Evolving Deep Convolutional Neural Networks Using Backpropagation-Assisted Mutations
【24h】

EvoCNN: Evolving Deep Convolutional Neural Networks Using Backpropagation-Assisted Mutations

机译:EvoCNN:使用反向传播辅助突变技术发展深层卷积神经网络

获取原文
获取外文期刊封面目录资料

摘要

In this abstract we present our initial results with a novel genetic algorithms based method for evolving convolutional neural networks (CNN). Currently, standard backpropagation is the primary training method for neural networks (NN) in general, including CNNs. In the past several methods proposed using genetic algorithms (GA) for training neural networks. These methods involve representing the weights of the NN as a chromosome, creating a randomly initialized population of such chromosomes (each chromosome represents one NN), and then evolving the population by performing the steps (1) measure the fitness of each chromosome (the lower the average loss over the training set, the better), (2) select the fitter chromosomes for breeding, (3) perform crossover between the parents (randomly choose weights from the parents to create the offspring), and (4) mutate the offspring. While in smaller NNs these methods obtained results comparable with backpropagation, their results deteriorate as the size of NN grows, and are impractical for training deep neural nets. Nowadays these methods have largely been abandoned due to this inefficiency. We propose a combination of GA-based evolution and backpropagation for evolving CNN as follows. Similar to the abovementioned methods we create an initial population of N chromosomes, each representing the weights of one CNN, and then evolve the chromosomes by applying fitness evaluation, crossover, and mutations, but with several key differences: (1) During crossover, instead of randomly selecting weights from each of the two parents, randomly select entire filters from each parent (this ensures that a useful filter is copied in its entirety, rather than being disrupted), (2) During mutation, modify weights by performing standard backpropagation, instead of random changes; and then randomly set a small portion of weights to zero (these steps allow for a more goal-oriented evolution, and zeroing some weights encourages sparsity in the network and has a regularizing effect). We refer to this method as EvoCNN. To measure the performance of our method, we ran several experiments on the MNIST handwritten digit recognition dataset. A standard CNN architecture was used containing the following layers: [Input size 28 × 28]-[convolution with 128 filters of size 5 × 5]-[max-pooling]-[convolution with 256 filters of size 3 × 3]-[max-pooling]-[fully connected layer of size 1000]-[softmax layer of size 10].
机译:在此摘要中,我们用一种基于新颖遗传算法的进化卷积神经网络(CNN)方法展示了我们的初步结果。当前,标准的反向传播通常是包括CNN在内的神经网络(NN)的主要训练方法。在过去,提出了几种使用遗传算法(GA)训练神经网络的方法。这些方法涉及将NN的权重表示为一条染色体,创建一个随机初始化的此类染色体的种群(每个染色体代表一个NN),然后通过执行以下步骤来进化种群:(1)测量每个染色体的适应度(较低的(2)选择适合的染色体进行繁殖,(3)在父母之间进行交叉(从父母那里随机选择权重以创建后代),以及(4)对后代进行突变。尽管在较小的NN中,这些方法获得的结果可与反向传播相媲美,但它们的结果随着NN大小的增长而恶化,对于训练深层神经网络不切实际。如今,由于这种效率低下,这些方法在很大程度上已被放弃。我们提出了基于GA的演化和反向传播相结合的方法,用于演化CNN,如下所示。与上述方法类似,我们创建了N个染色体的初始种群,每个种群代表一个CNN的权重,然后通过适用性评估,交叉和突变对染色体进行进化,但有几个主要区别:(1)在交叉过程中,取而代之的是从两个亲本中的每个亲本中随机选择权重,从每个亲本中随机选择整个滤镜(这样可以确保完整复制一个有用的滤镜,而不是被破坏),(2)在变异过程中,通过执行标准的反向传播来修改权重,而不是随机变化;然后将一小部分权重随机设置为零(这些步骤可实现面向目标的演变,将一些权重置零会鼓励网络中的稀疏性并具有正则化效果)。我们将此方法称为EvoCNN。为了衡量我们方法的性能,我们在MNIST手写数字识别数据集上进行了一些实验。使用的标准CNN架构包含以下层:[输入大小28×28]-[具有128个大小为5×5的滤波器的卷积]-[最大池]-[具有256个大小为3×3的滤波器的卷积]-[最大池]-[大小为1000的完全连接层]-[大小为10的softmax层]。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号