首页> 外文会议>International Conference for Convergence in Technology >Activation Functions: Experimentation and Comparison
【24h】

Activation Functions: Experimentation and Comparison

机译:激活功能:实验和比较

获取原文
获取外文期刊封面目录资料

摘要

Activation functions are mathematical functions that are used to activate the neurons of an Artificial Neural Network. Non-linear activation functions mainly help a neural network to converge faster while learning and finding patterns in the complex input data. A neural network learns by updating the weights, which is done using the Back Propagation algorithm, which uses first-order derivatives of the activation functions to calculate the gradient descent. This paper tests various existing and proposed activation functions against Minst and Cifar10 datasets for image classification using a shallow Convolutional Neural Network (CNN) Architecture. Based on the results, some of the proposed activation functions: SMod = $x$ *tanh ($x$), the Absolute/Mod Function, a scaled version of Swish, and some other activation functions, are found to be promising. Some of these are then tested against Deeper Neural Networks for various datasets, and it is observed that the average error rate is improved by 2.77. Along with that, suggestions on which activation functions to be used for shallow and deep layers of a Deep Neural Network are provided, resulting in better performance.
机译:激活功能是用于激活人工神经网络的神经元的数学函数。非线性激活功能主要帮助神经网络在学习和查找复杂输入数据中的模式时收敛得更快。神经网络通过更新使用后传播算法进行的权重来学习,该权重将使用激活函数的一阶导数来计算梯度下降。本文使用浅卷积神经网络(CNN)架构测试针对Minst和CIFAR10数据集的各种现有和提出的激活功能,用于图像分类。基于结果,一些提议的激活功能:Smod = $ x $ * Tanh( $ x $ ),发现绝对/ mod函数,速度版本的速度和一些其他激活功能,是有前途的。然后将其中一些用于针对各种数据集的更深的神经网络测试,并且观察到平均误差率为2.77。除此之外,提供了用于深度神经网络的浅层和深层的激活功能的建议,导致性能更好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号