首页> 外文期刊>Engineering Applications of Artificial Intelligence >Learning with Type-2 Fuzzy activation functions to improve the performance of Deep Neural Networks
【24h】

Learning with Type-2 Fuzzy activation functions to improve the performance of Deep Neural Networks

机译:使用Type-2模糊激活函数进行学习以提高深度神经网络的性能

获取原文
获取原文并翻译 | 示例
           

摘要

In this study, we propose a novel Interval Type-2 (IT2) Fuzzy activation layer that is composed of Single input IT2 (SIT2) Fuzzy Rectifying Units (FRUs) to improve the learning performances of Deep Neural Networks (DNNs). The novel SIT2-FRU has tunable parameters that not only define the slopes of the positive and negative quadrants but also the characteristic of the input-output mapping of the activation function. The novel SIT2-FRU also alleviates vanishing gradient problem and has a fast convergence rate since it can push the mean activation to around zero by processing the inputs defined in the negative quadrant. Thus, SIT2-FRU gives the opportunity to the DNN to have a better learning behavior as it is capable to express linear or sophisticated input-output mapping by simply tuning the footprint of uncertainty of its IT2 fuzzy sets. In order to examine the performance of the SIT2-FRU, comparative experimental studies are performed on the MNIST, Quickdraw Pictionary and CIFAR-10 benchmark datasets. The proposed SIT2-FRU is compared with the state of the art activation functions which are the Rectified Linear Unit (ReLU), Parametric ReLU (PReLU) and Exponential Linear Unit (ELU). Comparative experimental results and analyses clearly show the enhancement in the learning performance of DNNs that include activation layer(s) composed of SIT2-FRUs. It is shown that the learning performance of the SIT2-FRU is robust to different parameter settings of the learning rates and mini batch sizes. Furthermore, the experimental results show that SIT2-FRU can result with a high performance with or without batch normalization layers unlike the other employed activation units. It is concluded that DNNs with SIT2-FRUs have a satisfactory generalization capability, a robust and high learning performance when compared to the ReLU, PReLU and ELU activation functions.
机译:在这项研究中,我们提出了一种新颖的间隔类型2(IT2)模糊激活层,该层由单输入IT2(SIT2)模糊整流单元(FRU)组成,以改善深度神经网络(DNN)的学习性能。新型SIT2-FRU具有可调参数,这些参数不仅定义了正象限和负象限的斜率,而且还定义了激活函数的输入-输出映射的特性。新型SIT2-FRU还可以缓解梯度消失的问题,并且收敛速度快,因为它可以通过处理负象限中定义的输入将平均激活推到零附近。因此,SIT2-FRU通过简单地调整其IT2模糊集的不确定性足迹即可表达线性或复杂的输入输出映射,从而为DNN提供了更好的学习行为。为了检查SIT2-FRU的性能,对MNIST,Quickdraw Pictionary和CIFAR-10基准数据集进行了对比实验研究。将拟议的SIT2-FRU与最新的激活函数(整流线性单元(ReLU),参数化ReLU(PReLU)和指数线性单元(ELU))进行比较。对比实验结果和分析清楚地表明,DNN的学习性能得到了增强,这些DNN包括由SIT2-FRU组成的激活层。结果表明,SIT2-FRU的学习性能对于学习率和最小批量大小的不同参数设置具有鲁棒性。此外,实验结果表明,与其他采用的激活单元不同,无论是否具有批处理归一化层,SIT2-FRU都能获得高性能。结论是,与ReLU,PReLU和ELU激活功能相比,具有SIT2-FRU的DNN具有令人满意的泛化能力,强大的学习性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号