【24h】

Fast Activation Function Approach for Deep Learning Based Online Anomaly Intrusion Detection

机译:基于深度学习的在线异常入侵检测快速激活函数方法

获取原文

摘要

The piecewise-linear activation functions such as ReLU become the catalyst that revolutionizes the training of the deep neural networks. Common nonlinear activation functions used in neural networks such as the tanh and the sigmoid activation functions su?er from saturation during training. The saturation behavior causes the problem of vanishing stochastic gradient decent. We propose a fast activation function, namely the Adaptive Linear Function (ALF) to increase the convergence speed and accuracy of the deep leaning structure for real-time applications. The ALF reduces the saturation effects caused by the soft activation functions and the vanishing gradient caused by the negative values of the ReLU. We evaluate the training method for an online anomaly intrusion detection system using Deep Belief Network (DBN) and simulating four bench mark datasets. The activation function increases the convergence speed of the DBN, with the entire training time reduced 80% compared to the sigmoid, ReLU, and tanh activation functions. The method achieves an accuracy rate of 98.59% on the total 10% KDDCUP'99 test dataset, 96.2% on the NSL-KDD dataset, 98.4% on the Kyoto dataset, and 96.57% on the CSIC HTTP dataset. The proposed activation function outperformed the results obtained when any of the three activation functions-sigmoid, ReLu, or tanh- was used on the test stream of the four datasets. Furthermore, the DBN structure outperforms state-of-the-art networks such as the Stacked Sparse AutoEncoder Based Extreme Learning Machine (SSAELM) in both accuracy and convergence speed.
机译:分段线性激活函数(例如ReLU)成为推动深层神经网络训练革命的催化剂。神经网络中常用的非线性激活函数(例如tanh和S型激活函数)在训练过程中会饱和。饱和行为会导致随机梯度下降消失的问题。我们提出了一种快速激活函数,即自适应线性函数(ALF),以提高实时应用的深度倾斜结构的收敛速度和准确性。 ALF减少了由软激活功能引起的饱和效应以及由ReLU的负值引起的消失梯度。我们评估了使用深度信任网络(DBN)并模拟四个基准数据集的在线异常入侵检测系统的训练方法。与S型,ReLU和tanh激活功能相比,激活功能提高了DBN的收敛速度,整个训练时间减少了80%。该方法在总共10%的KDDCUP'99测试数据集上达到98.59%的准确率,在NSL-KDD数据集上达到96.2%,在Kyoto数据集上达到98.4%,在CSIC HTTP数据集上达到96.57%。当在四个数据集的测试流上使用三个激活函数(Sigmoid,ReLu或tanh)中的任何一个时,建议的激活函数均优于获得的结果。此外,DBN结构在准确性和收敛速度方面都优于诸如堆叠式基于稀疏自动编码器的极限学习机(SSAELM)之类的最新网络。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号