首页> 外文会议>USM-JIRCAS Joint International Symposium >ADVANCED NEURAL NETWORK LEARNING APPLIED TO MODELING OF PULPING OF SUGAR MAPLE
【24h】

ADVANCED NEURAL NETWORK LEARNING APPLIED TO MODELING OF PULPING OF SUGAR MAPLE

机译:高级神经网络学习应用于糖枫木制浆的建模

获取原文

摘要

This paper reports work done to improve the modeling of complex processes when only small experimental datum sets are available. Neural networks are used to capture the nonlinear underlying phenomena contained in the data set and to partly eliminate the burden of having to specify completely the structure of the model. Two different types of neural networks were used for the application of Pulping of Sugar Maple problem. A three layer feed forward neural networks, using the Preconditioned Conjugate Gradient (PCG) methods were used in this investigation. Preconditioning is a method to improve convergence by lowering the condition number and increasing the eigenvalues clustering. The idea is to solve the modified problem M~(-1) Ax = M~(-1) b where Mis a positive-definite preconditioner that is closely related to A. We mainly focused on Preconditioned Conjugate Gradient- based training methods which originated from optimization theory, namely Preconditioned Conjugate Gradient with Fletcher-Reeves Update (PCGF), Preconditioned Conjugate Gradient with Polak-Ribiere Update (PCGP) and Preconditioned Conjugate Gradient with Powell-Beale Restarts (PCGB). The computational experiments revealed that the PCG methods, through its preconditioner matrix, reduced drastically the mean squared error during training. The CG and PCG methods have a much faster convergence rate than the BP; since it uses second order information to calculate the new direction (Rao, 1978). All the trials for the PCG methods converged to the required solution indicating that the choice of initial weights is appropriate. Even though it does not guarantee convergence to the required mean squared error, it reduces the possibility of getting stuck in local minima. The PCG algorithms have proved its capability in the generalization aspect, at the same time preserving the efficient convergence of the algorithm. The variation in the network outputs is explained very well by the corresponding targets. The results of the simulations suggest that, using preconditioning techniques, the condition numbers of a matrix will be improved. If the eigenvalues of the input matrix have been clustered, we can iteratively solve the problem more quickly. The behavior of the PCG methods in the simulations proved to be robust against phenomenon such as oscillations due to large step size.
机译:本文报告了在仅提供小型实验基准组时改善复杂过程的建模。神经网络用于捕获数据集中包含的非线性底层现象,并且部分地消除了必须完全指定模型结构的负担。两种不同类型的神经网络用于甘蔗问题的应用。在本研究中使用了三层馈电前向神经网络,使用预处理的共轭梯度(PCG)方法。预处理是通过降低条件数和增加特征值聚类来改善收敛的方法。这个想法是解决修改的问题m〜(-1)x = m〜(-1)b,其中MIS是与A密切相关的正定的预处理器。我们主要集中在前提的共轭梯度的培训方法上起源于从优化理论,附属于闪光灯-Reeves更新(PCGF)的预处理缀合物梯度,使用POLAK-RIBIERE更新(PCGP)和PATLEL-Beale重启(PCGB)的预处理共轭梯度和预先处理的共轭梯度(PCGB)。计算实验表明,通过其预处理器矩阵,PCG方法彻底减少了训练期间平均平均误差。 CG和PCG方法具有比BP更快的收敛速度;由于它使用二阶信息来计算新方向(Rao,1978)。 PCG方法的所有试验都会融合到所需的解决方案,表明初始重量的选择是合适的。尽管它不保证收敛到所需的平均方形错误,但它降低了在局部最小值中陷入困境的可能性。 PCG算法已经证明了其在泛化方面的能力,同时保留了算法的有效收敛性。通过相应的目标解释网络输出的变化。仿真结果表明,使用预处理技术,将提高矩阵的条件数量。如果输入矩阵的特征值已经聚集了,我们可以迭代地解决问题。在模拟中的PCG方法的行为被证明是对诸如跨越阶梯尺寸的振荡等现象的稳健。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号