The paper demonstrates that the largest Lyapunov exponent /spl lambda/ of recurrent neural networks can be controlled efficiently by a stochastic gradient method. An essential core of the proposed method is a novel stochastic approximate formulation of the Lyapunov exponent /spl lambda/ as a function of the network parameters such as connection weights and thresholds of neural activation functions. By a gradient method, a direct calculation to minimize a square error (/spl lambda/-/spl lambda//sup obj/)/sup 2/, where /spl lambda//sup obj/ is a desired exponent value, needs gradient collection through time which are given by a recursive calculation from past to present values. The collection is computationally expensive and causes unstable control of the exponent for networks with chaotic dynamics because of chaotic instability. The stochastic formulation derived in the paper gives us an approximation of the gradient collection in a fashion without the recursive calculation. This approximation can realize not only a faster calculation of the gradients, where only O(N/sup 2/) run time is required while a direct calculation needs O(N/sup 5/T) run time for networks with N neurons and T evolution, but also stable control for chaotic dynamics. It is also shown by simulation studies that the approximation is a robust formulation for the network size and that proposed method can control the chaos dynamics in recurrent neural networks effectively.
展开▼