An accelerated learning algorithm for multilayer perceptron networks
- 1 May 1994
- journal article
- Published by Institute of Electrical and Electronics Engineers (IEEE) in IEEE Transactions on Neural Networks
- Vol. 5 (3), 493-497
- https://doi.org/10.1109/72.286921
Abstract
An accelerated learning algorithm (ABP-adaptive back propagation) is proposed for the supervised training of multilayer perceptron networks. The learning algorithm is inspired from the principle of "forced dynamics" for the total error functional. The algorithm updates the weights in the direction of steepest descent, but with a learning rate a specific function of the error and of the error gradient norm. This specific form of this function is chosen such as to accelerate convergence. Furthermore, ABP introduces no additional "tuning" parameters found in variants of the backpropagation algorithm. Simulation results indicate a superior convergence speed for analog problems only, as compared to other competing methods, as well as reduced sensitivity to algorithm step size parameter variations.Keywords
This publication has 3 references indexed in Scilit:
- Experiments in finding neural network weightsPublished by Office of Scientific and Technical Information (OSTI) ,1990
- Parallel Distributed ProcessingPublished by MIT Press ,1986
- Tracking control of non-linear systems using sliding surfaces, with application to robot manipulators†International Journal of Control, 1983