𝔖 Bobbio Scriptorium
✦   LIBER   ✦

Evaluation of nonlinear optimization methods for the learning algorithm of artificial neural networks

✍ Scribed by Hideyuki Takagi; Shigeo Sakaue; Hayato Togawa


Publisher
John Wiley and Sons
Year
1992
Tongue
English
Weight
980 KB
Volume
23
Category
Article
ISSN
0882-1666

No coin nor oath required. For personal study only.

✦ Synopsis


Abstract

This paper describes the implementation of nonlinear optimization methods into the learning of neural networks (NN) and the speed efficiency of four proposed improvements into the backpropagation algorithm. The problems of the backpropagation learning method are pointed out first, and the efficiency of implementing a nonlinear optimization method as a solution to this problems is described.

Two nonlinear optimization methods are selected after inspecting several nonlinear methods from the viewpoint of NN learning to avoid the problem of the backpropagation algorithm. These are the linear search method by Davies, Swann, and Campey (DSC), and the conjugate gradient method by Fletcher and Reeves.

The NN learning algorithms with these standard methods being implemented are formulated. Moreover, the following four improvements of the nonlinear optimization methods are proposed to shorten the NN learning time: (a) fast forward calculation in linear search by consuming a larger amount of memories; (b) avoiding the trap to local minimum point in an early stage of linear search; (c) applying a linear search method suitable for parallel processing; and (d) switching the gradient direction using the conjugate gradient method. The evaluation results have shown that all methods described here are effective in shortening the learning time.


πŸ“œ SIMILAR VOLUMES


On a class of efficient learning algorit
✍ Frank BΓ€rmann; Friedrich Biegler-KΓΆnig πŸ“‚ Article πŸ“… 1992 πŸ› Elsevier Science 🌐 English βš– 457 KB

The ability of a neural network with one hidden layer to accurately learn a specified learning set increases with the number of nodes in the hidden layer; if a network has exactly the same number of internal nodes as the number of examples to be learnt, it is theoretically able to learn these exampl