𝔖 Bobbio Scriptorium
✦   LIBER   ✦

An adaptive conjugate gradient learning algorithm for efficient training of neural networks

✍ Scribed by H. Adeli; S.L. Hung


Book ID
107884655
Publisher
Elsevier Science
Year
1994
Tongue
English
Weight
1003 KB
Volume
62
Category
Article
ISSN
0096-3003

No coin nor oath required. For personal study only.


πŸ“œ SIMILAR VOLUMES


A parallel algorithm for gradient traini
✍ ZdenΔ›k HanzΓ‘lek πŸ“‚ Article πŸ“… 1998 πŸ› Elsevier Science 🌐 English βš– 742 KB

This paper presents a message-passing architecture simulating multilayer neural networks, adjusting its weights for each pair, consisting of an input vector and a desired output vector. First, the multilayer neural network is defined, and the difficulties arising from parallel implementation are cla

Adaptive stepsize algorithms for on-line
✍ G.D. Magoulas; V.P. Plagianakos; M.N. Vrahatis πŸ“‚ Article πŸ“… 2001 πŸ› Elsevier Science 🌐 English βš– 312 KB

In this paper a method for adapting the stepsize in on-line network training is presented. The proposed technique derives from the stochastic gradient descent proposed by Almeida et al. [On-line Learning in Neural Networks, 111-134, Cambridge University Press, 1998]. The new aspect of our approach c

Efficient Partition of Learning Data Set
✍ Igor V. Tetko; Alessandro E.P. Villa πŸ“‚ Article πŸ“… 1997 πŸ› Elsevier Science 🌐 English βš– 641 KB

This study investigates the emerging possibilities of combining unsupervised and supervised learning in neural network ensembles. Such strategy is used to get an efficient partition of a noisy input data set in order to focus the training of neural networks on the most complex and informative domain