๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

Neural accelerator for parallelization of back-propagation algorithm

โœ Scribed by Edoardo Franzi


Publisher
Elsevier Science
Year
1993
Weight
910 KB
Volume
38
Category
Article
ISSN
0165-6074

No coin nor oath required. For personal study only.


๐Ÿ“œ SIMILAR VOLUMES


Acceleration by prediction for error bac
โœ Arihiro Kanda; Satoshi Fujita; Tadashi Ae ๐Ÿ“‚ Article ๐Ÿ“… 1994 ๐Ÿ› John Wiley and Sons ๐ŸŒ English โš– 687 KB

## Abstract This paper proposes a speed improvement of the error back propagation algorithm, which is employed widely in the multilayered neural network, by introducing the prediction. The idea is to realize a larger acceleration by introducing the differential factor for the moment terms in the er

An adaptive training algorithm for back
โœ L.-W. Chan; F. Fallside ๐Ÿ“‚ Article ๐Ÿ“… 1987 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 628 KB

The effect of the coefficients used in the conventional back propagation algorithm on training connectionist models is discussed, using a vowel recognition task in speech processing as an example. Some weaknesses of the use of fixed coefficients are described and an adaptive algorithm using variable

Back-propagation neural network for long
โœ Tsong-Lin Lee ๐Ÿ“‚ Article ๐Ÿ“… 2004 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 261 KB

During the recent years, the availability of accurate ocean tide models has become increasingly important, as tides are the main contributor to disposal and movement of sediments, tracers and pollutants, and to a whole range of offshore applications in engineering, environmental observations, explor

Accelerating the convergence of the back
โœ T. P. Vogl; J. K. Mangis; A. K. Rigler; W. T. Zink; D. L. Alkon ๐Ÿ“‚ Article ๐Ÿ“… 1988 ๐Ÿ› Springer-Verlag ๐ŸŒ English โš– 621 KB

The utility of the back-propagation method in establishing suitable weights in a distributed adaptive network has been demonstrated repeatedly. Unfortunately, in many applications, the number of iterations required before convergence can be large. Modifications to the back-propagation algorithm desc