๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

Learning and Approximation Capabilities of Adaptive Spline Activation Function Neural Networks

โœ Scribed by Lorenzo Vecci; Francesco Piazza; Aurelio Uncini


Publisher
Elsevier Science
Year
1998
Tongue
English
Weight
297 KB
Volume
11
Category
Article
ISSN
0893-6080

No coin nor oath required. For personal study only.

โœฆ Synopsis


In this paper, we study the theoretical properties of a new kind of artificial neural network, which is able to adapt its activation functions by varying the control points of a Catmull-Rom cubic spline. Most of all, we are interested in generalization capability, and we can show that our architecture presents several advantages. First of all, it can be seen as a sub-optimal realization of the additive spline based model obtained by the reguralization theory. Besides, simulations confirm that the special learning mechanism allows to use in a very effective way the network's free parameters, keeping their total number at lower values than in networks with sigmoidal activation functions. Other notable properties are a shorter training time and a reduced hardware complexity, due to the surplus in the number of neurons.


๐Ÿ“œ SIMILAR VOLUMES


Neural network approximation of continuo
โœ M.B. Stinchcombe ๐Ÿ“‚ Article ๐Ÿ“… 1999 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 164 KB

This article characterizes the set of activation functions, bounded or unbounded, that allow feedforward network approximation of the continuous functions on the classic two-point compactification of R 1 . The characterization fails when the set of targets are continuous functions on the classic com

Properties and performance of orthogonal
โœ Chieh F. Sher; Ching-Shiow Tseng; Chen-San Chen ๐Ÿ“‚ Article ๐Ÿ“… 2001 ๐Ÿ› John Wiley and Sons ๐ŸŒ English โš– 216 KB

Backpropagation neural network has been applied successfully to solving uncertain problems in many fields. However, unsolved drawbacks still exist such as the problems of local minimum, slow convergence speed, and the determination of initial weights and the number of processing elements. In this pa

Approximation of a function and its deri
โœ Pierre Cardaliaguet; Guillaume Euvrard ๐Ÿ“‚ Article ๐Ÿ“… 1992 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 998 KB

This paper deals with the approximation of both a function and its derivative by feedforward neural networks. We propose an explicit formula of approximation which is noise resistant and can be easily modified with the patterns. We apply these results to approach a function defined implicitly, which