𝔖 Bobbio Scriptorium
✦   LIBER   ✦

Cubic approximation neural network for multivariate functions

✍ Scribed by Doron Stein; Arie Feuer


Publisher
Elsevier Science
Year
1998
Tongue
English
Weight
374 KB
Volume
11
Category
Article
ISSN
0893-6080

No coin nor oath required. For personal study only.

✦ Synopsis


This paper introduces a novel neural network architecture-cubic approximation neural network (CANN), capable of local approximation of multivariate functions. It is particularly simple in concept and in structure. Its simplicity enables a quantitative evaluation of its approximation capabilities, namely, for a desired error bound the size of the needed network can be calculated. In addition, if a training session is used, a thorough analysis of the learning process performance is performed. The trade-off between the rate of learning and the steady-state performance is clearly demonstrated. On the other hand, this approach suffers from the problem common to all local approximation networks-the number of neurons grows exponentially with the dimension of the input vector.


πŸ“œ SIMILAR VOLUMES


Multivariate sigmoidal neural network ap
✍ George A. Anastassiou πŸ“‚ Article πŸ“… 2011 πŸ› Elsevier Science 🌐 English βš– 268 KB

Here we study the multivariate quantitative constructive approximation of real and complex valued continuous multivariate functions on a box or RN, N∈N, by the multivariate quasi-interpolation sigmoidal neural network operators. The "right" operators for our goal are fully and precisely described. T

Neural network approximation of continuo
✍ M.B. Stinchcombe πŸ“‚ Article πŸ“… 1999 πŸ› Elsevier Science 🌐 English βš– 164 KB

This article characterizes the set of activation functions, bounded or unbounded, that allow feedforward network approximation of the continuous functions on the classic two-point compactification of R 1 . The characterization fails when the set of targets are continuous functions on the classic com

Orthogonal considerations in the design
✍ B. Francois πŸ“‚ Article πŸ“… 1996 πŸ› Elsevier Science 🌐 English βš– 604 KB

Two problems occur in the design of feedforward neural networks: the choice of the optimal architecture and the initialization. Generally, input and output data of a system (or a function) are measured and recorded. Then, experimenters wish to design a neural network to map exactly these output valu

Variable selection for neural networks i
✍ FrΓ©dΓ©ric Despagne; DΓ©sirΓ©-Luc Massart πŸ“‚ Article πŸ“… 1998 πŸ› Elsevier Science 🌐 English βš– 397 KB

The problem of variable selection for neural network modeling is discussed in this paper. Two methods that gave the best results in a previous comparative study are presented. One of these methods is a modified version of the Hinton diagrams, the other method is based on saliency estimation and is p