Let D/R d be a compact set and let 8 be a uniformly bounded set of D ร R functions. For a given real-valued function f defined on D and a given natural number n, we are looking for a good uniform approximation to f of the form n i=1 a i , i , with , i # 8, a i # R. Two main cases are considered: (1)
Universal approximation in p-mean by neural networks
โ Scribed by Robert M. Burton; Herold G. Dehling
- Publisher
- Elsevier Science
- Year
- 1998
- Tongue
- English
- Weight
- 112 KB
- Volume
- 11
- Category
- Article
- ISSN
- 0893-6080
No coin nor oath required. For personal study only.
โฆ Synopsis
A feedforward neural net with d input neurons and with a single hidden layer of n neurons is given by
where a j , v j , w ji สฆ R. In this paper we study the approximation of arbitrary functions f: R d โ R by a neural net in an L p (m) norm for some finite measure m on R d . We prove that under natural moment conditions, a neural net with non-polynomial function can approximate any given function.
๐ SIMILAR VOLUMES
Fluid neural networks can be used as a theoretical framework for a wide range of complex systems as social insects. In this article we show that collective logical gates can be built in such a way that complex computation can be possible by means of the interplay between local interactions and the c
Constructive theorems of three-layer artificial neural networks with (1) trigonometric, (2) piecewise linear, and (3) sigmoidal hidden-layer units are proved in this paper. These networks approximate 2p-periodic pth-order Lebesgue-integrable functions (L p 2p ) on R m to R n for p ี 1 with L p 2p ยน
In this paper, we construct two types of feed-forward neural networks (FNNs) which can approximately interpolate, with arbitrary precision, any set of distinct data in the metric space. Firstly, for analytic activation function, an approximate interpolation FNN is constructed in the metric space, an
In this paper, we prove that any finite time trajectory of a given n-dimensional dynamical system can be approximately realized by the internal state of the output units of a continuous time recurrent neural network with n output units, some hidden units, and an appropriate initial condition. The es