Approximation by neural networks is not continuous
✍ Scribed by Paul C. Kainen; Věra Kůrková; Andrew Vogt
- Book ID
- 114297201
- Publisher
- Elsevier Science
- Year
- 1999
- Tongue
- English
- Weight
- 188 KB
- Volume
- 29
- Category
- Article
- ISSN
- 0925-2312
No coin nor oath required. For personal study only.
📜 SIMILAR VOLUMES
In this paper, we prove that any finite time trajectory of a given n-dimensional dynamical system can be approximately realized by the internal state of the output units of a continuous time recurrent neural network with n output units, some hidden units, and an appropriate initial condition. The es
Let D/R d be a compact set and let 8 be a uniformly bounded set of D Ä R functions. For a given real-valued function f defined on D and a given natural number n, we are looking for a good uniform approximation to f of the form n i=1 a i , i , with , i # 8, a i # R. Two main cases are considered: (1)
In this paper, we prove that any continuous mapping can be approximately realized by Rumelhart-Hinton-Williams' multilayer neural networks with at least one hidden layer whose output functions are sigmoid functions. The starting point of the proof for the one hidden layer case is an integral formula