Networks can be considered as approximation schemes. Multilayer networks of the perceptron type can approximate arbitrarily well continuous functions (Cybenko 1988(Cybenko , 1989;;Funahashi 1989;Stinchcombe and White 1989). We prove that networks derived from regularization theory and including Radi
Neural networks and the best trigomometric approximation
โ Scribed by Jianjun Wang; Zongben Xu
- Book ID
- 107347282
- Publisher
- Academy of Mathematics and Systems Science, Chinese Academy of Sciences
- Year
- 2011
- Tongue
- English
- Weight
- 442 KB
- Volume
- 24
- Category
- Article
- ISSN
- 1009-6124
No coin nor oath required. For personal study only.
๐ SIMILAR VOLUMES
Let D be a set with a probability measure +, +(D)=1, and let K be a compact subset of L q (D, +), where the infimum is taken over all g n of the form g n = n i=1 a i , i , with arbitrary , i # K and a i # R. It is shown that for f # conv(K \_ (&K )), under some mild restrictions, \ n ( f, K ) C q =
It is well known that artificial neural networks are universal approximators. But what about fuzzy neural networks? Only Buckl, ey and Hayashi [1] presented a theoretical result for these networks: They showed that there are fuzzy functions which cannot be approximated by a certain fuzzy neural netw