𝔖 Bobbio Scriptorium
✦   LIBER   ✦

Limitations of multi-layer perceptron networks - steps towards genetic neural networks

✍ Scribed by H Mühlenbein


Book ID
107919266
Publisher
Elsevier Science
Year
1990
Tongue
English
Weight
614 KB
Volume
14
Category
Article
ISSN
0167-8191

No coin nor oath required. For personal study only.

✦ Synopsis


In this paper we investigate multi-layer perceptron networks in the task domain of Boolean functions. We demystify the multi-layer perceptron network by showing that it just divides the input space into regions constrained by hyperplanes. We use this information to construct minimal training sets. Despite using minimal training sets, the learning time of multi-layer perceptron networks with backpropagation scales exponentially for complex Boolean functions. But modular neural networks which consist of independently trained subnetworks scale very well. We conjecture that the next generation of neural networks will be genetic neural networks which evolve their structure. We confirm Minsky and Papert: " The future of neural networks is tied not to the search for some single, universal scheme to solve all problems at once, but to the evolution of a many-faceted technology of network design.


📜 SIMILAR VOLUMES


Daily pan evaporation modelling using mu
✍ Özgür Kişi 📂 Article 📅 2009 🏛 John Wiley and Sons 🌐 English ⚖ 527 KB

## Abstract This paper reports on investigations of the abilities of three different artificial neural network (ANN) techniques, multi‐layer perceptrons (MLP), radial basis neural networks (RBNN) and generalized regression neural networks (GRNN) to estimate daily pan evaporation. Different MLP mode