Learning in small connectionist networks does not generalize to large networks
โ Scribed by D. Feldman-Stewart; D. J. K. Mewhort
- Book ID
- 104768178
- Publisher
- Guilford Publishing Inc
- Year
- 1994
- Tongue
- English
- Weight
- 592 KB
- Volume
- 56
- Category
- Article
- ISSN
- 0340-0727
No coin nor oath required. For personal study only.
โฆ Synopsis
A number of recent models of human information processing have been based on connectionist architectures. Such models are designed to illustrate specific psychological principles and are usually implemented in small networks. The assumption implicit in the work is that principles illustrated in small networks can be applied easily to brain-size networks by scaling up as required. To consider the scaling question, we used a back-propagation algorithm to compare learning in both large and small networks and found that learning depended on the size of the network. In small networks, increasing I] (the rate-oflearning parameter) beyond 1 increased the rate of learning; in large networks, the same manipulation reduced the rate of learning. The example illustrates the difficulty of generalizing across network size and calls into question the assumption that principles illustrated in small networks can be applied to the brain by expansion of the network.
๐ SIMILAR VOLUMES
Multimedia Interactive Protocols and Systems (MIPS) is the brand new name of a workshop that has been successfully held for the ?rst time in 2002 in Coimbra, as the ?rst joint edition of two well established series of workshops: Interactive Distributed Multimedia Systems (IDMS) and Protocols for Mul