In this paper, optimal encoding schemes for linear associative memories are derived for biased association under both the white-noise and colored-noise situations. Analysis and simulation results all show that the biased encodings thus derived are optimal and superior to existing models in their per
Time for retrieval in recurrent associative memories
โ Scribed by Alessandro Treves; Edmund T. Rolls; Martin Simmen
- Publisher
- Elsevier Science
- Year
- 1997
- Tongue
- English
- Weight
- 737 KB
- Volume
- 107
- Category
- Article
- ISSN
- 0167-2789
No coin nor oath required. For personal study only.
โฆ Synopsis
Experimental evidence shows that certain types of visual information processing, such as face recognition, may be extremely rapid: in a few tens of milliseconds a neuron may yield most of the information about a visual stimulus that can ever be extracted from the response of that neuron. These data might be taken to indicate that there is insufficient time for recurrent or feedback processing. However, a novel analytical method allows the analysis of the dynamics of an associative memory network of integrate-and-fire neurons, laterally connected through realistically modelled synapses. The analysis, supported by preliminary simulations, indicates that the network may retrieve information from memory over a time determined mainly by excitatory synaptic conductance time constants, that is in times in the order of tens of milliseconds. The results thus show that the dynamics of recurrent processing is sufficiently rapid for it to contribute to processing within the short times observed in neurophysiological experiments.
๐ SIMILAR VOLUMES
In this paper a global design method for associative memories using discrete-time cellular neural networks (DTCNNs) is presented. The proposed synthesis technique enables to realize associative memories with several advantageous features. First of all, grey-level as well as bipolar images can be sto
The multipletraining concept first appliedto Bidirectional Associative Memory trainingis appliedto the back-propagation (BP) algorithm for use in associative memories. This new algorithm. which assigns different weights to the various pairsin the energyfunction, is calledmultiple training back-propa