An explanation of reasoning neural networks
β Scribed by R.R. Tsaih
- Publisher
- Elsevier Science
- Year
- 1998
- Tongue
- English
- Weight
- 693 KB
- Volume
- 28
- Category
- Article
- ISSN
- 0895-7177
No coin nor oath required. For personal study only.
β¦ Synopsis
tioning Neural Networks (RN) adopts the layered feedforward network structure, and its learning algorithm belongs to the weight-and-structure-change category of learning algorithm. In this paper, we firstly explain that, in the layered feedforward network, the essential characteristic of the mapping between two consecutive layers is the level-adjacent mapping, in which level-adjacent patterns in the previous-layer spsce are mapped to similar patterns in the latter-layer space. Then, we explain how RN's learning algorithm handles the undesired predicaments associated with the back propagation learning algorithm.
π SIMILAR VOLUMES
This paper give a methodology, PROLOG code, as well as an example of an explanation facility EN applicable to most neural networks. It involves How?, Why? and TRACE facilities, and is based on a general explanation degree calculation in a multilayer neural network, as well as on input node character
The Boltzmann machine is a probabilistic neural network describing the associative dependency of variables. It yields a probability distribution, which is a special case of the distribution generated by probabilistic inference networks. Hence both types of networks can be combined allowing to integr