Subset-based training and pruning of sigmoid neural networks
โ Scribed by Guian Zhou; J. Si
- Book ID
- 104348918
- Publisher
- Elsevier Science
- Year
- 1999
- Tongue
- English
- Weight
- 170 KB
- Volume
- 12
- Category
- Article
- ISSN
- 0893-6080
No coin nor oath required. For personal study only.
โฆ Synopsis
In the present paper we develop two algorithms, subset-based training (SBT) and subset-based training and pruning (SBTP), using the fact that the Jacobian matrices in sigmoid network training problems are usually rank deficient. The weight vectors are divided into two parts during training, according to the Jacobian rank sizes. Both SBT and SBTP are trust-region methods. Compared with the standard Levenberg-Marquardt (LM) method, these two algorithms can achieve similar convergence properties as the LM but with fewer memory requirements. Furthermore the SBTP combines training and pruning of a network into one comprehensive procedure. The effectiveness of the two algorithms is evaluated using three examples. Comparisons are made with some existing algorithms. Some convergence properties of the two algorithms are given to qualitatively evaluate the performance of the algorithms.
๐ SIMILAR VOLUMES
We introduce a new method for proving explicit upper bounds on the VC dimension of general functional basis networks and prove as an application, for the first time, that the VC dimension of analog neural networks with the sigmoidal activation function \_( y)=1ร1+e & y is bounded by a quadratic poly