Neural networks (NN) are famous for their advantageous flexibility for problems when there is insufficient knowledge to set up a proper model. On the other hand, this flexibility can cause overfitting and can hamper the generalization of neural networks. Many approaches to regularizing NN have been
Feed-forward neural networks for secondary structure prediction
โ Scribed by T.W. Barlow
- Publisher
- Elsevier Science
- Year
- 1995
- Tongue
- English
- Weight
- 744 KB
- Volume
- 13
- Category
- Article
- ISSN
- 0263-7855
No coin nor oath required. For personal study only.
๐ SIMILAR VOLUMES
In this paper we present a new algorithm, which is orders of magnitude faster than the delta rule, for training feed-forward neural networks. It provides a substantial improvement over the method of Scalero and Tepedelenlioglu (IEEE Trans. Signal Process. 40(1) (1992)) in both training time and nume
Fast training of feed-forward neural networks became increasingly important as the neural network field moves toward maturity. This paper begins with a review of various criteria proposed for training feed-forward neural networks, which include the frequently used quadratic error criterion, the rela