๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

Effect of data standardization on neural network training

โœ Scribed by M. Shanker; M.Y. Hu; M.S. Hung


Book ID
113322938
Publisher
Elsevier Science
Year
1996
Tongue
English
Weight
993 KB
Volume
24
Category
Article
ISSN
0305-0483

No coin nor oath required. For personal study only.


๐Ÿ“œ SIMILAR VOLUMES


Training neural networks with heterogene
โœ John A. Drakopoulos; Ahmad Abdulkader ๐Ÿ“‚ Article ๐Ÿ“… 2005 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 140 KB

Data pruning and ordered training are two methods and the results of a small theory that attempts to formalize neural network training with heterogeneous data. Data pruning is a simple process that attempts to remove noisy data. Ordered training is a more complex method that partitions the data into

Data qualification: Logic analysis appli
โœ Bryan P. Bergeron; Richard S. Shiffman; Ronald L. Rouse ๐Ÿ“‚ Article ๐Ÿ“… 1994 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 704 KB

Abatraet-For neural networks to develop good internal representations for pattern mapping, noise in the training set data must be controlled. Because of the many difficulties associated with manually validating training data, we have focused on using decision table techniques as a practical, domain-

Efficient Partition of Learning Data Set
โœ Igor V. Tetko; Alessandro E.P. Villa ๐Ÿ“‚ Article ๐Ÿ“… 1997 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 641 KB

This study investigates the emerging possibilities of combining unsupervised and supervised learning in neural network ensembles. Such strategy is used to get an efficient partition of a noisy input data set in order to focus the training of neural networks on the most complex and informative domain