𝔖 Bobbio Scriptorium
✦   LIBER   ✦

Reducing data dimensionality through optimizing neural network inputs

✍ Scribed by Shufeng Tan; Michael L. Mayrovouniotis


Book ID
102696495
Publisher
American Institute of Chemical Engineers
Year
1995
Tongue
English
Weight
916 KB
Volume
41
Category
Article
ISSN
0001-1541

No coin nor oath required. For personal study only.

✦ Synopsis


A neural network method for reducing data dimensionality based on the concept of input training, in which each input pattern is not fKed but adjusted along with internal network parameters to reproduce its corresponding output pattern, is presented. With input adjustment, a properly configured network can be trained to reproduce a given data set with minimum distortion; the trained network inputs provide reduced data.

A three-layer network with input training can perform all functions of a five-layer autoassociative network, essentially capturing nonlinear correlations among data. In addition, simultaneous training of a network and its inputs is shown to be significantly more eflcient in reducing data dimensionality than training an autoassociative network. The concept of input training is closely related to principal component analysis (PCA) and the principal curve method, which is a nonlinear extension of PCA.


πŸ“œ SIMILAR VOLUMES


Feedforward neural network's sensitivity
✍ Igor T. Podolak πŸ“‚ Article πŸ“… 1999 πŸ› Elsevier Science 🌐 English βš– 434 KB

Neural networks can be used to develop solutions to problems which are strictly symbolic. A question arises how to represent symbols in terms of number vectors understandable to neural networks. Data representation used should promote good generalization and reduce simulation uncertainty of the resu