Reducing data dimensionality through optimizing neural network inputs
β Scribed by Shufeng Tan; Michael L. Mayrovouniotis
- Book ID
- 102696495
- Publisher
- American Institute of Chemical Engineers
- Year
- 1995
- Tongue
- English
- Weight
- 916 KB
- Volume
- 41
- Category
- Article
- ISSN
- 0001-1541
No coin nor oath required. For personal study only.
β¦ Synopsis
A neural network method for reducing data dimensionality based on the concept of input training, in which each input pattern is not fKed but adjusted along with internal network parameters to reproduce its corresponding output pattern, is presented. With input adjustment, a properly configured network can be trained to reproduce a given data set with minimum distortion; the trained network inputs provide reduced data.
A three-layer network with input training can perform all functions of a five-layer autoassociative network, essentially capturing nonlinear correlations among data. In addition, simultaneous training of a network and its inputs is shown to be significantly more eflcient in reducing data dimensionality than training an autoassociative network. The concept of input training is closely related to principal component analysis (PCA) and the principal curve method, which is a nonlinear extension of PCA.
π SIMILAR VOLUMES
Neural networks can be used to develop solutions to problems which are strictly symbolic. A question arises how to represent symbols in terms of number vectors understandable to neural networks. Data representation used should promote good generalization and reduce simulation uncertainty of the resu