๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

An information theoretic approach for combining neural network process models

โœ Scribed by D.V. Sridhar; E.B. Bartlett; R.C. Seagrave


Publisher
Elsevier Science
Year
1999
Tongue
English
Weight
481 KB
Volume
12
Category
Article
ISSN
0893-6080

No coin nor oath required. For personal study only.

โœฆ Synopsis


Typically neural network modelers in chemical engineering focus on identifying and using a single, hopefully optimal, neural network model. Using a single optimal model implicitly assumes that one neural network model can extract all the information available in a given data set and that the other candidate models are redundant. In general, there is no assurance that any individual model has extracted all relevant information from the data set. Recently, Wolpert (Neural Networks, 5(2), 241 (1992)) proposed the idea of stacked generalization to combine multiple models. Sridhar, Seagrave and Barlett (AIChE J., 42, 2529 (1996)) implemented the stacked generalization for neural network models by integrating multiple neural networks into an architecture known as stacked neural networks (SNNs). SNNs consist of a combination of the candidate neural networks and were shown to provide improved modeling of chemical processes. However, in Sridhar's work SNNs were limited to using a linear combination of artificial neural networks. While a linear combination is simple and easy to use, it can utilize only those model outputs that have a high linear correlation to the output. Models that are useful in a nonlinear sense are wasted if a linear combination is used. In this work we propose an information theoretic stacking (ITS) algorithm for combining neural network models. The ITS algorithm identifies and combines useful models regardless of the nature of their relationship to the actual output. The power of the ITS algorithm is demonstrated through three examples including application to a dynamic process modeling problem. The results obtained demonstrate that the SNNs developed using the ITS algorithm can achieve highly improved performance as compared to selecting and using a single hopefully optimal network or using SNNs based on a linear combination of neural networks.


๐Ÿ“œ SIMILAR VOLUMES


An efficient neural network approach for
โœ M. S. Alam; A. Kranti; G. A. Armstrong ๐Ÿ“‚ Article ๐Ÿ“… 2009 ๐Ÿ› John Wiley and Sons ๐ŸŒ English โš– 401 KB

The present paper demonstrates the suitability of artificial neural network (ANN) for modelling of a FinFET in nano-circuit simulation. The FinFET used in this work is designed using careful engineering of source-drain extension, which simultaneously improves maximum frequency of oscillation f max b

An information theoretic approach for im
โœ Alfonso Montuori; Giovanni Raimondo; Eros Pasero ๐Ÿ“‚ Article ๐Ÿ“… 2008 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 778 KB

We present the results of an information theory-based approach to select an optimal subset of features for the prediction of protein model quality. The optimal subset of features was calculated by means of a backward selection procedure. The performances of a probabilistic classifier modeled by mean

Combining dynamic neural networks and im
โœ Shang Ming Zhou ๐Ÿ“‚ Article ๐Ÿ“… 1999 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 165 KB

This paper describes how to build a quality prediction model for complex industrial production processes using dynamic neural networks. It is known that it is difficult to analyze the mechanisms of many complex industrial production processes and build models by employing classical methods. In this