𝔖 Bobbio Scriptorium
✦   LIBER   ✦

Construction of confidence intervals for neural networks based on least squares estimation

✍ Scribed by I. Rivals; L. Personnaz


Publisher
Elsevier Science
Year
2000
Tongue
English
Weight
424 KB
Volume
13
Category
Article
ISSN
0893-6080

No coin nor oath required. For personal study only.

✦ Synopsis


We present the theoretical results about the construction of confidence intervals for a nonlinear regression based on least squares estimation and using the linear Taylor expansion of the nonlinear model output. We stress the assumptions on which these results are based, in order to derive an appropriate methodology for neural black-box modeling; the latter is then analyzed and illustrated on simulated and real processes. We show that the linear Taylor expansion of a nonlinear model output also gives a tool to detect the possible ill-conditioning of neural network candidates, and to estimate their performance. Finally, we show that the least squares and linear Taylor expansion based approach compares favorably with other analytic approaches, and that it is an efficient and economic alternative to the nonanalytic and computationally intensive bootstrap methods.


πŸ“œ SIMILAR VOLUMES


Comments on construction of confidence i
✍ Tomas Morsing; Claes Ekman πŸ“‚ Article πŸ“… 1998 πŸ› John Wiley and Sons 🌐 English βš– 63 KB πŸ‘ 1 views

Faber and Kowalski recently proposed a method to calculate confidence regions for the regression coefficients in a linear model when partial least squares (PLS) has been used as an estimation method (J. Chemometrics, 11, 181 (1997)). In this short communication we show that the proposed confidence r

A learning algorithm for multilayered ne
✍ Friedrich Biegler-KΓΆnig; Frank BΓ€rmann πŸ“‚ Article πŸ“… 1993 πŸ› Elsevier Science 🌐 English βš– 365 KB

An algorithm ./or the training of mtdtilayered neural networks solely based on linear algebraic methods is presented. Its convergence speed up to a certain limit t~flearning accura~3' is orders o./magnitude better than that of the classical back propagation. Furthermore. its learning aptitude increa