This paper deals with the problem of Stein-rule prediction in a general linear model. Our study extends the work of by assuming that the covariance matrix of the model's disturbances is unknown. Also, predictions are based on a composite target function that incorporates allowance for the simultane
Improved Multivariate Prediction under a General Linear Model
โ Scribed by C.A. Gotway; N. Cressie
- Publisher
- Elsevier Science
- Year
- 1993
- Tongue
- English
- Weight
- 554 KB
- Volume
- 45
- Category
- Article
- ISSN
- 0047-259X
No coin nor oath required. For personal study only.
โฆ Synopsis
Assuming a general linear model with known covariance matrix, several linear and nonlinear predictors are presented and their properties are discussed. In the context of simultaneous multiple prediction, a total sum of squared errors is suggested as a loss function for comparing predictors. Based on a fundamental relationship between prediction and estimation, a very general class of predictors is developed from which predictors with uniformly smaller risk than that of the classical best linear unbiased (i.e., universal kriging) predictor can be constructed. r. 199.3 Academic Press, Inc.
๐ SIMILAR VOLUMES
For multivariate nonlinear regression and multivariate generalized linear regression models, with repeated measurements and possible missing values, we derive the asymptotic normality of a general estimating equations estimator of the regression matrix. We also provide consistent estimators of the c
A matrix formulation of the Kruskal-Wallis analysis of variance is presented. This formulation illustrates the paralle1 nature of the parametric general linear model and the Kruskal-Wallis model. Using the matrix formulation, it is shown that the Kruskai-Wailis method can be implemented on a digital