Estimating parameters of a multiple autoregressive model by the modified maximum likelihood method
✍ Scribed by Özlem Türker Bayrak; Ayşen D. Akkaya
- Publisher
- Elsevier Science
- Year
- 2010
- Tongue
- English
- Weight
- 586 KB
- Volume
- 233
- Category
- Article
- ISSN
- 0377-0427
No coin nor oath required. For personal study only.
✦ Synopsis
We consider a multiple autoregressive model with non-normal error distributions, the latter being more prevalent in practice than the usually assumed normal distribution. Since the maximum likelihood equations have convergence problems (Puthenpura and Sinha, 1986) [11], we work out modified maximum likelihood equations by expressing the maximum likelihood equations in terms of ordered residuals and linearizing intractable nonlinear functions (Tiku and Suresh, 1992) [8]. The solutions, called modified maximum estimators, are explicit functions of sample observations and therefore easy to compute. They are under some very general regularity conditions asymptotically unbiased and efficient (Vaughan and Tiku, 2000) [4]. We show that for small sample sizes, they have negligible bias and are considerably more efficient than the traditional least squares estimators. We show that our estimators are robust to plausible deviations from an assumed distribution and are therefore enormously advantageous as compared to the least squares estimators. We give a real life example.
📜 SIMILAR VOLUMES
When experiments are conducted there is always a chance of the occurrence of large measurement errors (outliers). Common identification methods like generalized least squares, maximum likelihood etc. may not converge in these situations due to the presence of oufliers. Here we present a method for t
The program estimates the optimal gene-frequencies from observed values of phenotypes by the method of maximumlikelihood. The program does not make assumptions about a specific blood-group-system. This information has to be supplied by input cards. A straightforward iteration process is used for ob