Estimating one of two normal means when their difference is bounded
β Scribed by Constance van Eeden; James V. Zidek
- Publisher
- Elsevier Science
- Year
- 2001
- Tongue
- English
- Weight
- 104 KB
- Volume
- 51
- Category
- Article
- ISSN
- 0167-7152
No coin nor oath required. For personal study only.
β¦ Synopsis
In this paper, we address the problem of estimating Γ1 when Yj βΌ ind N(Γj; 2 j ); j = 1; 2, are observed, the j are known and |Γ1 -Γ2|6c for a known constant c. Assuming the loss is squared error, we derive a generalized Bayes estimator which is admissible. It uses Y2 to achieve a uniformly smaller risk than that of the classical UMVU estimator, Y1. The proofs use a combination of Stein and Kubokawa methods.
π SIMILAR VOLUMES
In this paper we address the problem of estimating y 1 when Y i B ind NΓ°y i ; s 2 i Γ; i ΒΌ 1; 2; are observed and jy 1 Γ y 2 jpc for a known constant c: Clearly Y 2 contains information about y 1 : We show how the so-called weighted likelihood function may be used to generate a class of estimators t
For estimating under squared-error loss the mean of a p-variate normal distribution when this mean lies in a ball of radius m centered at the origin and the covariance matrix is equal to the identity matrix, it is shown that the Bayes estimator with respect to a uniformly distributed prior on the bo