Three information-theoretical methods to estimate a continuous univariate distribution are proposed for estimation when the distribution type is uncertain, when data are scarce, or when extremes are important. The first is a new version of Jaynes' MaxEnt Method. The second, minimizing Shannon's inf
Estimating random variables from random sparse observations
✍ Scribed by Andrea, Montanari
- Publisher
- John Wiley and Sons
- Year
- 2008
- Tongue
- English
- Weight
- 238 KB
- Volume
- 19
- Category
- Article
- ISSN
- 1124-318X
- DOI
- 10.1002/ett.1289
No coin nor oath required. For personal study only.
✦ Synopsis
Abstract
Let X~1~, … , X~n~, be a collection of iid discrete random variables, and Y~1~, … , Y~m~, a set of noisy observations of such variables. Assume each observation Y~a~, to be a random function of a random subset of the X~i~,s, and consider the conditional distribution of X~i~, given the observations, namely µ~i~,(x~i~,) ≡ ${\cal P}${X~i~, = x~i~,|Y} (a posteriori probability).
We establish a general decoupling principle among the X~i~,s, as well as a relation between the distribution of µ~i~, and the fixed points of the associated density evolution operator. These results hold asymptotically in the large system limit, provided the average number of variables an observation depends on is bounded. We discuss the relevance of our result to a number of applications, ranging from sparse graph codes and multi‐user detection, to group testing. Copyright © 2008 John Wiley & Sons, Ltd.
📜 SIMILAR VOLUMES
Formulas for covariance matrix between a random vector and its ordered components are derived for different distributions including multivariate normal, t, and F. The present formulas and related results obtained here lead to some known results in the literature as special cases.