We have a probabilistic experiment with two random variables and with marginal densities . The experiment is performed, such that and are realized. However, we are only able to observe . Our goal is then to estimate the unknown sample value of and from the observation . In particular, we want to choose the best estimate of , , given and .
- In this setting is called the a priori density of and is called the a posteriori density.
Example: Estimating Noisy Binary Signal
MMSE Estimation
Given an observation , find the best estimate of , i.e. such that we minimize
This is called the minimum mean square error (MMSE) estimator problem.
The key idea is that to minimize the expected square error, the best choice of is the conditional mean of given the observation:
- The conditional mean is the best estimate in the sense that it makes the variance of the estimation error as small as possible.
Derivation
Let’s choose some arbitrary estimator (any vector or value). Then:
Letting , we have
The only term that depends on is . Thus, the term is minimized when . So the optimal estimator is just
Example: MMSE Estimation
Conditional Mean for Gaussian
For and , the conditional pdf is also Gaussian with its mean given by
MMSE Estimation for Gaussian