We have a probabilistic experiment with two random variables and with marginal densities . The experiment is performed, such that and are realized. However, we are only able to observe . Our goal is then to estimate the unknown sample value of and from the observation . In particular, we want to choose the best estimate of , , given and .

  • In this setting is called the a priori density of and is called the a posteriori density.

Example: Estimating Noisy Binary Signal

MMSE Estimation

Given an observation , find the best estimate of , i.e. such that we minimize

This is called the minimum mean square error (MMSE) estimator problem.

The key idea is that to minimize the expected square error, the best choice of is the conditional mean of given the observation:

  • The conditional mean is the best estimate in the sense that it makes the variance of the estimation error as small as possible.

Derivation

Let’s choose some arbitrary estimator (any vector or value). Then:

Letting , we have

The only term that depends on is . Thus, the term is minimized when . So the optimal estimator is just

Example: MMSE Estimation

Conditional Mean for Gaussian

For and , the conditional pdf is also Gaussian with its mean given by

MMSE Estimation for Gaussian