Conditional entropy tells us the information needed to specify given for a joint distribution .
Formulation
Consider the joint distribution given by from which we draw pairs of values and .
If the value of is already known, then the additional information needed to specify the corresponding value of is given by . Thus, the average amount of information needed to specify can be written as:
which is called the conditional entropy of given .
The product rule tells us that the conditional entropy satisfies the relation:
where:
- is the differential entropy of
- is the differential entropy of the marginal distribution
Thus, the information needed to describe and is given by: