Conditional entropy tells us the information needed to specify given for a joint distribution .

Formulation

Consider the joint distribution given by from which we draw pairs of values and .

If the value of is already known, then the additional information needed to specify the corresponding value of is given by . Thus, the average amount of information needed to specify can be written as:

which is called the conditional entropy of given .

The product rule tells us that the conditional entropy satisfies the relation:

where:

  • is the differential entropy of
  • is the differential entropy of the marginal distribution

Thus, the information needed to describe and is given by: