Jump to content

Source coding theorem

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by WikiIT (talk | contribs) at 19:17, 18 May 2006. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In information theory, the source coding theorem (Shannon 1948) informally states that:

"N i.i.d. random variables each with entropy H(X) can be compressed into more than NH(X) bits with negligible risk of information loss, as N tends to infinity; but conversely, if they are compressed into fewer than NH(X) bits it is virtually certain that information will be lost." (MacKay 2003).

Devising coding strategies to achieve successfully this compression is the basis of the field of entropy encoding.

A more mathematical statement of the theorem is:

Let X be an ensemble with entropy H(X)=H bits. Given ε > 0 and 0 < δ < 1, there exists a positive integer N0 such that for N > N0,
where Hδ(X)=log2|Sδ(X)|; and Sδ(X) is the smallest subset of values of X such that the probability that x is not in Sδ is less than δ. (MacKay 2003).

The source coding theorem is closely related to the asymptotic equipartition property, and the notion of the typical set.