Jump to content

Binary entropy function

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by 85.64.144.107 (talk) at 20:16, 10 November 2009 (Taylor series). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
Entropy of a Bernoulli trial as a function of success probability, called the binary entropy function.

In information theory, the binary entropy function, denoted or , is defined as the entropy of a Bernoulli trial with probability of success p. Mathematically, the Bernoulli trial is modelled as a random variable X that can take on only two values: 0 and 1. The event is considered a success and the event is considered a failure. (These two events are mutually exclusive and exhaustive.)

If then and the entropy of X is given by

where is taken to be 0. The logarithms in this formula are usually taken (as shown in the graph) to the base 2. See binary logarithm.

When the binary entropy function attains its maximum value. This is the case of the unbiased bit, the most common unit of information entropy.

is distinguished from the entropy function by its taking a single scalar constant parameter. For tutorial purposes, in which the reader may not distinguish the appropriate function by its argument, is often used; however, this could confuse this function with the analogous function related to Rényi entropy, so (with "b" not in italics) should be used to dispel ambiguity.

Derivative

The derivative of the binary entropy function may be expressed as the negative of the logit function:

Taylor series

The taylor series of the binary entropy function in a neighbourhood of 1/2 is

for

See also

References