Entropy

Overview


The primary definition in information theory is Entropy. It provides a measure of how uncertain a random variable is. While there are other measures that could be deemed measures of uncertainty (see variance for example), , entropy is central to information theory because of its connection to data compression.

Entropy


The entropy of a random variable is defined as
{% H(X) = - \sum p(x) \times log [p(x)] %}
The entropy can be recast in terms of expectation.
{% H(X) = \mathbb{E} [log (1/p(x))] %}


The plot below shows the entropy of a bernoulli variable with probability p on the x axis. (the other probability is 1-p )

copy

Contents