Information Theory
Overview
The theory of information takes several different forms. This corner focuses on the theory of information initially
developed by Claude Shannon.
Topics
- Entropy - the central notion in information theory, it is a measure of uncertainty or information
- Joint Entropy -
the amount of entropy contained in two random variables jointly
- Conditional Entropy - is a measure of the additional amount of uncertainty that one
variable contributes to the join entropy.
- Entropy Rate - the entropy of a stochastic process
- Mutual Information
Information Theory, Statistics and Artificial Intelligence
- Entropy Optimization
is a principle which has broad applicability in statistics and machine learning.