Difference between revisions of "Entropy, Relative Entropy, Mutual Information"
Jump to navigation
Jump to search
Line 59: | Line 59: | ||
* a measure of the uncertainty associated with a set of variables | * a measure of the uncertainty associated with a set of variables | ||
− | The joint entropy of a pair of discrete random variables <math>\left(X, Y\right)</math> with a joint distribution <math>p\left(x, y\right)</math> is defined as | + | The ''joint entropy'' of a pair of discrete random variables <math>\left(X, Y\right)</math> with a joint distribution <math>p\left(x, y\right)</math> is defined as |
+ | |||
+ | {{NumBlk|:|<math>H\left(X, Y\right)=-\sum_{x \in \mathcal{X}} \sum_{y \in \mathcal{Y}} p\left(x, y\right)\cdot \log_2 p\left(x, y\right)</math>|{{EquationRef|6}}}} |
Revision as of 18:32, 25 June 2020
Contents
Definitions
Entropy
- a measure of the uncertainty of a random variable
- The entropy of a random variable is a measure of the uncertainty of the random variable
- it is a measure of the amount of information required on the average to describe the random variable
Relative Entropy
- a measure of the distance between two distributions
- a measure of the inefficiency of assuming that the distribution is when the true distribution is .
Mutual Information
- a measure of the amount of information that one random variable contains about another random variable
Entropy
Definitions:
- a measure of the uncertainty of a random variable
- The entropy of a random variable is a measure of the uncertainty of the random variable
- it is a measure of the amount of information required on the average to describe the random variable
The entropy of a discrete random variable, , is
-
(1)
where has a probability mass function (pmf), , and an alphabet .
Expected Value
For a discrete random variable, , with probability mass function, , the expected value of is
-
(2)
For a discrete random variable, , with probability mass function, , the expected value of is
-
(3)
Consider the case where . We get
-
(4)
Lemma 1: Entropy is greater than or equal to zero
-
(5)
Proof: Since , then , and subsequently, . Thus from Eq. (4) we get .
Lemma 2: Changing the logarithm base
-
(6)
Proof:
- Given that
- And since
- We get
Note that the entropy, , has units of bits for , or nats (natural units) for , or dits (decimal digits) for .
Joint Entropy
Definition:
- a measure of the uncertainty associated with a set of variables
The joint entropy of a pair of discrete random variables with a joint distribution is defined as
-
(6)