Difference between revisions of "Entropy, Relative Entropy, Mutual Information"

From Microlab Classes
Jump to navigation Jump to search
Line 47: Line 47:
  
 
'''Proof''':  
 
'''Proof''':  
* Given that <math>H_b\left(X\right)=-\sum_{x\in \mathcal{X}} p\left(x\right)\cdot \log_b p\left(x\right)</math>, and <math>H_a\left(X\right)=-\sum_{x\in \mathcal{X}} p\left(x\right)\cdot \log_a p\left(x\right)</math>
+
* Given that  
 +
** <math>H_b\left(X\right)=-\sum_{x\in \mathcal{X}} p\left(x\right)\cdot \log_b p\left(x\right)</math>
 +
** <math>H_a\left(X\right)=-\sum_{x\in \mathcal{X}} p\left(x\right)\cdot \log_a p\left(x\right)</math>
 
* And since <math>\log_b p = \log_b a \cdot log_a p</math>
 
* And since <math>\log_b p = \log_b a \cdot log_a p</math>
 
* We get <math>H_b\left(X\right)=\left(\log_b a\right)\cdot H_a\left(X\right)</math>
 
* We get <math>H_b\left(X\right)=\left(\log_b a\right)\cdot H_a\left(X\right)</math>
 +
 +
Note that the entropy, <math>H_b\left(X\right)</math>, has units of ''bits'' for <math>b=2</math>, or ''nats'' (natural units) for <math>b=e</math>, or ''dits'' (decimal digits) for <math>b=10</math>.

Revision as of 18:22, 25 June 2020

Definitions

Entropy

  • a measure of the uncertainty of a random variable
  • The entropy of a random variable is a measure of the uncertainty of the random variable
    • it is a measure of the amount of information required on the average to describe the random variable

Relative Entropy

  • a measure of the distance between two distributions
  • a measure of the inefficiency of assuming that the distribution is when the true distribution is .

Mutual Information

  • a measure of the amount of information that one random variable contains about another random variable

Entropy

Definitions:

  • a measure of the uncertainty of a random variable
  • The entropy of a random variable is a measure of the uncertainty of the random variable
    • it is a measure of the amount of information required on the average to describe the random variable

The entropy of a discrete random variable, , is

 

 

 

 

(1)

where has a probability mass function (pmf), , and an alphabet .

Expected Value

For a discrete random variable, , with probability mass function, , the expected value of is

 

 

 

 

(2)

For a discrete random variable, , with probability mass function, , the expected value of is

 

 

 

 

(3)

Consider the case where . We get

 

 

 

 

(4)

Lemma 1: Entropy is greater than or equal to zero

 

 

 

 

(5)

Proof: Since , then , and subsequently, . Thus from Eq. (4) we get .

Lemma 2: Changing the logarithm base

 

 

 

 

(6)

Proof:

  • Given that
  • And since
  • We get

Note that the entropy, , has units of bits for , or nats (natural units) for , or dits (decimal digits) for .