Difference between revisions of "161-A3.1"

From Microlab Classes
Jump to navigation Jump to search
Line 126: Line 126:
 
{{NumBlk|::|<math>\begin{align}
 
{{NumBlk|::|<math>\begin{align}
 
\frac{\partial I\left(A;B\right)}{\partial \alpha} & = \frac{\partial}{\partial \alpha}\left( \alpha\log_2\frac{1}{\alpha} + \left(1-\alpha\right)\log_2\frac{1}{1-\alpha} \right) = 0\\
 
\frac{\partial I\left(A;B\right)}{\partial \alpha} & = \frac{\partial}{\partial \alpha}\left( \alpha\log_2\frac{1}{\alpha} + \left(1-\alpha\right)\log_2\frac{1}{1-\alpha} \right) = 0\\
& = \frac{\partial}{\partial \alpha}\left(\frac{\alpha}{\ln 2}\ln\frac{1}{\alpha} + \frac{1-\alpha}{\ln 2}\ln\frac{1}{1-\alpha}\right) \\
+
& = \frac{\partial}{\partial \alpha}\left(\frac{\alpha}{\ln 2}\ln\frac{1}{\alpha} + \frac{1-\alpha}{\ln 2}\ln\frac{1}{1-\alpha}\right) = 0\\
& =  
+
& = \frac{1}{\ln 2}\ln\frac{1}{\alpha} - \alpha\frac{1}{\alpha^2}\frac{\alpha}{\ln 2} - \frac{1}{\ln 2}\ln \frac{1}{1-\alpha} - \frac{1-\alpha}{\ln 2}\frac{1-\alpha}{\left(1-\alpha\right)^2} = 0\\
 
\end{align}</math>|{{EquationRef|15}}}}
 
\end{align}</math>|{{EquationRef|15}}}}
  

Revision as of 18:47, 2 October 2020

  • Activity: Mutual Information and Channel Capacity
  • Instructions: In this activity, you are tasked to
    • Walk through the examples.
    • Calculate the channel capacity of different channel models.
  • Should you have any questions, clarifications, or issues, please contact your instructor as soon as possible.

Example 1: Mutual Information

Given the following probabilities:

Table 1: : Blood Type, : Chance for Skin Cancer
A B AB O
Very Low 1/8 1/16 1/32 1/32
Low 1/16 1/8 1/32 1/32
Medium 1/16 1/16 1/16 1/16
High 1/4 0 0 0

To get the entropies of and , we need to calculate the marginal probabilities:

 

 

 

 

(1)

 

 

 

 

(2)

And since:

 

 

 

 

(3)

We get:

 

 

 

 

(4)

 

 

 

 

(5)

Calculating the conditional entropies using:

 

 

 

 

(6)

 

 

 

 

(7)

Note that . Calculating the mutual information, we get:

 

 

 

 

(8)

Or equivalently:

 

 

 

 

(9)

Let us try to understand what this means:

  • If we only consider , we have the a priori probabilities, for each blood type, and we can calculate the entropy, , i.e. the expected value of the information we get when we observe the results of the blood test.
  • Will our expectations change if we do not have access to the blood test, but instead, we get to access (1) the person's susceptibility to skin cancer, and (2) the joint probabilities in Table 1? Since we are given more information, we expect one of the following:
    • The uncertainty to be equal to the original uncertainty if and are independent, since , thus .
    • A reduction in the uncertainty equal to , due to the additional information given by .
    • If and are perfectly correlated, we reduce the uncertainty to zero since and .

Example 2: A Noiseless Binary Channel

Consider transmitting information over a noiseless binary channel shown in Fig. 1. The input, has a priori probabilities and , and the output is . Calculating the output probabilities:

 

 

 

 

(10)

 

 

 

 

(11)

Thus, the entropy at the output is:

 

 

 

 

(12)

The conditional entropy, is then:

 

 

 

 

(13)

Which is expected since and are perfectly correlated, and there is no uncertainty in once is given. Thus, the mutual information is:

 

 

 

 

(14)

To get the channel capacity, we get the maximum over all . Thus, we can take the derivative of the mutual information and set it equal to zero, to find the optimal . And noting that , we get:

 

 

 

 

(15)

Example 3: A Noisy Channel with Non-Overlapping Outputs

Example 4: The Binary Symmetric Channel (BSC)

Sources

  • Yao Xie's slides on Entropy and Mutual Information