Difference between revisions of "Huffman codes"

From Microlab Classes
Jump to navigation Jump to search
Line 3: Line 3:
  
 
== Optimality ==
 
== Optimality ==
The optimality of the Huffman code hinges on the following two properties of optimal prefix-free codes:
+
The optimality of the Huffman code hinges upon the following two properties of optimal prefix-free codes:
*  
+
* Shorter codewords are assigned to more probable symbols.
*  
+
* There exists an optimal code in which the codewords assigned to the two smallest probabilities are siblings.
 +
 
 +
Let us prove the first statement by contradiction. Let us suppose that <math>l_1 < l_2</math> instead for some symbols <math>x_1,x_2</math> such that <math> p_1 < p_2</math>. If we calculate the expected length due to all symbols, <math>x_1</math> and <math>x_2</math> will contribute the term <math>p_1l_1 + p_2 l_2</math>. We proceed by showing the term obtained by swapping the assignment, <math>p_1l_2 + p_2l_1</math>, is strictly lower.
 +
 
 +
<math>(p_1l_2 + p_2l_1) - (p_1l_1 + p_2l_2) = (p_1 - p_2)(l_2 - l_1) <0,</math>
 +
 
 +
since <math>p_1 < p_2</math> and <math>l_1 < l_2</math>. This contradicts the supposed optimality of the code, and so it must be the case that <math>l_1 \ge l_2</math> whenever <math> p_1 < p_2</math>.
 +
 
 +
We only outline the proof for the second statement. First, we must note that there must be at least two codewords which are siblings of each other. Otherwise, we can "trim" the tree by moving up a codeword by one level. Next, we can use the first statement to perform an exchange argument that will give the desired property.
  
 
== Achieving Shannon's limit ==
 
== Achieving Shannon's limit ==

Revision as of 03:23, 15 March 2021

Construction

Huffman-example.svg

Optimality

The optimality of the Huffman code hinges upon the following two properties of optimal prefix-free codes:

  • Shorter codewords are assigned to more probable symbols.
  • There exists an optimal code in which the codewords assigned to the two smallest probabilities are siblings.

Let us prove the first statement by contradiction. Let us suppose that instead for some symbols such that . If we calculate the expected length due to all symbols, and will contribute the term . We proceed by showing the term obtained by swapping the assignment, , is strictly lower.

since and . This contradicts the supposed optimality of the code, and so it must be the case that whenever .

We only outline the proof for the second statement. First, we must note that there must be at least two codewords which are siblings of each other. Otherwise, we can "trim" the tree by moving up a codeword by one level. Next, we can use the first statement to perform an exchange argument that will give the desired property.

Achieving Shannon's limit

Using Huffman codes to encoder every single symbol from can potentially incur a large penalty since the expected length can be at most 1 D-ary digit away from the entropy bound (using base- logarithm). If we group symbols together at a time, we can achieve the Shannon limit to arbitrary precision.

The idea behind achieving Shannon's compression limit is to use a larger source alphabet. If take symbols at a time and create a Huffman code, we are essentially using the larger input alphabet . Using the bounds for Huffman code, the expected length is given by

.

Here, accounts for the encoding of a block of symbols. We can normalize the inequality and obtain a tighter expression:

.

Since we are encoding independent and identically distributed symbols at a time, the joint entropy . Hence,

,

where the upper bound can be made to approach by letting be sufficiently large.

The proof above is representative of many arguments in information theory that take the blocklength to arbitrarily large values. In that respect, we can look at as an asymptotic compression limit. However, it is also important to keep in mind that directly applying the above construction is very impractical. If we take the English alphabet, with 26 letters, and group them five-at-a-time, then the effective alphabet blows up quickly to above 10 million! Fortunately, more modern coding schemes exist that approach the entropy bound without the heavy exponential overhead. The interested student is encouraged to look up "arithmetic coding."