# 6.22: Compression and the Huffman Code

- Page ID
- 1874

Learning Objectives

- The Huffman source coding algorithm is provably maximally efficient.

Shannon's Source Coding Theorem has additional applications in **data compression**. Here, we have a symbolic-valued signal source, like a computer file or an image, that we want to represent with as few bits as possible. Compression schemes that assign symbols to bit sequences are known as **lossless** if they obey the Source Coding Theorem; they are **lossy** if they use fewer bits than the alphabet's entropy. Using a lossy compression scheme means that you cannot recover a symbolic-valued signal from its compressed version without incurring some error. You might be wondering why anyone would want to intentionally create errors, but lossy compression schemes are frequently used where the efficiency gained in representing the signal outweighs the significance of the errors.

Shannon's Source Coding Theorem states that symbolic-valued signals require **on the average** at least **H(A)** number of bits to represent each of its values, which are symbols drawn from the alphabet **A**. In the module on the Source Coding Theorem we find that using a so-called **fixed rate** source coder, one that produces a fixed number of bits/symbol, may not be the most efficient way of encoding symbols into bits. What is not discussed there is a procedure for designing an efficient source coder: one **guaranteed** to produce the fewest bits/symbol on the average. That source coder is not unique, and one approach that does achieve that limit is the **Huffman source coding algorithm**.

Note

In the early years of information theory, the race was on to be the first to find a **provably** maximally efficient source coding algorithm. The race was won by then MIT graduate student David Huffman in 1954, who worked on the problem as a project in his information theory course. We're pretty sure he received an “A.”

- Create a vertical table for the symbols, the best ordering being in decreasing order of probability.
- Form a binary tree to the right of the table. A binary tree always has two branches at each node. Build the tree by merging the two lowest probability symbols at each level, making the probability of the node equal to the sum of the merged nodes' probabilities. If more than two nodes/symbols share the lowest probability at a given level, pick any two; your choice won't affect \[\overline{B(A)}\]
- At each node, label each of the emanating branches with a binary number. The bit sequence obtained from passing from the tree's root to the symbol is its Huffman code.

Example \(\PageIndex{1}\):

The simple four-symbol alphabet used in the Entropy and Source Coding modules has a four-symbol alphabet with the following probabilities,

\[Pr[a_{0}]=\frac{1}{2}\]

\[Pr[a_{1}]=\frac{1}{4}\]

\[Pr[a_{2}]=\frac{1}{8}\]

\[Pr[a_{3}]=\frac{1}{8}\]

and an entropy of 1.75 bits. This alphabet has the Huffman coding tree shown in Fig. 6.22.1

The code thus obtained is not unique as we could have labeled the branches coming out of each node differently. The average number of bits required to represent this alphabet equals **1.75 **bits, which is the Shannon entropy limit for this source alphabet. If we had the symbolic-valued signal

\[s(m)=\left \{ a_{2},a_{3},a_{1},a_{4},a_{1},a_{2,...} \right \}\]

Our Huffman code would produce the bitstream

\[b(n)=101100111010...\]

If the alphabet probabilities were different, clearly a different tree, and therefore different code, could well result. Furthermore, we may not be able to achieve the entropy limit. If our symbols had the probabilities

\[Pr[a_{1}]=\frac{1}{2},Pr[a_{2}]=\frac{1}{4},Pr[a_{3}]=\frac{1}{5},\; and\; Pr[a_{4}]=\frac{1}{20}\]

The average number of bits/symbol resulting from the Huffman coding algorithm would equal **1.75 **bits. However, the entropy limit is 1.68 bits. The Huffman code does satisfy the Source Coding Theorem—its average length is within one bit of the alphabet's entropy—but you might wonder if a better code existed. David Huffman showed mathematically that no other code could achieve a shorter average code than his. We can't do better.

Exercise \(\PageIndex{1}\)

Derive the Huffman code for this second set of probabilities, and verify the claimed average code length and alphabet entropy.

**Solution**

The Huffman coding tree for the second set of probabilities is **identical** to that for the first (Fig. 6.22.1). The average code length is

\[\frac{1}{2}1+\frac{1}{4}2+\frac{1}{5}3+\frac{1}{20}3=1.75\; bits\]

The entropy calculation is straightforward:

\[H(A)=-\left ( \frac{1}{2}\log_{2}\frac{1}{2}+\frac{1}{4}\log_{2}\frac{1}{4}+\frac{1}{5}\log_{2}\frac{1}{5}+\frac{1}{20}\log_{2}\frac{1}{20}\right )=1.68\; bits\]

## Contributor

- ContribEEOpenStax