Very easy to follow tutorial on implementing a Huffman tree structure.
From Rosetta Code.org:
Huffman encoding is a way to assign binary codes to symbols that reduces the overall number of bits used to encode a typical string of those symbols.
For example, if you use letters as symbols and have details of the frequency of occurrence of those letters in typical strings, then you could just encode each letter with a fixed number of bits, such as in ASCII codes. You can do better than this by encoding more frequently occurring letters such as e and a, with smaller bit strings; and less frequently occurring letters such as q and x with longer bit strings.
Any string of letters will be encoded as a string of bits that are no-longer of the same length per letter. To successfully decode such as string, the smaller codes assigned to letters such as ‘e’ cannot occur as a prefix in the larger codes such as that for ‘x’.
If you were to assign a code 01 for ‘e’ and code 011 for ‘x’, then if the bits to decode started as 011… then you would not know if you should decode an ‘e’ or an ‘x’.
The Huffman coding scheme takes each symbol and its weight (or frequency of occurrence), and generates proper encodings for each symbol taking account of the weights of each symbol, so that higher weighted symbols have fewer bits in their encoding.
A Huffman encoding can be computed by first creating a tree of nodes:
Create a leaf node for each symbol and add it to the priority queue.
While there is more than one node in the queue:
Remove the node of highest priority (lowest probability) twice to get two nodes.
Create a new internal node with these two nodes as children and with probability equal to the sum of the two nodes’ probabilities.
Add the new node to the queue.
The remaining node is the root node and the tree is complete.