Difference between revisions of "Huffman coding"

From Hydrogenaudio Knowledgebase
Jump to: navigation, search
 
(13 intermediate revisions by 3 users not shown)
Line 1: Line 1:
Most [[lossy]] audio encoders use the classic technique of the Huffman algorithm. It acts at the end of the compression to code information and this is not therefore itself a compression algorithm but rather a coding method, known as (entropy coding). This coding creates variable length codes on a whole number of bits. Higher probability symbols have shorter codes. Huffman codes have the property to have a unique prefix, they can therefore be decoded correctly in spite of their variable length. The decoding step is very fast (via a correspondence table). This kind of coding allows to save on the average a bit less than 20% of space.
+
{{stub}}
  
It is an ideal complement of the perceptual coding: During big polyphonies, the perceptual coding is very efficient because many sounds are masked or lessened, but little information is identical, so the Huffmann algorithm is very seldom efficient. During "pure" sounds there are few [[masking]] effects, but Huffman is then very efficient because digitalized sound contains many repetitive bytes, that will then be replaced by shorter codes.
 
  
text © Gabriel Bouvigne - http://mp3-tech.org/
+
Most [[lossy]] audio encoders (MP3, Vorbis, AAC) use a common algorithmic technique, known as Huffman coding. This is not a compression algorithm, but rather a coding method, known as (entropy coding). This coding creates variable length codes on a whole number of bits. Higher probability symbols have shorter codes. Huffman codes have the property to have a unique prefix, they can therefore be decoded correctly in spite of their variable length. The decoding step is very fast (via a correspondence table). This kind of coding allows to save on the average a bit less than 20% of space.
 +
 
 +
It is an ideal complement of the perceptual coding: During big polyphonies, the perceptual coding is very efficient, due to the fact that many sounds are masked or lessened, but little information is identical, so the Huffman algorithm is very seldomly efficient. During "pure" sounds there are few [[masking]] effects, but Huffman is then very efficient because digitalized sound contains many repetitive bytes. These bytes will later be replaced by shorter codes.
 +
 
 +
text © Gabriel Bouvigne - www.mp3-tech.org
 +
 
 +
==References==
 +
* [http://compression.ru/download/articles/huff/huffman_1952_minimum-redundancy-codes.pdf Huffman Original Research Paper from 1952] on entropy coding. 
 +
* [http://bcl.sourceforge.net/ Huffman source code] an ANSI C implementation of the algorithm.
 +
* [http://alexvn.freeservers.com/s1/huffman_template_algorithm.html#label_Algorithm n-ary Huffman Template algorithm] written in C++ (GPL license).
 +
 
 +
 
 +
[[Category:Technical]]
 +
[[Category:Algorithms]]

Latest revision as of 00:43, 7 September 2006


Most lossy audio encoders (MP3, Vorbis, AAC) use a common algorithmic technique, known as Huffman coding. This is not a compression algorithm, but rather a coding method, known as (entropy coding). This coding creates variable length codes on a whole number of bits. Higher probability symbols have shorter codes. Huffman codes have the property to have a unique prefix, they can therefore be decoded correctly in spite of their variable length. The decoding step is very fast (via a correspondence table). This kind of coding allows to save on the average a bit less than 20% of space.

It is an ideal complement of the perceptual coding: During big polyphonies, the perceptual coding is very efficient, due to the fact that many sounds are masked or lessened, but little information is identical, so the Huffman algorithm is very seldomly efficient. During "pure" sounds there are few masking effects, but Huffman is then very efficient because digitalized sound contains many repetitive bytes. These bytes will later be replaced by shorter codes.

text © Gabriel Bouvigne - www.mp3-tech.org

References