Multimedia
Compression
BY
V.PRADEEPA
II – M.SC(CS&IT)
NADAR SARASWATHI
COLLEGE OF ARTS
AND SCIENCE, THENI
Why Compress
 Raw data are huge.
 Audio:
CD quality music
44.1kHz*16bit*2 channel=1.4Mbps
 Video:
near-DVD quality true color animation
640px*480px*30fps*24bit=220Mbps
 Impractical in storage and bandwidth
Outline
 Generic Compression Overview
 Content specific Compression
 Lossy Compression
Introduction to
Generic Compression
Algorithm
Lossless Compression
Generic Compression
 Also called Entropy Encoding
 Lossless Compression Algorithms
 Entropy can defined as:
 Need statistical knowledge of data
 Well-known Algorithms:
 Rice coding
 Huffman coding
 Arithmetic coding
Huffman encoding
Input: ABACDEAACCAABEAABACBDDABCADDBCEAEAAADBE
Order-0 model
Symbol A B C D E
Count 15 7 6 6 5
total:39*3=117 bits
Output:
15*1+(7+6+6+5)*3=87 bits
Compression ratio:
117/87 = 1.34
Property of Huffman encoding
 Easy to implement, high encoding speed
 Unique Prefix Property: no code is a
prefix to any other code
 Adaptive Huffman encoding:
statistical knowledge not available
update Huffman tree when needed
Arithmetic Encoding
 Symbol X, Y
prob(X) = 2/3
prob(Y) = 1/3
Property of Arithmetic Encoding
 Prevent entropy wasting in Huffman
coding, for the number of bits to represent
a symbol can be non-integer
 About 5~10% smaller than Huffman
coding
 Computational intensive
 US patented!!
 Both Huffman and Arithmetic are used in
the entropy encoding stage in JPEG
Application of General
Compression
 Generic file compression like Zip, Rar,
gzip, bzip, etc.
 Final stage of content specific
compression
JPEG uses Huffman or Arithmetic
Monkey’s Audio (ape) uses Rice
Lossless Audio (La) uses Arithmetic
Content specific
Compression
Further De-correlation
De-correlation
 Correlation means redundancy
 However, general algorithm may not find
content-specific correlation
 General algorithm of higher order may not be
efficient enough
 No matter lossy or lossless, multimedia file
format use content-specific pre-filter as 1st
step to
reduce data redundancy.
Correlation in Multimedia
 Audio:
Temporal, Channel
 Still Image:
Color space, Spatial, Stereo
 Video:
Temporal
Audio Channel Correlation
 Correlation between
L/R channels
 L/R to mid/pass band
conversion
 More complex
decorrelation in more
channels

Multimedia compression

Editor's Notes

  • #6 For example, in an image with uniform distribution of gray-level intensity, i.e. pi = 1/256, then the number of bits needed to code each gray level is 8 bits. The entropy of this image is 8. Q: How about an image in which half of the pixels are white (I = 220) and half are black (I = 10)?
  • #7 Content from http://www.cs.cf.ac.uk/Dave/Multimedia/node207.html
  • #9 Symbol with more probability have longer part. From http://www.cs.cf.ac.uk/Dave/Multimedia/node207.html
  • #14 Audio: Temporal: Smooth waveform indicates redundancy Waveform prediction, encode the residue