Upcoming SlideShare
×

Thanks for flagging this SlideShare!

Oops! An error has occurred.

×

# Saving this for later?

### Get the SlideShare app to save on your phone or tablet. Read anywhere, anytime - even offline.

Standard text messaging rates apply

# vector QUANTIZATION

1,263
views

Published on

Published in: Technology, Art & Photos

0 Likes
Statistics
Notes
• Full Name
Comment goes here.

Are you sure you want to Yes No
• Be the first to comment

• Be the first to like this

Views
Total Views
1,263
On Slideshare
0
From Embeds
0
Number of Embeds
0
Actions
Shares
0
46
0
Likes
0
Embeds 0
No embeds

No notes for slide

### Transcript

• 1. Vector Quantization Aniruddh Tyagi 02-06-12
• 2. Voronoi Region• Blocks: o A sequence of audio. o A block of image pixels. Formally: vector example: (0.2, 0.3, 0.5, 0.1)• A vector quantizer maps k-dimensional vectors in the vector space Rk into a finite set of vectors Y = {yi: i = 1, 2, ..., N}. Each vector yi is called a code vector or a codeword. and the set of all the codewords is called a codebook. Associated with each codeword, yi, is a nearest neighbor region called Voronoi region, and it is defined by:• The set of Voronoi regions partition the entire space Rk .
• 3. Two Dimensional Voronoi DiagramCodewords in 2-dimensional space. Input vectors are marked with anx, codewords are marked with red circles, and the Voronoi regions areseparated with boundary lines.
• 4. The Schematic of a VectorQuantizer
• 5. Compression Formula• Amount of compression: o Codebook size is K, input vector of dimension L o In order to inform the decoder of which code vector is selected, we need to use bits.  E.g. need 8 bits to represent 256 code vectors. o Rate: each code vector contains the reconstruction value of L source output samples, the number of bits per sample would be: . o Sample: a scalar value in vector. o K: level of vector quantizer.
• 6. VQ vs SQAdvantage of VQ over SQ:• For a given rate, VQ results in a lower distortion than SQ.• If the source output is correlate, vectors of source output values will tend to fall in clusters. o E.g. Sayood’s book Exp 9.3.1• Even if no dependency: greater flexibility. o E.g. Sayood’s book Exp 9.3.2
• 7. Algorithms• Lloyd algorithm: pdf-optimized quantizer, assume that distribution is known• LBG: (VQ) o Continuous (require integral ooperation) o Modified: with training set.
• 8. LBG Algorithm– Determine the number of codewords, N, or the size of the codebook.– Select N codewords at random, and let that be the initial codebook. The initial codewords can be randomly chosen from the set of input vectors.– Using the Euclidean distance measure clusterize the vectors around each codeword. This is done by taking each input vector and finding the Euclidean distance between it and each codeword. The input vector belongs to the cluster of the codeword that yields the minimum distance.
• 9. LBG Algorithm (contd.)4.Compute the new set of codewords. This is done by obtaining theaverage of each cluster. Add the component of each vector and divide bythe number of vectors in the cluster.where i is the component of each vector (x, y, z, ... directions), m is thenumber of vectors in the cluster.5.Repeat steps 2 and 3 until the either the codewords dont change orthe change in the codewords is small.
• 10. Other Algorithms• Problems: LBG is a greedy algorithm, may fall into Local minimum.• Four methods selecting initial vectors: o Random o Splitting ( with perturbation vector) Animation o Train with different subset o PNN (pairwise nearest neighbor)• Empty cell problem: o No input correspond to a output vector o Solution: give to other clusters, e.g. most populate cluster.
• 11. LBG for image compression• Taking blocks of images as vector L=NM.• If K vectors in code book: o need to use bits. o Rate:• The higher the value K, the better quality, but lower compression ratio.• Overhead to transmit code book:• Train with a set of images.
• 12. Rate_Dimension Product• Rate-dimension product o The size of the codebook increase exponentially with the rate. o Suppose we want to encode a source using R bits/sample. If we use an L-d quantizer, we would group L samples together into vectors. This means that we would have RL bits available to represent wach vector. o With RL bits, we can represent 2^(RL) output vectors.
• 13. Tree structured VQ• Set vectors in different quadrant. Only signs of vectors need to be compared. Thus reduce the number of comparisons by 2^L for L-d vector problem.• It works well for symmetric distribution. But not when we lose more and more symmetry.
• 14. Tree Structured Vector Quantizer• Extend to non-symmetric case: o Divide the set of output points into two groups, g0 and g1, and assign to each group a test vector s.t. output points in each group are closer to test vector assigned to that group than to the test vector assigned to the other group. o Label the two test vectors 0 and 1. o When we got an input vector, compare it against the test vectors. Depending on the outcome, the input is compared to the output points associated with the test vector closest to the input. o After these two comparisons, we can discard half of the output points. o Comparison with the test vectors takes the place of looking at the signs of the components to decide which set of output points to discard from contention. o If the total number of output points is K, we make( K/2)+2 comparisons instead of K comparisons. o Can continue to expand the number of groups. Finally: 2logK comparisons instead of K.( 2 comparisons with the test vectors and a total of logK stages
• 15. Tree Structured VQ (continued)• Since the test vectors are assigned to groups: 0, 1, 00,01,10,11,000,001,010,011,100,101,110,111 etc. which are the nodes of a binary tree, the VQ has the name “Tree Structured VQ”.• Penalty: o Possible increase in distortion: it is possible that at some stage the input is closer to one test vector while at the same time being closest to an output belonging to the rejected group. o Increase storage: output points from VQ codebook plus the test vectors.