The document discusses lossy compression techniques. It begins by explaining that lossy compression algorithms compress data by discarding some information, yielding much higher compression ratios than lossless compression but resulting in distorted approximations of the original data. It then covers various lossy compression methods including quantization, transform coding using the discrete cosine transform (DCT), wavelet-based coding using the discrete wavelet transform (DWT), and techniques like vector quantization (VQ) and the Karhunen-Loeve transform (KLT) that aim to decorrelate signal components before quantization. Key aspects like rate-distortion theory, various distortion measures, and algorithms for quantization are also described.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
We would send hard copy of Journal by speed post to the address of correspondence author after online publication of paper.
We will dispatched hard copy to the author within 7 days of date of publication
Image Processing involves the immense utilisation of Wavelet Transforms, and to apply on images require the knowledge of its application two dimensions.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
We would send hard copy of Journal by speed post to the address of correspondence author after online publication of paper.
We will dispatched hard copy to the author within 7 days of date of publication
Image Processing involves the immense utilisation of Wavelet Transforms, and to apply on images require the knowledge of its application two dimensions.
Fixed Point Realization of Iterative LR-Aided Soft MIMO Decoding AlgorithmCSCJournals
Multiple-input multiple-output (MIMO) systems have been widely acclaimed in order to provide high data rates. Recently Lattice Reduction (LR) aided detectors have been proposed to achieve near Maximum Likelihood (ML) performance with low complexity. In this paper, we develop the fixed point design of an iterative soft decision based LR-aided K-best decoder, which reduces the complexity of existing sphere decoder. A simulation based word-length optimization is presented for physical implementation of the K-best decoder. Simulations show that the fixed point result of 16 bit precision can keep bit error rate (BER) degradation within 0.3 dB for 8×8 MIMO systems with different modulation schemes.
Thanks to image compression, we can shrink those files and save a lot of memory space. JPEG is a popular standard for compressing images, so a 3 MB file can be reduced to as low as 700 KB.
Image compression has four stages. Compression can be achieved at three of these stages, and the level of compression depends on how efficiently it is done and how much image clarity is being compromised. The four stages are explained here.
After an image has been segmented into regions ; the resulting pixels is usually is represented and described in suitable form for further computer processing.
Fixed Point Realization of Iterative LR-Aided Soft MIMO Decoding AlgorithmCSCJournals
Multiple-input multiple-output (MIMO) systems have been widely acclaimed in order to provide high data rates. Recently Lattice Reduction (LR) aided detectors have been proposed to achieve near Maximum Likelihood (ML) performance with low complexity. In this paper, we develop the fixed point design of an iterative soft decision based LR-aided K-best decoder, which reduces the complexity of existing sphere decoder. A simulation based word-length optimization is presented for physical implementation of the K-best decoder. Simulations show that the fixed point result of 16 bit precision can keep bit error rate (BER) degradation within 0.3 dB for 8×8 MIMO systems with different modulation schemes.
Thanks to image compression, we can shrink those files and save a lot of memory space. JPEG is a popular standard for compressing images, so a 3 MB file can be reduced to as low as 700 KB.
Image compression has four stages. Compression can be achieved at three of these stages, and the level of compression depends on how efficiently it is done and how much image clarity is being compromised. The four stages are explained here.
After an image has been segmented into regions ; the resulting pixels is usually is represented and described in suitable form for further computer processing.
In computer science and information theory, data compression, source coding,[1] or bit-rate reduction involves encoding information using fewer bits than the original representation.[2] Compression can be either lossy or lossless. Lossless compression reduces bits by identifying and eliminating statistical redundancy. No information is lost in lossless compression.
Design and Performance Analysis of Convolutional Encoder and Viterbi Decoder ...IJERA Editor
In digital communication forward error correction methods have a great practical importance when channel is
noisy. Convolutional error correction code can correct both type of errors random and burst. Convolution
encoding has been used in digital communication systems including deep space communication and wireless
communication. The error correction capability of convolutional code depends on code rate and constraint
length. The low code rate and high constraint length has more error correction capabilities but that also
introduce large overhead. This paper introduces convolutional encoders for various constraint lengths. By
increasing the constraint length the error correction capability can be increased. The performance and error
correction also depends on the selection of generator polynomial. This paper also introduces a good generator
polynomial which has high performance and error correction capabilities.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Beyond Floating Point – Next Generation Computer Arithmeticinside-BigData.com
John Gustafson from the National University of Singapore presented this talk at Stanford.
“A new data type called a “posit” is designed for direct drop-in replacement for IEEE Standard 754 floats. Unlike unum arithmetic, posits do not require interval-type mathematics or variable size operands, and they round if an answer is inexact, much the way floats do. However, they provide compelling advantages over floats, including simpler hardware implementation that scales from as few as two-bit operands to thousands of bits. For any bit width, they have a larger dynamic range, higher accuracy, better closure under arithmetic operations, and simpler exception-handling. For example, posits never overflow to infinity or underflow to zero, and there is no “Not-a-Number” (NaN) value. Posits should take up less space to implement in silicon than an IEEE float of the same size. With fewer gate delays per operation as well as lower silicon footprint, the posit operations per second (POPS) supported by a chip can be significantly higher than the FLOPs using similar hardware resources. GPU accelerators, in particular, could do more arithmetic per watt and per dollar yet deliver superior answer quality.”
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Watch the video presentation: http://wp.me/p3RLHQ-gjH
In order to be able to visulaize the data, or simply to speed up the process of learning without loosing the important features, we apply dimensionality reduction. methods.
We will talk about 2 methods: PCA and manifold.
[Notebook](https://colab.research.google.com/drive/1_ksjf1K49dUA8XtyDGoL5V3JEajHvFHb)
Capacitance Sensing - Signal-to-Noise Ratio Requirement for CapSense Applicat...Ruth Moore
A minimum Signal-to-Noise Ratio (SNR) requirement of 5:1 is derived for CapSense applications. This is a system level requirement for all implementations of CapSense
Introduction to Convolutional Codes
Convolutional Encoder Structure
Convolutional Encoder Representation(Vector, Polynomial, State Diagram and Trellis Representations )
Maximum Likelihood Decoder
Viterbi Algorithm
MATLAB Simulation
Hard and Soft Decisions
Bit Error Rate Tradeoff
Consumed Time Tradeoff
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofsAlex Pruden
This paper presents Reef, a system for generating publicly verifiable succinct non-interactive zero-knowledge proofs that a committed document matches or does not match a regular expression. We describe applications such as proving the strength of passwords, the provenance of email despite redactions, the validity of oblivious DNS queries, and the existence of mutations in DNA. Reef supports the Perl Compatible Regular Expression syntax, including wildcards, alternation, ranges, capture groups, Kleene star, negations, and lookarounds. Reef introduces a new type of automata, Skipping Alternating Finite Automata (SAFA), that skips irrelevant parts of a document when producing proofs without undermining soundness, and instantiates SAFA with a lookup argument. Our experimental evaluation confirms that Reef can generate proofs for documents with 32M characters; the proofs are small and cheap to verify (under a second).
Paper: https://eprint.iacr.org/2023/1886
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
1. Chapter 8 Lossy Compression
8.1 Introduction
81
Algorithms
8.1
8 1 Introduction Lossless compression algorithms do not
8.2 Distortion Measures deliver compression ratios that are high
8.3
8 3 The Rate-Distortion Theory
Rate Distortion
enough. Hence, most multimedia
8.4 Quantization
compression algorithms are lossy.
p g y
8.5
8 5 Transform Coding
8.6 Wavelet-Based Coding What is lossy compression ?
8.7 W l t P k t
8 7 Wavelet Packets ◦ The compressed data is not the same as the
8.8 Embedded Zerotree of Wavelet Coefficients original data, but a close approximation of it.
8.9 S Partitioning i Hierarchical T
8 9 Set P i i i in Hi hi l Trees (SPIHT) ◦ Yi ld a much hi h compression ratio than
Yields h higher i i h
8.10 Further Exploration that of loss-less compression.
1 2
8.2 Distortion Measures
82 8.3 The Rate-Distortion Theory
83 Rate-
The three most commonly used distortion measures
in image compression are: Provides a framework for the study of
◦ mean square error (MSE) σ 2, tradeoffs between Rate and Distortion.
where xn yn and N are the input data sequence
xn, yn,and sequence,
reconstructed data sequence, and length of the data
sequence respectively.
◦ signal to noise ratio (S ) in decibel units (dB),
(SNR), ( )
◦ peak signal to noise ratio (PSNR),
3 4
2. 8.4 Quantization
84 Uniform Scalar Quantization
Reduce the number of distinct output A uniform scalar quantizer partitions the domain
of input values into equally spaced intervals,
values to a much smaller set, via except p
p possibly at the two outer intervals.
y
quantization. ◦ The output or reconstruction value corresponding to
each interval is taken to be the midpoint of the
Main source of the "loss" in lossyy interval.
i l
compression. ◦ The length of each interval is referred to as the step
Three diff
Th different f
t forms of quantization.
f ti ti size,
size denoted by the symbol ∆ ∆.
Two types of uniform scalar quantizers:
◦ Uniform: midrise and midtread quantizers.
◦ Midrise quantizers have even number of output levels
levels.
◦ Nonuniform: companded quantizer. ◦ Midtread quantizers have odd number of output
◦ Vector Quantization.
Q levels, including zero as one of them (see Fig. 8.2).
5 6
For the special case where ∆ = 1, we can
1
simply compute the output values for these
q
quantizers as:
Performance of an M level quantizer. Let B =
{b0,b1,...,bM }be the set of decision
boundaries and Y = { 1,y2,. ..,yM } the set of
{y }be
reconstruction or output values.
Suppose the input is uniformly di ib d i
S h i i if l distributed in
the interval [−Xmax, Xmax]. The rate of the
quantizer is:
7 8
3. Quantization Error of Uniformly
Distributed Source
Granular distortion: quantization error caused by the quantizer for
q y q
bounded input.
To get an overall figure for granular distortion, notice that decision
boundaries bi for a midrise quantizer are [(i − 1)∆, i∆], i =1..M/2,
covering positive data X ( d another h lf for negative X values).
i ii d (and h half f i l )
Output values yi are the midpoints i∆ − ∆/2, i =1..M/2, again just
considering the positive data. The total distortion is twice the sum
over the positive data or
data,
where we divide by the range of X to normalize to a value of at
most 1.
Since the reconstruction values yi are the midpoints of each
interval, the quantization error must lie within the values. For a
uniformly distributed source, the graph of quantization error is
shown in Fig. 8 3
Fig 8.3.
9 10
Therefore,
Therefore the average squared error is Signal variance is , so
the same as the variance of the if the quantizer is n bits, M=2n, then from
quantization error calculated f
l l d from just Eq. (8.2) we have
the interval [0, Δ] with error values in
.
The error value at x is e(x) = x – Δ/2 so
Δ/2,
the variance of errors is given by
11 12
4. Nonuniform Scalar Quantization Minimize the total distortion by setting the
derivative of Eq. (8.12) to zero.
Two common approaches for nonuniform ◦ Differentiating with respect to yi yields the set of
quantization: Lloyd-Max quantizer and reconstruction values
companded quantizer.
Lloyd-Max quantizer
y q ◦ Th optimal reconstruction value i the weighted
The i l i l is h i h d
◦ The probability distribution fX(x), the decision centroid of the x interval.
boundaries bi and the reconstruction values yi ◦ Differentiating with respect to bi and setting the
Total distortion measure result to zero yields
◦ The decision boundary at the midpoint of two
adjacent reconstruction values
13 14
Companded Quantizer
Algorithm 8 1
8.1
LLOYD-MAX QUANTIZATION
BEGIN
Choose initial level set y0
I=0
Repeat
Compute bi using Eq. 8.14
I=I+1 Companded quantization is nonlinear
nonlinear.
Compute yi using Eq. 8.13 As shown above, a compander consists of a
Until | yi - yi-1 | < ε compressor function G a uniform quantizer
G, quantizer,
and an expander function G−1.
END
The two commonly used companders are
15
the µ-law and A-law companders. 16
5. Vector Quantization (VQ)
According to Shannon’s original work on
Shannon s
information theory, any compression system
p
performs better if it operates on vectors or
p
groups of samples rather than individual
symbols or samples.
Form vectors of input samples b simply
F f l by l
concatenating a number of consecutive
samples into a single vector
vector.
Instead of single reconstruction values as in
scalar quantization, in VQ code vectors with
quantization
n components are used. A collection of
these code vectors form the codebook.
17 18
8.5 Transform Coding
85 Spatial Frequency and DCT
The rationale behind transform coding: Spatial frequency indicates how many times
If Y is the result of a linear transform T of the input pixel values change across an image block.
vector X in such a way that the components of Y are
much less correlated then Y can be coded more
correlated, The DCT formalizes this notion with a
efficiently than X. measure of how much the image contents
If most information is accurately described by the
y y change in correspondence to the number of
first few components of a transformed vector, then
the remaining components can be coarsely quantized, cycles of a cosine wave per block.
or even set to zero, with little signal distortion. The role of the DCT is to decompose the
Discrete Cosine Transform (DCT) will be studied first. original signal into its DC and AC
In addition, we will examine the Karhunen-Loeve
Transform (KLT) which optimally d
T f hi h ti ll decorrelates th
l t the components; the role of the IDCT is to
components of the input X. reconstruct (re-compose) the signal.
19 20
6. Definition of DCT:
DCT:
Given an input function f(i j) over two
f(i,
integer variables i and j (a piece of an image),
the 2D DCT transforms it into a new
function F(u, v), with integer u and v running
over the same range as i and j.
The general definition of the transform is:
21 22
23 24
8. The DCT is a linear transform
transform.
In general, a transform T (or function) is
linear, iff
T (αp + βq)= αT (p)+ βT (q)
βq) (8.21)
(8 21)
where α and β are constants, p and q are
any ffunctions, variables or constants.
bl
From the definition in Eq. 8.17 or 8.19,
q ,
this property can readily be proven for
the DCT because it uses only simple
29
arithmetic operations. 30
The Cosine Basis Functions
31 32
9. 2D Separable Basis Comparison of DCT and DFT
The 2D DCT can be separated into a sequence of The discrete cosine transform is a close counterpart
two, 1D DCT steps: to the Discrete Fourier Transform (DFT). DCT is a
transform that only involves the real part of the DFT.
For a continuous signal we define the continuous
signal,
Fourier
Because the use of digital computers requires us to
It is straightforward to see that this simple change discretize the input signal we define a DFT that
signal,
saves many arithmetic steps. The number of operates on 8 samples of the input signal {f0,f1,...,f7}
as:
iterations required is reduced from 8×8 to 8+8.
8 8 8 8.
33 34
Writing the sine and cosine terms explicitly, we have
The formulation of the DCT that allows it to use only
the cosine basis functions of the DFT is that we can
cancel out the imaginary part of the DFT by making a
symmetric copy of the original input signal.
DCT of 8 input samples corresponds to DFT of the
16 samples made up of original 8 input samples and a
symmetric copy of these, as shown in Fig. 8.10.
35 36
10. A Simple Comparison of DCT and
DFT
Table 8 1 and Fig. 8.11 show the comparison
8.1 Fig 8 11
of DCT and DFT on a ramp function, if only
the first three t
th fi t th terms are used.
d
37 38
Karhunen-
Karhunen-Loeve Transform (KLT)
(KLT) Our goal is to find a transform T such that the
g
components of the output Y are uncorrelated,
i.e ,if t ≠s. Thus, the autocorrelation
The Karhunen-Loeve transform is a matrix of Y takes on the form of a positive diagonal
matrix.
reversible linear transform that exploits the
Since any autocorrelation matrix is symmetric and
statistical properties of the vector non-negative d fi i there are k orthogonal
i definite, h h l
representation. eigenvectors u1,u2,...,uk and k corresponding real and
nonnegative eigenvalues λ1 ≥ λ2 ≥ ···≥λk ≥0.
g g
It optimally decorrelates the input signal
signal.
If we define the Karhunen-Loeve transform as
To understand the optimality of the KLT,
consider the autocorrelation matrix RX of
the input vector X defined as Then, the autocorrelation matrix of Y becomes
39 40
11. KLT Example
To illustrate the mechanics of the KLT KLT, The eigenvalues of RX are λ1=6 1963
=6.1963,
consider the four 3D input vectors x1 =(4, λ2=0.2147, and λ3=0.0264. The
4, 5), x2 =(3, 2, 5), x3 =(5, 7, 6), and x4 =(6, corresponding eigenvectors are
)
7, 7).
41 42
8.6 Wavelet-Based Coding
8 6 Wavelet-
Subtracting the mean vector from each input vector and
g p The objective of the wavelet transform is to
apply the KLT decompose the input signal into components that are
easier to deal with, have special interpretations, or
have some components that can be thresholded away away,
for compression purposes.
We want to be able to at least approximately
reconstruct the original signal given these
h i i l i l i h
Since the rows of T are orthonormal vectors, the inverse components.
transform is just the transpose: T−1= TT ,and The basis functions of the wavelet transform are
localized in both time and frequency.
There are two types of wavelet transforms: the
In
I general, after the KLT most of the "energy" of the
l f h f h " " f h continuous wavelet t
ti l t transform (CWT) and th
f d the
transform coefficients are concentrated within the first few discrete wavelet transform (DWT).
components. This is the "energy compaction" property of the
KLT.
KLT
43 44
12. The Continuous Wavelet Transform
The continuous wavelet transform (CWT)
of f L2(R)at time u and scale s is defined
as:
The inverse of the continuous wavelet
transform is:
t f i
45 46
Multiresolution Analysis in the
The Discrete Wavelet Transform
Wavelet Domain
Discrete wavelets are again formed from a Multiresolution analysis provides the tool to adapt
mother wavelet, but with scale and shift in signal resolution to only relevant details for a
particular task.
discrete steps.
p
The approximation component is then recursively
The DWT makes the connection between decomposed into approximation and detail at
wavelets in the continuous time domain and successively coarser scales.
"filter banks"
"f l b k " in the discrete time domain in
h d d Wavelet functions ψ(t) are used to characterize detail
a multiresolution analysis framework. information. The averaging (approximation)
information is formally determined by a kind of dual
It is possible to show that the dil t d and
i ibl t h th t th dilated d to the mother wavelet, called the "scaling function"
translated family of wavelets ψ φ(t).
Wavelets
W l t are set up such that the approximation at
t h th t th i ti t
resolution 2−j contains all the necessary information
to compute an approximation at coarser resolution
form
f rm an orthonormal basis of L2(R)
rth n rmal f (R). 2−(j+1)
(j+1)
47 48
13. The scaling function must satisfy the so-called dilation
equation:
The wavelet at the coarser level is also expressible as
a sum of translated scaling functions:
The vectors h0[n] and h1[n] are called the low pass
low-pass
and high-pass analysis filters. To reconstruct the
original input, an inverse operation is needed. The
inverse filt
i filters are called synthesis filt
ll d th i filters.
49 50
Wavelet Transform Example
Suppose we are given the following input sequence.
pp g g p q Form a new sequence having length equal to that of
the original sequence by concatenating the two
sequences {xn−1,i } and {dn−1,i }. The resulting sequence
Consider h
C id the transform that replaces the original sequence
f h l h i i l is
with its pairwise average xn−1,i and difference dn−1,i defined as
follows:
This sequence has exactly the same number of
elements as the input sequence — the transform did
not increase the amount of data
data.
Since the first half of the above sequence contain
averages from the original sequence, we can view it as
g g q
The
Th averages and d ff
d differences are applied only on consecutive
l d l a coarser approximation to the original signal. The
pairs of input sequences whose first element has an even second half of this sequence can be viewed as the
index. Therefore, the number of elements in each set {xn−1,i } details or approximation errors of the first half.
pp
and {dn−1,i } i exactly half of th number of elements i th
d is tl h lf f the b f l t in the
original sequence.
51 52
14. It is easily verified that the original sequence
y g q
can be reconstructed from the transformed
sequence using the relations
This transform is the discrete Haar wavelet
transform.
53 54
55 56
15. 57 58
Biorthogonal Wavelets
For orthonormal wavelets, the forward transform
and its inverse are transposes of each other and
the analysis filters are identical to the synthesis
filters.
filters
Without orthogonality, the wavelets for analysis
and synthesis are called “biorthogonal”. The
biorthogonal .
synthesis filters are not identical to the analysis
filters. We denote them as and .
To specify a biorthogonal wavelet transform, we
require both
59 60
16. 2D Wavelet Transform
For an N by N input image, the two-dimensional
two dimensional
DWT proceeds as follows:
◦ Convolve each row of the image with h0[n] and h1[n],
discard the odd numbered columns of the resulting arrays
arrays,
and concatenate them to form a transformed row.
◦ After all rows have been transformed, convolve each
column of the result with h0[n]and h1[n] Again discard the
[n].
odd numbered rows and concatenate the result.
After the above two steps, one stage of the DWT is
complete. Th transformed i
l The f d image now contains four
i f
subbands LL, HL, LH, and HH, standing for low-low,
high-low, etc.
g
The LL subband can be further decomposed to yield
yet another level of decomposition. This process can
be continued until the desired number of
decomposition levels is reached.
61 62
2D Wavelet Transform Example
The input image is a sub-sampled version of
the image Lena. The size of the input is
16×16.
16×16 The filter used in the example is the
Antonini 9/7 filter set
63 64
17. The input image is shown in numerical
form below. Convolve the first row with both h0[n] and h1[n]
and discarding the values with odd-numbered
index. The results of these two operations are:
p
Form the transformed output row by
concatenating the resulting coefficients. The first
g g
row of the transformed image is then:
First,
First we need to compute the analysis
and synthesis high-pass filters. Continue the same process for the remaining
rows.
65 66
Apply the filters to the columns of the resulting
image.
image Apply both h0[n] and h1[n] to each column
and discard the odd indexed results:
The result after all rows have been
processed Concatenate the above results into a single column
and apply the same procedure to each of the
remaining columns.
67 68
18. This completes one stage of the discrete
wavelet transform. We can perform another
transform
stage of the DWT by applying the same
transform procedure illustrated above to the
f d ill d b h
upper left 8 × 8 DC image of I12(x, y). The
resulting two-stage transformed image is
69 70
8.7 Wavelet Packets
87 Discrete Wavelet Transform
In the usual dyadic wavelet decomposition, only
the low-pass filtered subband is recursively
decomposed and thus can be represented by a
logarithmic
l arithmic tree str ct re
structure.
A wavelet packet decomposition allows the
decomposition to be represented by any pruned
subtree of the full tree topology.
The wavelet packet decomposition is very flexible
since a best wavelet basis in the sense of some
cost metric can be found within a large library of
permissible bases
bases.
The computational requirement for wavelet
packet decomposition is relatively low as each
decomposition can be computed in the order of
NlogN using fast filter banks. 71 72
19. Wavelet Packet Decomposition 8.8 Embedded Zerotree of Wavelet
(WPD) Coefficients
Effective and computationally efficient for image
coding.
The EZW algorithm addresses two problems:
◦ obtaining the best image quality for a given bit-rate,
and
◦ accomplishing this task in an embedded fashion
fashion.
Using an embedded code allows the encoder to
terminate the encoding at any point. Hence, the
g yp
encoder is able to meet any target bit-rate exactly.
Similarly, a decoder can cease to decode at any point
and can produce reconstructions corresponding to all
lower-rate encodings.
embedded code contains all lower-rate codes “embedded” at the
beginning of the bitstream
73 74
The Zerotree Data Structure
The EZW algorithm efficiently codes the
"significance map" which indicates the locations of
nonzero quantized wavelet coefficients.
This
Th is achieved using a new data structure called
h d d ll d
the zerotree.
Using the hierarchical wavelet decomposition
presented earlier, we can relate every coefficient
at a given scale to a set of coefficients at the next
finer scale of similar orientation.
The coefficient at the coarse scale is called the
"parent" while all corresponding coefficients are
parent
the next finer scale of the same spatial location
and similar orientation are called "children".
75 76
20. Given a threshold T, a coefficient x is an element of
the zerotree if it is insignificant and all of its
descendants are insignificant as well.
The significance map is coded using the zerotree with
a four-symbol alphabet:
◦ The zerotree root: The root of the zerotree is encoded
with a special symbol indicating that the insignificance of
the coefficients at finer scales is completely predictable.
◦ Isolated zero: The coefficient is insignificant but has some
g
significant descendants.
◦ Positive significance: The coefficient is significant with a
p
positive value.
◦ Negative significance: The coefficient is significant with a
negative value.
77 78
Successive Approximation
Dominant Pass
Quantization
Motivation: Coefficients having their coordinates on the
◦ Takes advantage of the efficient encoding of the dominant list implies that they are not yet
significance map using the zerotree data structure by g
significant.
allowing it to encode more significance maps
maps.
◦ Produce an embedded code that provides a coarse- Coefficients are compared to the threshold Ti to
to-fine, multiprecision logarithmic representation of
p g p determine their significance. If a coefficient is
the scale space corresponding to the wavelet- found to be significant, its magnitude is appended
transformed image. to the subordinate list and the coefficient in the
The SAQ method sequentially applies a sequence wavelet t
l t transform array i set to 0 to enable the
f is t t t bl th
of thresholds T0,...,TN−1 to determine the possibility of the occurrence of a zerotree on
significance of each coefficient.
future dominant passes at smaller thresholds
thresholds.
A dominant list and a subordinate list are
maintained during the encoding and decoding The resulting significance map is zerotree coded.
process.
process
79 80
21. Subordinate Pass
All coefficients on the subordinate list are scanned
and their magnitude (as it is made available to the EZW Example
decoder) is refined to an additional bit of precision.
The width of the uncertainty interval for the true
magnitude of the coefficients is cut in half.
For each magnitude on the subordinate list, the
g ,
refinement can be encoded using a binary alphabet
with a "1" indicating that the true value falls in the
upper half of the uncertainty interval and a "0" 0
indicating that it falls in the lower half.
After the completion of the subordinate pass, the
magnitudes on the subordinate li t are sorted i
it d th b di t list t d in
decreasing order to the extent that the decoder can
perform the same sort.
81 82
Encoding The
Th coefficient −29 i i i ifi
ffi i 29 is insignificant, b contains a significant
but
descendant 33 in LH1. Therefore, it is coded as z.
i i ifi
Since the largest coefficient is 57 the initial
57, The coefficient 30 is also insignificant, and all its descendant
are insignificant, so it is coed as t.
threshold T0 is 32. Continuing in this manner, the dominant pass outputs the
At the beginning, the dominant list contains following symbols:
g y
the coordinates of all the coefficients. {57(p), −37(n), −29(z), 30(t), 39(p), −20(t), 17(t), 33(p), 14(t),
6(z), 10(t), 19(t), 3(t), 7(t), 8(t), 2(t), 2(t), 3(t), 12(t), −9(t),
The following is the list of coefficients visited
g 33(p), 20(t), 2(t), 4(t)}
(p), ( ), ( ), ( )}
in the order of the scan:
{57, −37, −29, 30, 39, −20, 17, 33, 14, 6, 10, 19, There are five coefficients found to be significant: 57, -37, 39,
33,
33 and another 33. Since we know that no coefficients are
33
3, 7, 8, 2, 2, 3, 12, −9, 33, 20, 2, 4}
} greater than 2T0 = 64 and the threshold used in the first
With respect to the threshold T0 = 32, it is dominant pass is 32, the uncertainty interval is thus [32, 64).
easy to see that the coefficients 57 and -37
h h ffi i d 37 The subordinate pass following the dominant pass refines the
magnitude of these coefficients by indicating whether they lie
are significant. Thus, we output a p and a n in the first half or the second half of the uncertainty interval.
to represent them them.
83 84
22. Now the dominant list contains the coordinates of all Before we move on to the second round of
the coefficients except those found to be significant dominant and subordinate passes, we need
and the subordinate list contains the values:
to set the values of the significant
g
dominant list: { 29 30 39, −20, 17, 14, 6, 10, 19, 3, 7, 8,
{−29, 30, 39 20 17 14 6 10 19 3 7 8 coefficients to 0 in the wavelet transform
2, 2, 3, 12, −9, 20, 2, 4}
subordinate list: {57, 37, 39, 33, 33}.
{ , , , , }
array so that they do not prevent the
Now, we attempt to rearrange the values in the
emergence of a new zerotree
zerotree.
subordinate list such that larger coefficients appear The new threshold for second dominant
before smaller ones with the constraint that the
ones, pass is T1 = 16 Using the same procedure as
16.
decoder is able do exactly the same. above, the dominant pass outputs the
The decoder is able to distinguish values from [32, 48)
g [ ) following symbols
g y
and [48, 64). Since 39 and 37 are not distinguishable in
the decoder, their order will not be changed.
The subordinate list is now:
85 86
The subordinate pass that follows will halve
each of the three current uncertainty Decoding
intervals [48, 64), [32, 48), and [16, 32). The Suppose we only received information from the first
pp y
subordinate pass outputs the following bits:
b d h f ll b dominant and subordinate pass.
From the symbols in D0 we can obtain the position of the
significant coefficients.
Then, using the bits decoded from S0, we can reconstruct the
value of these coefficients using the center of the uncertainty
The output of the subsequent dominant and
p q interval.
subordinate passes are shown below:
87 88
23. 8.9 Set Partitioning in Hierarchical
If the decoder received only D0, S0, D1, S1, Trees (SPIHT)
The SPIHT algorithm is an extension of the EZW
D2, and only the first 10 bits of S2, then algorithm.
the reconstruction is The SPIHT algorithm significantly improved the
performance of its predecessor by changing the way
subsets of coefficients are partitioned and how
refinement information is conveyed.
A unique property of the SPIHT bitstream is its
compactness. The resulting bitstream from the SPIHT
algorithm is so compact that passing it through an
entropy coder would only produce very marginal gain
in compression.
No d i i f
N ordering information i explicitly transmitted t
ti is li itl t itt d to
the decoder. Instead, the decoder reproduces the
execution path of the encoder and recovers the
ordering information.
f
89 90
8.10 Further Explorations
8 10
Text books:
◦ Introduction to Data Compression by Khalid Sayood
◦ Vector Quantization and Signal Compression by Allen Gersho
and Robert M. Gray
◦ Digital Image Processing by Rafael C. Gonzales and Richard E.
Woods
◦ Probability and Random Processes with Applications to Signal
Processing by Henry Stark and John W. Woods
◦ A Wavelet Tour of Signal Processing by Stephane G. Mallat
Web sites: → Link to Further Exploration for Chapter 8
8..
including:
◦ An online graphics-based demonstration of the wavelet
transform.
transform
◦ Links to documents and source code related to quantization,
Theory of Data Compression webpage, FAQ for
comp.compression, etc.
◦ A link to an excellent article Image Compression – from DCT to
Wavelets : A Review. 91