SlideShare a Scribd company logo
Report On
Medical Image Compression with Security and
Water Marking
By:
ANKIT KUMAR CHAUDHARY
(15/IEE/055)
Under The Guidance Of
Dr. M.A. ANSARI
Department of Electrical Engineering
School of Engineering
Gautam Buddha University
Gautam Budh Nagar UP, India
CONTENTS
1. Abstract
2. Introduction
3. Review
4. Principles behind compression
5. Different classes of compression techniques
6. Normal Compression Vs Medical Image Compression
7. Medical Image Compression Standards
8. Conclusion
9. References
1)Abstract
The increasing utilization of medical imaging in clinical practice and the growing
dimensions of data volumes generated by various medical imaging modalities,
the distribution, storage, and management of digital medical image data sets
requires data compression. Over the past few decades, several image compression
standards have been proposed by international standardization organizations.
This paper discusses the current status of these image compression standards in
medical imaging applications together with some of the legal and regulatory
issues surrounding the use of compression in medical settings.
2)Introduction
Medical imaging has become an indispensable tool in clinical practice. Studies
have shown links between the use of medical imaging exams and declines in
mortality, reduced need for exploratory surgery, fewer hospital admissions,
shorter hospital stays, and longer life expectancy. As a result, the utilization of
medical imaging has risen sharply during the early part of the last decade. In 2003,
the percentage of medical visits in the US by patients aged ≥ 65 years that resulted
in medical imaging was estimated to be 12.8%. While earlier medical imaging
exams were recorded on radiological film, most exams are now acquired digitally.
In addition to increased utilization, there have also been major advances in
medical imaging technology that have resulted in significant increases in the
quantity of digital medical imaging data during the last few decades. For example,
in early 1990s, a typical Computed Tomography (CT) exam of the thorax would
have consisted of 25 slices with 10mm thickness, yielding a data size of roughly
12 megabytes (MBs). Today, a similar exam on a modern CT scanner can yield
sub milli-meter slice thickness with increased in-plane resolution resulting in 600
MB to a gigabyte (GB) of data. In a modern hospital, Picture Archiving and
Communication Systems (PACS) handle the short- and long-term storage,
retrieval, management, distribution, processing and presentation of these large
datasets. Data compression plays an important role in these systems. Since the
earliest days of PACS, compression of medical images has been anticipated and
novel compression techniques have been proposed before standardized
compression approaches were available. However, proprietary compression
techniques greatly increase the cost and effort required to migrate data between
different systems, and interoperability and compatibility of these systems
necessitate the use of standards for digital communications. In this paper, we
provide a review of the current status of image compression standards used for
medical imaging data in these systems (It is also worthwhile to point out that the
role of data compression in the medical setting is not limited to images. Modern
medical practice utilizes many physiological signals (e.g. electrocardiogram
(ECG), electroencephalogram (EEG), and Electromyogram (EMG)) which must
be stored and transmitted in clinical practice. Data compression has an important
role to play for management of such physiological signals as well. However, in
this paper, we limit our discussion to medical images). It is important to note that
there have been earlier reviews of medical image compression techniques. In this
paper, we focus on the image compression standards including the more recent
standards that have not been considered in these earlier reviews. We also compare
the compression performances of these standards on publicly available datasets.
3)Review
Uncompressed multimedia (graphics, audio and video) data requires considerable
storage capacity and transmission bandwidth. Despite rapid progress in mass-
storage density, processor speeds, and digital communication system
performance, demand for data storage capacity and data-transmission bandwidth
continues to outstrip the capabilities of available technologies. The recent growth
of data intensive multimedia-based web applications have not only sustained the
need for more efficient ways to encode signals and images but have made
compression of such signals central to storage and communication technology.
The figures in Table 1 show the qualitative transition from simple text to full-
motion video data and the disk space, transmission bandwidth, and transmission
time needed to store and transmit such uncompressed data.
The examples above clearly illustrate the need for sufficient storage space, large
transmission bandwidth, and long transmission time for image, audio, and video
data. At the present state of technology, the only solution is to compress
multimedia data before its storage and transmission and decompress it at the
receiver for play back. For example, with a compression ratio of 32:1, the space,
bandwidth, and transmission time requirements can be reduced by a factor of 32,
with acceptable quality.
4)Principles behind compression
A common characteristic of most images is that the neighbouring pixels are
correlated and therefore contain redundant information. The foremost task then
is to find less correlated representation of the image. In general, three types of
redundancy can be identified:
1) Spatial Redundancy or correlation between neighbouring pixel values.
2) Spectral Redundancy or correlation between different colour planes or spectral
bands.
3) Temporal Redundancy or correlation between adjacent frames in a sequence
of images.
5)Different classes of compression techniques
Two ways of classifying compression techniques are mentioned here.
(a) Lossless vs. Lossy compression: In lossless compression schemes, the
reconstructed image, after compression, is numerically identical to the original
image. However lossless compression can only a achieve a modest amount of
compression. An image reconstructed following lossy compression contains
degradation relative to the original. Often this is because the compression scheme
completely discards redundant information. However, lossy schemes are capable
of achieving much higher compression. Under normal viewing conditions, no
visible loss is perceived (visually lossless).
(b) Predictive vs. Transform coding: In predictive coding, information already
sent or available is used to predict future values, and the difference is coded. Since
this is done in the image or spatial domain, it is relatively simple to implement
and is readily adapted to local image characteristics. Differential Pulse Code
Modulation (DPCM) is one particular-example of predictive coding. Transform
coding, on the other hand, first transforms the image from its spatial domain
representation to a different type of representation using some well-known
transform and then codes the transformed values (coefficients). This method
provides greater data compression compared to predictive methods, although at
the expense of greater computation.
6)Normal Compression Vs Medical Image Compression
The coding of medical images differs from the coding of standard natural images
in that it is imperative that the integrity of the diagnostic information in medical
images are maintained while providing a reduction in storage space and network
transmission bandwidth requirements. Inevitably, the ultimate solution is through
reversible compression. However, at present, the existing state-of-the-art
reversible technologies cannot achieve a significant reduction in bit-rate deemed
adequate for the current practical applications in biomedical imaging.
There have been numerous compression research studies examining the use of
compression as applied to medical images. The papers can be categorised as
focusing on just a lossless compression method, on just a lossy compression
method, or focusing on both. Most have focused on lossless algorithms since the
medical community has been reluctant to adopt lossy techniques owing to the
legal and regulatory issues that are raised, but this situation may start to change
as more lossy research is performed.
Lossless image compression is typically performed in two steps, decorrelation
and coding. Image decorrelation attempts to reduce the redundancy within the
image. There are several common approaches that have been taken in the
literature to perform this redundancy reduction step including differential pulse
code modulation, hierarchical interpolation, bit-plane encoding and
multiplicative autoregression. Several popular approaches for encoding are
Huffman encoding, Lempel-Ziv encoding, arithmetic encoding and run-length
encoding.
7)Medical Image Compression Standards
Emphasis is placed on those techniques that have been adopted or proposed as
international standards. Particular attention is directed to the older JPEG lossless
processes, the new JPEG LS process and the lossless mode of the proposed JPEG
2000 scheme.
JPEG Predictive Lossless Standard
A predictor combines the values of up to three neighboring samples (A, B, and
C) to form a prediction of the sample indicated by X in Figure. This prediction is
then subtracted from the actual value of sample X, and the difference is encoded
losslessly by either of the entropy coding methods - Huffman or arithmetic. Any
one of the eight predictors listed in Table (under “selection-value”) can be used.
Selections 1, 2, and 3 are onedimensional predictors and selections 4, 5, 6 and 7
are two-dimensional predictors. Selection-value 0 can only be used for
differential coding in the hierarchical mode of operation. The encoders can use
any source image precision from 2 to 16 bits/sample, and can use any of the
predictors except selection-value 0. The decoders must handle any of the sample
precisions and any of the predictors. Lossless codecs typically produce around
2:1 compression for color images with moderately complex scenes.
The JPEG-LS Standard
JPEG-LS is the basis for new lossless/near-lossless compression standard for
compressing continuous-tone, greyscale, or colour digital still images , especially
Medical Images. The standard is based on the LOCO-I algorithm (Low
Complexity LOssless COmpression for Medical Images). The algorithm uses
context modeling. Context is a function of samples in the causal template used to
condition the coding of the present sample. Context modeling is the procedure
determining probability distribution of prediction error from the context. Each
sample value is conditioned on a small number of neighbouring samples.
Encoder
Context Modeling
context is determined from four neighbourhood reconstructed samples at
positions a, b, c,and d of the same component
context determines if the information in the sample x should be encoded in the
regular mode (neighbours not very alike) or run mode (when neighbours are very
alike).
Prediction ( Regular mode )
a, b, and c are used to form a prediction of the sample at position x. Prediction
error is computed as the difference between the actual sample value at position x
and its predicted value. This prediction error is then corrected by a context
dependent term to compensate for systematic biases in prediction.
Error encoding ( Regular mode )
The corrected prediction error (further quantized for near lossless coding) is then
encoded using a procedure derived from Golomb coding. The Golomb coding
procedures depend on the context determined by the values of the samples at
positions a, b, c, and d as well as prediction errors previously encoded for the
same context.
Run mode
This mode is selected when reconstructed values of a,b,c and d are identical or
within bounds when near-lossless coding. The mode skips prediction and error-
coding. The encoder looks, starting at x, for a sequence of consecutive samples
with values identical to the reconstructed value of the sample at a. The length
information is encoded.
Decoding process
Encoding and decoding processes are approximately symmetrical. Decoding
process is followed by a sample mapping procedure which uses the value of each
decoded sample as an index to a look-up table, provided in the compressed image
data. If no table is provided for a specific component the output of the sample
mapping procedure is identical to the input.
Lossless JPEG 2000 Standard
JPEG2000 coding is a kind of unified lossless/lossy coding. The differences
between lossless and lossy algorithms are two parts. The first part is in the
implementation of discrete wavelet transform (DWT), and the second part is in
the rate-control scheme.
DWT computation
DWT is carried out by the mallat decomposition of 2-channel filter banks in
JPEG2000. Filters in the filter banks are classified into two types: One is an
integer filter that has integer coefficients, and the other is a floating filter that has
non-integer coefficients. Lossless JPEG2000 coding uses integer DWT (IWT)
that is carried out by lifting schemes with integer filter and round operation.
Rate-control operation
In JPEG2000 coding, the use of two rate-control methods is allowed. One is code
truncation in the EBCOT algorithm called post-quantization, and the other is
prequantization using a scalar quantizer. Either the post-quantization or the
prequantization method can be used in lossy coding. Meanwhile, no rate-control
operation is required for the lossless coding.
EBCOT algorithm
EBCOT is one of the bit-plane based coding algorithms. The transformed
coefficients are decomposed into bit planes and are encoded by the MQ rithmetic
coder. Then, these encoded coefficients are truncated for the rate-control. When
IWT is used as DWT and pre quantization is skipped, there is no difference
between lossy coding and lossless coding until the code truncation is performed.
To perform lossless coding, we have to choose IWT and skip both the pre-
quantization and the post-quantization steps.
Scope
The proposed research is to develop some novel compression techniques, which
can be termed interframe compression and multistage compression.
Interframe compression
We use the fact that for every patient and at every image-taking session, several
almost identical images are taken. The approach is to designate one of these
images as a baseline image, compute the difference between it and the other
images, and then losslessly compress the baseline image and the difference
images. Since the difference images contain little data, the resulting compression
rate is expected to be over 4.
Multistage compression
In this approach an image is first compressed at a high compression rate but with
loss, and the error image is then compressed losslessly. The resulting compression
is not only strictly lossless, but also expected to yield a high compression rate,
especially if the lossy compression technique is good. This is because the error
image will consist of zero- or small-valued elements, thus allowing for lossless
compression at a high compression rate.
Algorithm of Huffman Code
1) Create sorted nodes based on probability/frequency
2) Start loop
3) Find & remove two smallest probability node
4) Create new node[W[Node]=W[N1]+W[N2]]
5) Insert new node, back to sorted list.
6) Repeat the loop until only one last node is present in the list
Flow Chart of Huffman Algorithm-
Algorithm of DCT
1) Read the image as a matrix.
2) Devide the matrix in block of 8x8.
3) Working from left to right, top to bottom, the DCT is applied to each block.
4) Each block compressed through the quantization.
5) The array of compressed blocks that constitute the image is stored in a
drastically reduce amount of space.
Flow Chart of DCT Algorithm
Result
Fig.1-Servical spine
Fig.2: Ultrasound display by lossless technique
Fig. 3: Knee display output by lossless technique
Fig. 4:Servical spine by lossy technique
Fig. 5: Ultrasound output by lossy technique
Fig. 6: Knee display output by lossy technique
8.Conclusion
Parameters Lossless Technique Lossy Technique
Information Have information without losses Have information some
losses
Size Reduce size Reduce more size compare to lossless
Transmission Harder to transmit compressed file Easy to transmit due to less bandwidth
9)References
1. Yong Rui and Thomas S. Huang, "Image Retrieval: Current Techniques, Promising
Directions, and Open Issues," J Visual Comm. And Image Representation, vol. 10, no.
4, Apr 2016.
2. Xin Yu Zhang and Tian Fu Wang, "Entropy- based Local Histogram Equalization for
Medical Ultrasound Image Enhancement," IEEE Intl. Con! 2015
3. Ivica Dimitrovski, Pero Guguljanov and Suzana Loskovska, "Implementation of Web
Based Medical Image Retrieval System in Oracle," IEEE 2nd Intl. Conference on
Adaptive Science & Technology 2017.
4. H. Greenspan and A. T. Pinhas, "Medical Image categorization and retrieval for
PACSusing the GMM-KL framework," IEEE Trans. 1'110.Tech Biomedicine., vol. 11,
no. 2, Mar. 2017.
5. Dimitris K. Iakovidis, Nikos Pelekis, Evangelos E. Kotsifakos, Ioannis Kopanakis,
Haralampos Karanikas and Yannis Theodoridis, "A Pattern Similarity Scheme for
Medical Image Retrieval," IEEE Trans. Info. Tech in Biomedicine, vol 13, no. 4, Jul.
2018.
6. Hua Yuan and Xiao-Ping Zhang, "Statistical Modeling in the Wavelet Domain for
Compact Feature Extraction and Similarity Measure of Images," IEEE Trans. Circuits
and Systems for Video Tech., vol. 20, no. 3, Mar 2018.
7. T. M. Lehmann, M. O. Guld, C. Thies, B. Plodowski, D. Keysers, B. Ott and H.
Schubeert, "IRMA - Content based image retrieval in medical applications," in Proc.
14th World Congr. Med. 1'110. (Medinfo), IDS, Amsterdam, The Netherlands, vol. 2,
2019.
8. Sharadh Ramaswamy and Kenneth Rose, "Towards Optimal Indexing for Relevance
Feedback in Large Image Databases," IEEE Trans. Image Processing, vol. 18, no. 12,
Dec 2019.

More Related Content

What's hot

Mri brain tumour detection by histogram and segmentation
Mri brain tumour detection by histogram and segmentationMri brain tumour detection by histogram and segmentation
Mri brain tumour detection by histogram and segmentation
iaemedu
 

What's hot (20)

Medical image analysis
Medical image analysisMedical image analysis
Medical image analysis
 
Lung Cancer Prediction using Image Classification
Lung Cancer Prediction using Image ClassificationLung Cancer Prediction using Image Classification
Lung Cancer Prediction using Image Classification
 
Image Binarization for the uses of Preprocessing to Detect Brain Abnormality ...
Image Binarization for the uses of Preprocessing to Detect Brain Abnormality ...Image Binarization for the uses of Preprocessing to Detect Brain Abnormality ...
Image Binarization for the uses of Preprocessing to Detect Brain Abnormality ...
 
A New Algorithm for Fully Automatic Brain Tumor Segmentation with 3-D Convolu...
A New Algorithm for Fully Automatic Brain Tumor Segmentation with 3-D Convolu...A New Algorithm for Fully Automatic Brain Tumor Segmentation with 3-D Convolu...
A New Algorithm for Fully Automatic Brain Tumor Segmentation with 3-D Convolu...
 
Brain Tumor Detection using CNN
Brain Tumor Detection using CNNBrain Tumor Detection using CNN
Brain Tumor Detection using CNN
 
Brain tumour segmentation based on local independent projection based classif...
Brain tumour segmentation based on local independent projection based classif...Brain tumour segmentation based on local independent projection based classif...
Brain tumour segmentation based on local independent projection based classif...
 
BRAIN TUMOR CLASSIFICATION IN 3D-MRI USING FEATURES FROM RADIOMICS AND 3D-CNN...
BRAIN TUMOR CLASSIFICATION IN 3D-MRI USING FEATURES FROM RADIOMICS AND 3D-CNN...BRAIN TUMOR CLASSIFICATION IN 3D-MRI USING FEATURES FROM RADIOMICS AND 3D-CNN...
BRAIN TUMOR CLASSIFICATION IN 3D-MRI USING FEATURES FROM RADIOMICS AND 3D-CNN...
 
Brain Tumor Segmentation in MRI Images
Brain Tumor Segmentation in MRI ImagesBrain Tumor Segmentation in MRI Images
Brain Tumor Segmentation in MRI Images
 
P180203105108
P180203105108P180203105108
P180203105108
 
Brain tumor detection using image segmentation ppt
Brain tumor detection using image segmentation pptBrain tumor detection using image segmentation ppt
Brain tumor detection using image segmentation ppt
 
15ICRASE130513 (1)
15ICRASE130513 (1)15ICRASE130513 (1)
15ICRASE130513 (1)
 
A Novel Adaptive Denoising Method for Removal of Impulse Noise in Images usin...
A Novel Adaptive Denoising Method for Removal of Impulse Noise in Images usin...A Novel Adaptive Denoising Method for Removal of Impulse Noise in Images usin...
A Novel Adaptive Denoising Method for Removal of Impulse Noise in Images usin...
 
Literature Survey on Detection of Brain Tumor from MRI Images
Literature Survey on Detection of Brain Tumor from MRI Images Literature Survey on Detection of Brain Tumor from MRI Images
Literature Survey on Detection of Brain Tumor from MRI Images
 
A Survey on Segmentation Techniques Used For Brain Tumor Detection
A Survey on Segmentation Techniques Used For Brain Tumor DetectionA Survey on Segmentation Techniques Used For Brain Tumor Detection
A Survey on Segmentation Techniques Used For Brain Tumor Detection
 
Brain tumor mri image segmentation and detection
Brain tumor mri image segmentation and detectionBrain tumor mri image segmentation and detection
Brain tumor mri image segmentation and detection
 
Brain Tumor Detection and Classification using Adaptive Boosting
Brain Tumor Detection and Classification using Adaptive BoostingBrain Tumor Detection and Classification using Adaptive Boosting
Brain Tumor Detection and Classification using Adaptive Boosting
 
34 107-1-pb
34 107-1-pb34 107-1-pb
34 107-1-pb
 
Application of-image-segmentation-in-brain-tumor-detection
Application of-image-segmentation-in-brain-tumor-detectionApplication of-image-segmentation-in-brain-tumor-detection
Application of-image-segmentation-in-brain-tumor-detection
 
Lung Cancer Detection on CT Images by using Image Processing
Lung Cancer Detection on CT Images by using Image ProcessingLung Cancer Detection on CT Images by using Image Processing
Lung Cancer Detection on CT Images by using Image Processing
 
Mri brain tumour detection by histogram and segmentation
Mri brain tumour detection by histogram and segmentationMri brain tumour detection by histogram and segmentation
Mri brain tumour detection by histogram and segmentation
 

Similar to Final Dessirtation-1 Report

A Comprehensive lossless modified compression in medical application on DICOM...
A Comprehensive lossless modified compression in medical application on DICOM...A Comprehensive lossless modified compression in medical application on DICOM...
A Comprehensive lossless modified compression in medical application on DICOM...
IOSR Journals
 
Combining 3D run-length encoding coding and searching techniques for medical...
Combining 3D run-length encoding coding and searching  techniques for medical...Combining 3D run-length encoding coding and searching  techniques for medical...
Combining 3D run-length encoding coding and searching techniques for medical...
IJECEIAES
 
REGION OF INTEREST BASED COMPRESSION OF MEDICAL IMAGE USING DISCRETE WAVELET ...
REGION OF INTEREST BASED COMPRESSION OF MEDICAL IMAGE USING DISCRETE WAVELET ...REGION OF INTEREST BASED COMPRESSION OF MEDICAL IMAGE USING DISCRETE WAVELET ...
REGION OF INTEREST BASED COMPRESSION OF MEDICAL IMAGE USING DISCRETE WAVELET ...
ijcsa
 
Joint, Image-Adaptive Compression and Watermarking by GABased Wavelet Localiz...
Joint, Image-Adaptive Compression and Watermarking by GABased Wavelet Localiz...Joint, Image-Adaptive Compression and Watermarking by GABased Wavelet Localiz...
Joint, Image-Adaptive Compression and Watermarking by GABased Wavelet Localiz...
CSCJournals
 
Compressed Medical Image Transfer in Frequency Domain
Compressed Medical Image Transfer in Frequency DomainCompressed Medical Image Transfer in Frequency Domain
Compressed Medical Image Transfer in Frequency Domain
CSCJournals
 

Similar to Final Dessirtation-1 Report (20)

A REVIEW OF IMAGE COMPRESSION TECHNIQUES
A REVIEW OF IMAGE COMPRESSION TECHNIQUESA REVIEW OF IMAGE COMPRESSION TECHNIQUES
A REVIEW OF IMAGE COMPRESSION TECHNIQUES
 
A Comprehensive lossless modified compression in medical application on DICOM...
A Comprehensive lossless modified compression in medical application on DICOM...A Comprehensive lossless modified compression in medical application on DICOM...
A Comprehensive lossless modified compression in medical application on DICOM...
 
Combining 3D run-length encoding coding and searching techniques for medical...
Combining 3D run-length encoding coding and searching  techniques for medical...Combining 3D run-length encoding coding and searching  techniques for medical...
Combining 3D run-length encoding coding and searching techniques for medical...
 
REGION OF INTEREST BASED COMPRESSION OF MEDICAL IMAGE USING DISCRETE WAVELET ...
REGION OF INTEREST BASED COMPRESSION OF MEDICAL IMAGE USING DISCRETE WAVELET ...REGION OF INTEREST BASED COMPRESSION OF MEDICAL IMAGE USING DISCRETE WAVELET ...
REGION OF INTEREST BASED COMPRESSION OF MEDICAL IMAGE USING DISCRETE WAVELET ...
 
COMPARISON OF CHANNEL ESTIMATION AND EQUALIZATION TECHNIQUES FOR OFDM SYSTEMS
COMPARISON OF CHANNEL ESTIMATION AND EQUALIZATION TECHNIQUES FOR OFDM SYSTEMSCOMPARISON OF CHANNEL ESTIMATION AND EQUALIZATION TECHNIQUES FOR OFDM SYSTEMS
COMPARISON OF CHANNEL ESTIMATION AND EQUALIZATION TECHNIQUES FOR OFDM SYSTEMS
 
Embedding Patient Information In Medical Images Using LBP and LTP
Embedding Patient Information In Medical Images Using LBP and LTPEmbedding Patient Information In Medical Images Using LBP and LTP
Embedding Patient Information In Medical Images Using LBP and LTP
 
Embedding Patient Information In Medical Images Using LBP and LTP
Embedding Patient Information In Medical Images Using LBP and LTPEmbedding Patient Information In Medical Images Using LBP and LTP
Embedding Patient Information In Medical Images Using LBP and LTP
 
A spatial image compression algorithm based on run length encoding
A spatial image compression algorithm based on run length encodingA spatial image compression algorithm based on run length encoding
A spatial image compression algorithm based on run length encoding
 
Joint, Image-Adaptive Compression and Watermarking by GABased Wavelet Localiz...
Joint, Image-Adaptive Compression and Watermarking by GABased Wavelet Localiz...Joint, Image-Adaptive Compression and Watermarking by GABased Wavelet Localiz...
Joint, Image-Adaptive Compression and Watermarking by GABased Wavelet Localiz...
 
An efficient lossless medical image
An efficient lossless medical imageAn efficient lossless medical image
An efficient lossless medical image
 
Iaetsd performance analysis of discrete cosine
Iaetsd performance analysis of discrete cosineIaetsd performance analysis of discrete cosine
Iaetsd performance analysis of discrete cosine
 
International Journal on Soft Computing ( IJSC )
International Journal on Soft Computing ( IJSC )International Journal on Soft Computing ( IJSC )
International Journal on Soft Computing ( IJSC )
 
Wavelet based Image Coding Schemes: A Recent Survey
Wavelet based Image Coding Schemes: A Recent Survey  Wavelet based Image Coding Schemes: A Recent Survey
Wavelet based Image Coding Schemes: A Recent Survey
 
Efficient Image Compression Technique using Clustering and Random Permutation
Efficient Image Compression Technique using Clustering and Random PermutationEfficient Image Compression Technique using Clustering and Random Permutation
Efficient Image Compression Technique using Clustering and Random Permutation
 
Efficient Image Compression Technique using Clustering and Random Permutation
Efficient Image Compression Technique using Clustering and Random PermutationEfficient Image Compression Technique using Clustering and Random Permutation
Efficient Image Compression Technique using Clustering and Random Permutation
 
Enhanced Image Compression Using Wavelets
Enhanced Image Compression Using WaveletsEnhanced Image Compression Using Wavelets
Enhanced Image Compression Using Wavelets
 
Compressed Medical Image Transfer in Frequency Domain
Compressed Medical Image Transfer in Frequency DomainCompressed Medical Image Transfer in Frequency Domain
Compressed Medical Image Transfer in Frequency Domain
 
M.sc.iii sem digital image processing unit v
M.sc.iii sem digital image processing unit vM.sc.iii sem digital image processing unit v
M.sc.iii sem digital image processing unit v
 
dFuse: An Optimized Compression Algorithm for DICOM-Format Image Archive
dFuse: An Optimized Compression Algorithm for DICOM-Format Image ArchivedFuse: An Optimized Compression Algorithm for DICOM-Format Image Archive
dFuse: An Optimized Compression Algorithm for DICOM-Format Image Archive
 
A hybrid predictive technique for lossless image compression
A hybrid predictive technique for lossless image compressionA hybrid predictive technique for lossless image compression
A hybrid predictive technique for lossless image compression
 

More from Ankit Chaudhary (7)

Energy Savings Potential of Hybrid Drivetrains
Energy Savings Potential of Hybrid DrivetrainsEnergy Savings Potential of Hybrid Drivetrains
Energy Savings Potential of Hybrid Drivetrains
 
Genetic Algorithm
Genetic AlgorithmGenetic Algorithm
Genetic Algorithm
 
Automatic Room Lights Controller Using Arduino & PIR Sensor
Automatic Room Lights Controller Using Arduino & PIR SensorAutomatic Room Lights Controller Using Arduino & PIR Sensor
Automatic Room Lights Controller Using Arduino & PIR Sensor
 
Automatic Room Light Controller Using Arduinom & PIR Sensor
Automatic Room Light Controller Using Arduinom & PIR SensorAutomatic Room Light Controller Using Arduinom & PIR Sensor
Automatic Room Light Controller Using Arduinom & PIR Sensor
 
Noida Metro Rail Corporation
Noida Metro Rail CorporationNoida Metro Rail Corporation
Noida Metro Rail Corporation
 
Noida Metro Rail Corporation
Noida Metro Rail CorporationNoida Metro Rail Corporation
Noida Metro Rail Corporation
 
Medical Image Compression with security & water marking
Medical Image Compression with security & water markingMedical Image Compression with security & water marking
Medical Image Compression with security & water marking
 

Recently uploaded

Introduction of Biology in living organisms
Introduction of Biology in living organismsIntroduction of Biology in living organisms
Introduction of Biology in living organisms
soumyapottola
 
527598851-ppc-due-to-various-govt-policies.pdf
527598851-ppc-due-to-various-govt-policies.pdf527598851-ppc-due-to-various-govt-policies.pdf
527598851-ppc-due-to-various-govt-policies.pdf
rajpreetkaur75080
 

Recently uploaded (14)

Introduction of Biology in living organisms
Introduction of Biology in living organismsIntroduction of Biology in living organisms
Introduction of Biology in living organisms
 
05232024 Joint Meeting - Community Networking
05232024 Joint Meeting - Community Networking05232024 Joint Meeting - Community Networking
05232024 Joint Meeting - Community Networking
 
Pollinator Ambassador Earth Steward Day Presentation 2024-05-22
Pollinator Ambassador Earth Steward Day Presentation 2024-05-22Pollinator Ambassador Earth Steward Day Presentation 2024-05-22
Pollinator Ambassador Earth Steward Day Presentation 2024-05-22
 
Writing Sample 2 -Bridging the Divide: Enhancing Public Engagement in Urban D...
Writing Sample 2 -Bridging the Divide: Enhancing Public Engagement in Urban D...Writing Sample 2 -Bridging the Divide: Enhancing Public Engagement in Urban D...
Writing Sample 2 -Bridging the Divide: Enhancing Public Engagement in Urban D...
 
Acorn Recovery: Restore IT infra within minutes
Acorn Recovery: Restore IT infra within minutesAcorn Recovery: Restore IT infra within minutes
Acorn Recovery: Restore IT infra within minutes
 
527598851-ppc-due-to-various-govt-policies.pdf
527598851-ppc-due-to-various-govt-policies.pdf527598851-ppc-due-to-various-govt-policies.pdf
527598851-ppc-due-to-various-govt-policies.pdf
 
Breathing in New Life_ Part 3 05 22 2024.pptx
Breathing in New Life_ Part 3 05 22 2024.pptxBreathing in New Life_ Part 3 05 22 2024.pptx
Breathing in New Life_ Part 3 05 22 2024.pptx
 
Oracle Database Administration I (1Z0-082) Exam Dumps 2024.pdf
Oracle Database Administration I (1Z0-082) Exam Dumps 2024.pdfOracle Database Administration I (1Z0-082) Exam Dumps 2024.pdf
Oracle Database Administration I (1Z0-082) Exam Dumps 2024.pdf
 
Hi-Tech Industry 2024-25 Prospective.pptx
Hi-Tech Industry 2024-25 Prospective.pptxHi-Tech Industry 2024-25 Prospective.pptx
Hi-Tech Industry 2024-25 Prospective.pptx
 
123445566544333222333444dxcvbcvcvharsh.pptx
123445566544333222333444dxcvbcvcvharsh.pptx123445566544333222333444dxcvbcvcvharsh.pptx
123445566544333222333444dxcvbcvcvharsh.pptx
 
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
 
Getting started with Amazon Bedrock Studio and Control Tower
Getting started with Amazon Bedrock Studio and Control TowerGetting started with Amazon Bedrock Studio and Control Tower
Getting started with Amazon Bedrock Studio and Control Tower
 
Eureka, I found it! - Special Libraries Association 2021 Presentation
Eureka, I found it! - Special Libraries Association 2021 PresentationEureka, I found it! - Special Libraries Association 2021 Presentation
Eureka, I found it! - Special Libraries Association 2021 Presentation
 
The Canoga Gardens Development Project. PDF
The Canoga Gardens Development Project. PDFThe Canoga Gardens Development Project. PDF
The Canoga Gardens Development Project. PDF
 

Final Dessirtation-1 Report

  • 1. Report On Medical Image Compression with Security and Water Marking By: ANKIT KUMAR CHAUDHARY (15/IEE/055) Under The Guidance Of Dr. M.A. ANSARI Department of Electrical Engineering School of Engineering Gautam Buddha University Gautam Budh Nagar UP, India
  • 2. CONTENTS 1. Abstract 2. Introduction 3. Review 4. Principles behind compression 5. Different classes of compression techniques 6. Normal Compression Vs Medical Image Compression 7. Medical Image Compression Standards 8. Conclusion 9. References
  • 3. 1)Abstract The increasing utilization of medical imaging in clinical practice and the growing dimensions of data volumes generated by various medical imaging modalities, the distribution, storage, and management of digital medical image data sets requires data compression. Over the past few decades, several image compression standards have been proposed by international standardization organizations. This paper discusses the current status of these image compression standards in medical imaging applications together with some of the legal and regulatory issues surrounding the use of compression in medical settings. 2)Introduction Medical imaging has become an indispensable tool in clinical practice. Studies have shown links between the use of medical imaging exams and declines in mortality, reduced need for exploratory surgery, fewer hospital admissions, shorter hospital stays, and longer life expectancy. As a result, the utilization of medical imaging has risen sharply during the early part of the last decade. In 2003, the percentage of medical visits in the US by patients aged ≥ 65 years that resulted in medical imaging was estimated to be 12.8%. While earlier medical imaging exams were recorded on radiological film, most exams are now acquired digitally. In addition to increased utilization, there have also been major advances in medical imaging technology that have resulted in significant increases in the quantity of digital medical imaging data during the last few decades. For example, in early 1990s, a typical Computed Tomography (CT) exam of the thorax would have consisted of 25 slices with 10mm thickness, yielding a data size of roughly 12 megabytes (MBs). Today, a similar exam on a modern CT scanner can yield sub milli-meter slice thickness with increased in-plane resolution resulting in 600 MB to a gigabyte (GB) of data. In a modern hospital, Picture Archiving and Communication Systems (PACS) handle the short- and long-term storage,
  • 4. retrieval, management, distribution, processing and presentation of these large datasets. Data compression plays an important role in these systems. Since the earliest days of PACS, compression of medical images has been anticipated and novel compression techniques have been proposed before standardized compression approaches were available. However, proprietary compression techniques greatly increase the cost and effort required to migrate data between different systems, and interoperability and compatibility of these systems necessitate the use of standards for digital communications. In this paper, we provide a review of the current status of image compression standards used for medical imaging data in these systems (It is also worthwhile to point out that the role of data compression in the medical setting is not limited to images. Modern medical practice utilizes many physiological signals (e.g. electrocardiogram (ECG), electroencephalogram (EEG), and Electromyogram (EMG)) which must be stored and transmitted in clinical practice. Data compression has an important role to play for management of such physiological signals as well. However, in this paper, we limit our discussion to medical images). It is important to note that there have been earlier reviews of medical image compression techniques. In this paper, we focus on the image compression standards including the more recent standards that have not been considered in these earlier reviews. We also compare the compression performances of these standards on publicly available datasets. 3)Review Uncompressed multimedia (graphics, audio and video) data requires considerable storage capacity and transmission bandwidth. Despite rapid progress in mass- storage density, processor speeds, and digital communication system performance, demand for data storage capacity and data-transmission bandwidth continues to outstrip the capabilities of available technologies. The recent growth of data intensive multimedia-based web applications have not only sustained the
  • 5. need for more efficient ways to encode signals and images but have made compression of such signals central to storage and communication technology. The figures in Table 1 show the qualitative transition from simple text to full- motion video data and the disk space, transmission bandwidth, and transmission time needed to store and transmit such uncompressed data. The examples above clearly illustrate the need for sufficient storage space, large transmission bandwidth, and long transmission time for image, audio, and video data. At the present state of technology, the only solution is to compress multimedia data before its storage and transmission and decompress it at the receiver for play back. For example, with a compression ratio of 32:1, the space,
  • 6. bandwidth, and transmission time requirements can be reduced by a factor of 32, with acceptable quality. 4)Principles behind compression A common characteristic of most images is that the neighbouring pixels are correlated and therefore contain redundant information. The foremost task then is to find less correlated representation of the image. In general, three types of redundancy can be identified: 1) Spatial Redundancy or correlation between neighbouring pixel values. 2) Spectral Redundancy or correlation between different colour planes or spectral bands. 3) Temporal Redundancy or correlation between adjacent frames in a sequence of images. 5)Different classes of compression techniques Two ways of classifying compression techniques are mentioned here. (a) Lossless vs. Lossy compression: In lossless compression schemes, the reconstructed image, after compression, is numerically identical to the original image. However lossless compression can only a achieve a modest amount of compression. An image reconstructed following lossy compression contains degradation relative to the original. Often this is because the compression scheme completely discards redundant information. However, lossy schemes are capable of achieving much higher compression. Under normal viewing conditions, no visible loss is perceived (visually lossless). (b) Predictive vs. Transform coding: In predictive coding, information already sent or available is used to predict future values, and the difference is coded. Since
  • 7. this is done in the image or spatial domain, it is relatively simple to implement and is readily adapted to local image characteristics. Differential Pulse Code Modulation (DPCM) is one particular-example of predictive coding. Transform coding, on the other hand, first transforms the image from its spatial domain representation to a different type of representation using some well-known transform and then codes the transformed values (coefficients). This method provides greater data compression compared to predictive methods, although at the expense of greater computation. 6)Normal Compression Vs Medical Image Compression The coding of medical images differs from the coding of standard natural images in that it is imperative that the integrity of the diagnostic information in medical images are maintained while providing a reduction in storage space and network transmission bandwidth requirements. Inevitably, the ultimate solution is through reversible compression. However, at present, the existing state-of-the-art reversible technologies cannot achieve a significant reduction in bit-rate deemed adequate for the current practical applications in biomedical imaging. There have been numerous compression research studies examining the use of compression as applied to medical images. The papers can be categorised as focusing on just a lossless compression method, on just a lossy compression method, or focusing on both. Most have focused on lossless algorithms since the medical community has been reluctant to adopt lossy techniques owing to the legal and regulatory issues that are raised, but this situation may start to change as more lossy research is performed. Lossless image compression is typically performed in two steps, decorrelation and coding. Image decorrelation attempts to reduce the redundancy within the image. There are several common approaches that have been taken in the literature to perform this redundancy reduction step including differential pulse code modulation, hierarchical interpolation, bit-plane encoding and
  • 8. multiplicative autoregression. Several popular approaches for encoding are Huffman encoding, Lempel-Ziv encoding, arithmetic encoding and run-length encoding. 7)Medical Image Compression Standards Emphasis is placed on those techniques that have been adopted or proposed as international standards. Particular attention is directed to the older JPEG lossless processes, the new JPEG LS process and the lossless mode of the proposed JPEG 2000 scheme. JPEG Predictive Lossless Standard A predictor combines the values of up to three neighboring samples (A, B, and C) to form a prediction of the sample indicated by X in Figure. This prediction is then subtracted from the actual value of sample X, and the difference is encoded losslessly by either of the entropy coding methods - Huffman or arithmetic. Any one of the eight predictors listed in Table (under “selection-value”) can be used. Selections 1, 2, and 3 are onedimensional predictors and selections 4, 5, 6 and 7 are two-dimensional predictors. Selection-value 0 can only be used for differential coding in the hierarchical mode of operation. The encoders can use any source image precision from 2 to 16 bits/sample, and can use any of the predictors except selection-value 0. The decoders must handle any of the sample precisions and any of the predictors. Lossless codecs typically produce around 2:1 compression for color images with moderately complex scenes.
  • 9. The JPEG-LS Standard JPEG-LS is the basis for new lossless/near-lossless compression standard for compressing continuous-tone, greyscale, or colour digital still images , especially Medical Images. The standard is based on the LOCO-I algorithm (Low Complexity LOssless COmpression for Medical Images). The algorithm uses context modeling. Context is a function of samples in the causal template used to condition the coding of the present sample. Context modeling is the procedure determining probability distribution of prediction error from the context. Each sample value is conditioned on a small number of neighbouring samples.
  • 10. Encoder Context Modeling context is determined from four neighbourhood reconstructed samples at positions a, b, c,and d of the same component context determines if the information in the sample x should be encoded in the regular mode (neighbours not very alike) or run mode (when neighbours are very alike).
  • 11. Prediction ( Regular mode ) a, b, and c are used to form a prediction of the sample at position x. Prediction error is computed as the difference between the actual sample value at position x and its predicted value. This prediction error is then corrected by a context dependent term to compensate for systematic biases in prediction. Error encoding ( Regular mode ) The corrected prediction error (further quantized for near lossless coding) is then encoded using a procedure derived from Golomb coding. The Golomb coding procedures depend on the context determined by the values of the samples at positions a, b, c, and d as well as prediction errors previously encoded for the same context. Run mode This mode is selected when reconstructed values of a,b,c and d are identical or within bounds when near-lossless coding. The mode skips prediction and error- coding. The encoder looks, starting at x, for a sequence of consecutive samples with values identical to the reconstructed value of the sample at a. The length information is encoded. Decoding process Encoding and decoding processes are approximately symmetrical. Decoding process is followed by a sample mapping procedure which uses the value of each decoded sample as an index to a look-up table, provided in the compressed image data. If no table is provided for a specific component the output of the sample mapping procedure is identical to the input.
  • 12. Lossless JPEG 2000 Standard JPEG2000 coding is a kind of unified lossless/lossy coding. The differences between lossless and lossy algorithms are two parts. The first part is in the implementation of discrete wavelet transform (DWT), and the second part is in the rate-control scheme. DWT computation DWT is carried out by the mallat decomposition of 2-channel filter banks in JPEG2000. Filters in the filter banks are classified into two types: One is an integer filter that has integer coefficients, and the other is a floating filter that has non-integer coefficients. Lossless JPEG2000 coding uses integer DWT (IWT) that is carried out by lifting schemes with integer filter and round operation.
  • 13. Rate-control operation In JPEG2000 coding, the use of two rate-control methods is allowed. One is code truncation in the EBCOT algorithm called post-quantization, and the other is prequantization using a scalar quantizer. Either the post-quantization or the prequantization method can be used in lossy coding. Meanwhile, no rate-control operation is required for the lossless coding. EBCOT algorithm EBCOT is one of the bit-plane based coding algorithms. The transformed coefficients are decomposed into bit planes and are encoded by the MQ rithmetic coder. Then, these encoded coefficients are truncated for the rate-control. When IWT is used as DWT and pre quantization is skipped, there is no difference between lossy coding and lossless coding until the code truncation is performed. To perform lossless coding, we have to choose IWT and skip both the pre- quantization and the post-quantization steps. Scope The proposed research is to develop some novel compression techniques, which can be termed interframe compression and multistage compression. Interframe compression We use the fact that for every patient and at every image-taking session, several almost identical images are taken. The approach is to designate one of these images as a baseline image, compute the difference between it and the other images, and then losslessly compress the baseline image and the difference images. Since the difference images contain little data, the resulting compression rate is expected to be over 4.
  • 14. Multistage compression In this approach an image is first compressed at a high compression rate but with loss, and the error image is then compressed losslessly. The resulting compression is not only strictly lossless, but also expected to yield a high compression rate, especially if the lossy compression technique is good. This is because the error image will consist of zero- or small-valued elements, thus allowing for lossless compression at a high compression rate. Algorithm of Huffman Code 1) Create sorted nodes based on probability/frequency 2) Start loop 3) Find & remove two smallest probability node 4) Create new node[W[Node]=W[N1]+W[N2]] 5) Insert new node, back to sorted list. 6) Repeat the loop until only one last node is present in the list
  • 15. Flow Chart of Huffman Algorithm-
  • 16. Algorithm of DCT 1) Read the image as a matrix. 2) Devide the matrix in block of 8x8. 3) Working from left to right, top to bottom, the DCT is applied to each block. 4) Each block compressed through the quantization. 5) The array of compressed blocks that constitute the image is stored in a drastically reduce amount of space. Flow Chart of DCT Algorithm
  • 17. Result Fig.1-Servical spine Fig.2: Ultrasound display by lossless technique
  • 18. Fig. 3: Knee display output by lossless technique Fig. 4:Servical spine by lossy technique
  • 19. Fig. 5: Ultrasound output by lossy technique Fig. 6: Knee display output by lossy technique
  • 20. 8.Conclusion Parameters Lossless Technique Lossy Technique Information Have information without losses Have information some losses Size Reduce size Reduce more size compare to lossless Transmission Harder to transmit compressed file Easy to transmit due to less bandwidth
  • 21. 9)References 1. Yong Rui and Thomas S. Huang, "Image Retrieval: Current Techniques, Promising Directions, and Open Issues," J Visual Comm. And Image Representation, vol. 10, no. 4, Apr 2016. 2. Xin Yu Zhang and Tian Fu Wang, "Entropy- based Local Histogram Equalization for Medical Ultrasound Image Enhancement," IEEE Intl. Con! 2015 3. Ivica Dimitrovski, Pero Guguljanov and Suzana Loskovska, "Implementation of Web Based Medical Image Retrieval System in Oracle," IEEE 2nd Intl. Conference on Adaptive Science & Technology 2017. 4. H. Greenspan and A. T. Pinhas, "Medical Image categorization and retrieval for PACSusing the GMM-KL framework," IEEE Trans. 1'110.Tech Biomedicine., vol. 11, no. 2, Mar. 2017. 5. Dimitris K. Iakovidis, Nikos Pelekis, Evangelos E. Kotsifakos, Ioannis Kopanakis, Haralampos Karanikas and Yannis Theodoridis, "A Pattern Similarity Scheme for Medical Image Retrieval," IEEE Trans. Info. Tech in Biomedicine, vol 13, no. 4, Jul. 2018. 6. Hua Yuan and Xiao-Ping Zhang, "Statistical Modeling in the Wavelet Domain for Compact Feature Extraction and Similarity Measure of Images," IEEE Trans. Circuits and Systems for Video Tech., vol. 20, no. 3, Mar 2018. 7. T. M. Lehmann, M. O. Guld, C. Thies, B. Plodowski, D. Keysers, B. Ott and H. Schubeert, "IRMA - Content based image retrieval in medical applications," in Proc. 14th World Congr. Med. 1'110. (Medinfo), IDS, Amsterdam, The Netherlands, vol. 2, 2019. 8. Sharadh Ramaswamy and Kenneth Rose, "Towards Optimal Indexing for Relevance Feedback in Large Image Databases," IEEE Trans. Image Processing, vol. 18, no. 12, Dec 2019.