The document compares Gaussian filtering and bilateral filtering on RGB images. Gaussian filtering replaces each pixel with a weighted average of neighboring pixels, with weights determined by a Gaussian function of distance. This blurs the image while reducing high-frequency components. Bilateral filtering also considers pixel intensity differences, using a range Gaussian to give higher weight to similar pixels, thus preserving edges while smoothing. The effect of varying the spatial and range parameters is examined, showing bilateral filtering better retains edges through the additional range parameter.
Gaussian Fuzzy Blocking Artifacts Removal of High DCT Compressed Imagesijtsrd
A new artifact removal method as cascade of Gaussian fuzzy edge decider and fuzzy image correction is proposed. In this design, a highly compressed i.e. low bit rate image is considered. Here, each overlapped block of image is fed to a Gaussian fuzzy based decider to check whether the central pixel of image block needs correction. Hence, the central pixel of overlapped block is corrected by fuzzy gradient of its neighbors. Experimental results shows remarkable improvement with presented gFAR algorithm compared to the past methods subjectively visual quality and objectively PSNR . Deepak Gambhir "Gaussian Fuzzy Blocking Artifacts Removal of High DCT Compressed Images" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-6 , October 2020, URL: https://www.ijtsrd.com/papers/ijtsrd33361.pdf Paper Url: https://www.ijtsrd.com/computer-science/multimedia/33361/gaussian-fuzzy-blocking-artifacts-removal-of-high-dct-compressed-images/deepak-gambhir
LOCAL DISTANCE AND DEMPSTER-DHAFER FOR MULTI-FOCUS IMAGE FUSION sipij
This work proposes a new method of fusion image using Dempster-Shafer theory and local variability (DST-LV). This method takes into account the behaviour of each pixel with its neighbours. It consists in calculating the quadratic distance between the value of the pixel I (x, y) of each point and the value of all the neighbouring pixels. Local variability is used to determine the mass function defined in DempsterShafer theory. The two classes of Dempster-Shafer theory studied are : the fuzzy part and the focused part. The results of the proposed method are significantly better when comparing them to results of other methods.
This document compares three image restoration techniques - Iterated Geometric Harmonics, Markov Random Fields, and Wavelet Decomposition - for removing noise from images. It describes each technique and the process used to test them. Noise was artificially added to images using different noise generation functions. Wavelet Decomposition and Markov Random Fields were then used to detect the noise locations. These noise locations were then used to create versions of the noisy images suitable for reconstruction via Iterated Geometric Harmonics. The reconstructed images were then compared to the original to evaluate the performance of each technique.
An efficient approach to wavelet image Denoisingijcsit
This document proposes an efficient approach to wavelet image denoising based on minimizing mean squared error. It uses Stein's unbiased risk estimate (SURE), which provides an accurate estimate of mean squared error without needing the original noiseless image. The key idea is to express the thresholding function as a linear combination of thresholds, allowing the minimization problem to be solved via a simple linear system rather than a nonlinear optimization. Experimental results show the proposed method achieves superior image quality compared to other techniques like BayesShrink and VisuShrink.
This document discusses different techniques for image segmentation, which is the process of partitioning an image into meaningful regions or objects. It covers several main methods of region segmentation, including region growing, clustering, and split-and-merge. It also discusses techniques for finding line and curve segments in an image, such as using the Hough transform or edge tracking procedures. Finally, it provides examples of applying these segmentation techniques to extract regions, straight lines, and circles from images.
This document discusses image segmentation techniques, specifically linking edge points through local and global processing. Local processing involves linking edge-detected pixels that are similar in gradient strength and direction within a neighborhood. Global processing uses the Hough transform to link edge points into lines by mapping points in the image space to the parameter space of slope-intercept or polar coordinates. Thresholding in parameter space identifies coherent lines composed of edge points. The Hough transform allows finding lines even if there are gaps or other defects in detected edge points.
VARIATION-FREE WATERMARKING TECHNIQUE BASED ON SCALE RELATIONSHIPcsandit
Most watermark methods use pixel values or coefficients as the judgment condition to embed or
extract a watermark image. The variation of these values may lead to the inaccurate condition
such that an incorrect judgment has been laid out. To avoid this problem, we design a stable
judgment mechanism, in which the outcome will not be seriously influenced by the variation.
The principle of judgment depends on the scale relationship of two pixels. From the observation
of common signal processing operations, we can find that the pixel value of processed image
usually keeps stable unless an image has been manipulated by cropping attack or halftone
transformation. This can greatly help reduce the modification strength from image processing
operations. Experiment results show that the proposed method can resist various attacks and
keep the image quality friendly.
The document presents a procedure for quantifying the roughness of diamond samples at the nanoscale. It involves calculating the ratio of the total surface area of the sample to its base area using 3D calculus. The procedure approximates the surface area formula and provides 11 steps to determine roughness factor from the data. It was tested on 3 samples and produced roughness factors of 26.17, 29.98, and 5.71 respectively. The goal was to create an easy-to-use method for the Materials Research Team to evaluate nano-scale coatings.
Gaussian Fuzzy Blocking Artifacts Removal of High DCT Compressed Imagesijtsrd
A new artifact removal method as cascade of Gaussian fuzzy edge decider and fuzzy image correction is proposed. In this design, a highly compressed i.e. low bit rate image is considered. Here, each overlapped block of image is fed to a Gaussian fuzzy based decider to check whether the central pixel of image block needs correction. Hence, the central pixel of overlapped block is corrected by fuzzy gradient of its neighbors. Experimental results shows remarkable improvement with presented gFAR algorithm compared to the past methods subjectively visual quality and objectively PSNR . Deepak Gambhir "Gaussian Fuzzy Blocking Artifacts Removal of High DCT Compressed Images" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-6 , October 2020, URL: https://www.ijtsrd.com/papers/ijtsrd33361.pdf Paper Url: https://www.ijtsrd.com/computer-science/multimedia/33361/gaussian-fuzzy-blocking-artifacts-removal-of-high-dct-compressed-images/deepak-gambhir
LOCAL DISTANCE AND DEMPSTER-DHAFER FOR MULTI-FOCUS IMAGE FUSION sipij
This work proposes a new method of fusion image using Dempster-Shafer theory and local variability (DST-LV). This method takes into account the behaviour of each pixel with its neighbours. It consists in calculating the quadratic distance between the value of the pixel I (x, y) of each point and the value of all the neighbouring pixels. Local variability is used to determine the mass function defined in DempsterShafer theory. The two classes of Dempster-Shafer theory studied are : the fuzzy part and the focused part. The results of the proposed method are significantly better when comparing them to results of other methods.
This document compares three image restoration techniques - Iterated Geometric Harmonics, Markov Random Fields, and Wavelet Decomposition - for removing noise from images. It describes each technique and the process used to test them. Noise was artificially added to images using different noise generation functions. Wavelet Decomposition and Markov Random Fields were then used to detect the noise locations. These noise locations were then used to create versions of the noisy images suitable for reconstruction via Iterated Geometric Harmonics. The reconstructed images were then compared to the original to evaluate the performance of each technique.
An efficient approach to wavelet image Denoisingijcsit
This document proposes an efficient approach to wavelet image denoising based on minimizing mean squared error. It uses Stein's unbiased risk estimate (SURE), which provides an accurate estimate of mean squared error without needing the original noiseless image. The key idea is to express the thresholding function as a linear combination of thresholds, allowing the minimization problem to be solved via a simple linear system rather than a nonlinear optimization. Experimental results show the proposed method achieves superior image quality compared to other techniques like BayesShrink and VisuShrink.
This document discusses different techniques for image segmentation, which is the process of partitioning an image into meaningful regions or objects. It covers several main methods of region segmentation, including region growing, clustering, and split-and-merge. It also discusses techniques for finding line and curve segments in an image, such as using the Hough transform or edge tracking procedures. Finally, it provides examples of applying these segmentation techniques to extract regions, straight lines, and circles from images.
This document discusses image segmentation techniques, specifically linking edge points through local and global processing. Local processing involves linking edge-detected pixels that are similar in gradient strength and direction within a neighborhood. Global processing uses the Hough transform to link edge points into lines by mapping points in the image space to the parameter space of slope-intercept or polar coordinates. Thresholding in parameter space identifies coherent lines composed of edge points. The Hough transform allows finding lines even if there are gaps or other defects in detected edge points.
VARIATION-FREE WATERMARKING TECHNIQUE BASED ON SCALE RELATIONSHIPcsandit
Most watermark methods use pixel values or coefficients as the judgment condition to embed or
extract a watermark image. The variation of these values may lead to the inaccurate condition
such that an incorrect judgment has been laid out. To avoid this problem, we design a stable
judgment mechanism, in which the outcome will not be seriously influenced by the variation.
The principle of judgment depends on the scale relationship of two pixels. From the observation
of common signal processing operations, we can find that the pixel value of processed image
usually keeps stable unless an image has been manipulated by cropping attack or halftone
transformation. This can greatly help reduce the modification strength from image processing
operations. Experiment results show that the proposed method can resist various attacks and
keep the image quality friendly.
The document presents a procedure for quantifying the roughness of diamond samples at the nanoscale. It involves calculating the ratio of the total surface area of the sample to its base area using 3D calculus. The procedure approximates the surface area formula and provides 11 steps to determine roughness factor from the data. It was tested on 3 samples and produced roughness factors of 26.17, 29.98, and 5.71 respectively. The goal was to create an easy-to-use method for the Materials Research Team to evaluate nano-scale coatings.
Texture Unit based Monocular Real-world Scene Classification using SOM and KN...IDES Editor
In this paper a method is proposed to discriminate
real world scenes in to natural and manmade scenes of similar
depth. Global-roughness of a scene image varies as a function
of image-depth. Increase in image depth leads to increase in
roughness in manmade scenes; on the contrary natural scenes
exhibit smooth behavior at higher image depth. This particular
arrangement of pixels in scene structure can be well explained
by local texture information in a pixel and its neighborhood.
Our proposed method analyses local texture information of a
scene image using texture unit matrix. For final classification
we have used both supervised and unsupervised learning using
K-Nearest Neighbor classifier (KNN) and Self Organizing
Map (SOM) respectively. This technique is useful for online
classification due to very less computational complexity.
Image Interpolation Techniques with Optical and Digital Zoom Concepts -semina...mmjalbiaty
full details about Spatial and Intensity Resolution , optical and digital zoom concepts and the common three interpolation algorithms for implementing zoom in image processing
3 intensity transformations and spatial filtering slidesBHAGYAPRASADBUGGE
This document discusses basics of intensity transformations and spatial filtering of digital images. It covers the following key points:
- Intensity transformations map input pixel intensities to output intensities using an operator T. Common transformations include log, power-law, and piecewise-linear functions.
- Spatial filters operate on neighborhoods of pixels. Linear filters perform averaging or correlation while non-linear filters use ordering like median.
- Basic filters include smoothing to reduce noise, sharpening to enhance edges using Laplacian or unsharp masking, and gradient for edge detection.
- Fuzzy set theory can be applied to intensity transformations by defining membership functions for concepts like dark/bright. It can also be used for spatial filtering by defining
Qcce quality constrained co saliency estimation for common object detectionKoteswar Rao Jerripothula
Despite recent advances in joint processing of images,
sometimes it may not be as effective as single image
processing for object discovery problems. In this paper while
aiming for common object detection, we attempt to address
this problem by proposing a novel QCCE: Quality Constrained
Co-saliency Estimation method. The approach here is to iteratively
update the saliency maps through co-saliency estimation
depending upon quality scores, which indicate the degree of
separation of foreground and background likelihoods (the easier
the separation, the higher the quality of saliency map). In this
way, joint processing is automatically constrained by the quality
of saliency maps. Moreover, the proposed method can be applied
to both unsupervised and supervised scenarios, unlike other
methods which are particularly designed for one scenario only.
Experimental results demonstrate superior performance of the
proposed method compared to the state-of-the-art methods.
Improved Characters Feature Extraction and Matching Algorithm Based on SIFTNooria Sukmaningtyas
The document describes an improved SIFT feature extraction and matching algorithm based on the MSER algorithm. It first uses MSER instead of DOG to detect maximally stable elliptical regions, increasing stability and reducing the number of features. It then divides each elliptical region into fan-shaped subregions instead of square subregions, and constructs a new SIFT descriptor using Gaussian-weighted gradient information. Experimental results showed the new algorithm has affine invariance while maintaining other properties of SIFT, making it faster and better suited for real-time image processing.
This document discusses different methods of image segmentation: thresholding, edge-based segmentation, and region-based segmentation. It provides details on various thresholding techniques including basic global thresholding, Otsu's method, multiple thresholding, and variable thresholding. For edge-based segmentation, it mentions basic edge detection, the Marr-Hildreth edge detector, and watersheds. Finally, it covers region-based segmentation and provides an algorithm for region growing.
Scaling Transform Methods For Compressing a 2D Graphical image acijjournal
This document summarizes a research paper that proposes using 2D scaling transformations to compress grayscale images. It begins by defining scaling and different types of scaling transformations. It then describes how to represent 2D scaling mathematically using transformation matrices. The paper applies 2D scaling with different factors to compress the Lena test image and evaluates the compressed images using PSNR and MSE metrics. Scaling by factors of 2, 4, and 8 are tested, with higher scaling factors achieving better compression but lower image quality. In conclusions, the paper finds the proposed 2D scaling technique provides comparable or better performance than other image transformation methods for compression.
This document provides an overview of image filtering techniques in the spatial domain. It discusses smoothing filters using averaging and Gaussian weighting. It introduces first derivative filters like Sobel operators that detect edges, and second derivative filters like the Laplacian that are useful for sharpening. The Laplacian highlights edges by finding the second spatial derivative. Sharpening is done by subtracting the Laplacian from the original image. Variations are discussed.
ER Publication,
IJETR, IJMCTR,
Journals,
International Journals,
High Impact Journals,
Monthly Journal,
Good quality Journals,
Research,
Research Papers,
Research Article,
Free Journals, Open access Journals,
erpublication.org,
Engineering Journal,
Science Journals,
SinGAN - Learning a Generative Model from a Single Natural ImageJishnu P
SinGAN is a generative adversarial network (GAN) that can learn the distribution of a single natural image and generate new realistic samples from that image distribution. Unlike other GANs that require large datasets, SinGAN only needs a single image for training. It uses a multi-scale architecture with multiple generators and discriminators at different scales. SinGAN was shown to generate high quality samples for tasks like super resolution, image editing, and animation from a single image. It also has some failure cases like generating unrealistic samples at the boundaries.
Image segmentation techniques
More information on this research can be found in:
Hussein, Rania, Frederic D. McKenzie. “Identifying Ambiguous Prostate Gland Contours from Histology Using Capsule Shape Information and Least Squares Curve Fitting.” The International Journal of Computer Assisted Radiology and Surgery ( IJCARS), Volume 2 Numbers 3-4, pp. 143-150, December 2007.
Effect of Block Sizes on the Attributes of Watermarking Digital ImagesDr. Michael Agbaje
This work examines the effect of block sizes on attributes (robustness, capacity, time of watermarking, visibility and distortion) of watermarked digital images using Discrete Cosine Transform (DCT) function. The DCT function breaks up the image into various frequency bands and allows watermark data to be easily embedded. The advantage of this transformation is the ability to pack input image data into a few coefficients. The block size 8 x 8 is commonly used in watermarking. The work investigates the effect of using block sizes below and above 8 x 8 on the attributes of watermark. The attributes of robustness and capacity increase as the block size increases (62-70db, 31.5-35.9 bit/pixel). The time for watermarking reduces as the block size increases. The watermark is still visible for block sizes below 8 x 8 but invisible for those above it. Distortion decreases sharply from a high value at 2 x 2 block size to minimum at 8 x 8 and gradually increases with block size. The overall observation indicates that watermarked image gradually reduces in quality due to fading above 8 x 8 block size. For easy detection of image against piracy the block size 16 x 16 gives the best output result because it closely resembles the original image in terms of visual quality displayed despite the fact that it contains a hidden watermark.
This document describes a project that uses photometric stereo to reconstruct 3D surfaces from images taken under different lighting conditions on a computer screen. Photometric stereo uses variations in pixel intensities across images to estimate surface normals and reconstruct the 3D shape. The project creates a MATLAB program that performs photometric stereo in real-time by flashing different light patterns on a screen and capturing images with a webcam. By using singular value decomposition, the program can reconstruct surfaces without knowing the exact positions of the light sources, overcoming a limitation of traditional photometric stereo. The reconstruction contains noise but demonstrates photometric stereo in a less controlled environment. Further work could explore tradeoffs of the method and improve efficiency.
BM3D is a denoising algorithm that uses block matching and collaborative filtering in 3D transform domains. It finds similar 2D image fragments, stacks them into 3D groups, applies a 3D transformation, shrinks the transform coefficients to attenuate noise, and inverse transforms to estimate denoised fragments. BM3D improves on previous methods by exploiting inter-fragment correlation both within and between similar image blocks grouped together. It provides multiple estimates for each pixel by processing overlapping blocks, then fuses the estimates to produce the final denoised image.
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
The document proposes a new algorithm to reduce blocking artifacts in compressed images using a combination of the SAWS technique, Fuzzy Impulse Artifact Detection and Reduction Method (FIDRM), and Noise Adaptive Fuzzy Switching Median Filter (NAFSM). FIDRM uses fuzzy rules to detect noisy pixels, while NAFSM uses a median filter to correct pixels based on local information. Experimental results on test images show the proposed approach achieves better PSNR than other deblocking methods.
This document discusses methods for multichannel image filtering and classification. It describes peculiarities of multichannel radar images including different types of noise. It discusses possible approaches to multichannel data processing including component filtering and nonlinear vector filtering. Vector filtering is able to remove residual superimposing errors between images. The document also covers multichannel data classification techniques including using neural networks and processing real data to identify soil erosion states.
Bilateral filtering for gray and color imagesHarshal Ladhe
The document summarizes a research paper on bilateral filtering, which is an edge-preserving smoothing technique. Bilateral filtering smooths images while preserving edges by combining nearby pixel values based on both their geometric closeness and photometric similarity. It can smooth colors in a way that is perceptually tuned to human vision. In contrast to standard filters, bilateral filtering does not produce phantom colors along edges in color images. The paper introduces the concept of bilateral filtering and discusses its advantages over traditional filtering methods for edge-preserving smoothing of both gray-scale and color images.
Linear filters like averaging and Gaussian filters can remove grain noise by averaging pixel values in a neighborhood. Median filters are better at removing outliers without reducing sharpness by setting a pixel to the median value in its neighborhood. The document demonstrates applying averaging and median filters in Matlab to remove noise, and using morphological opening to estimate and subtract a background illumination to rectify it.
Texture Unit based Monocular Real-world Scene Classification using SOM and KN...IDES Editor
In this paper a method is proposed to discriminate
real world scenes in to natural and manmade scenes of similar
depth. Global-roughness of a scene image varies as a function
of image-depth. Increase in image depth leads to increase in
roughness in manmade scenes; on the contrary natural scenes
exhibit smooth behavior at higher image depth. This particular
arrangement of pixels in scene structure can be well explained
by local texture information in a pixel and its neighborhood.
Our proposed method analyses local texture information of a
scene image using texture unit matrix. For final classification
we have used both supervised and unsupervised learning using
K-Nearest Neighbor classifier (KNN) and Self Organizing
Map (SOM) respectively. This technique is useful for online
classification due to very less computational complexity.
Image Interpolation Techniques with Optical and Digital Zoom Concepts -semina...mmjalbiaty
full details about Spatial and Intensity Resolution , optical and digital zoom concepts and the common three interpolation algorithms for implementing zoom in image processing
3 intensity transformations and spatial filtering slidesBHAGYAPRASADBUGGE
This document discusses basics of intensity transformations and spatial filtering of digital images. It covers the following key points:
- Intensity transformations map input pixel intensities to output intensities using an operator T. Common transformations include log, power-law, and piecewise-linear functions.
- Spatial filters operate on neighborhoods of pixels. Linear filters perform averaging or correlation while non-linear filters use ordering like median.
- Basic filters include smoothing to reduce noise, sharpening to enhance edges using Laplacian or unsharp masking, and gradient for edge detection.
- Fuzzy set theory can be applied to intensity transformations by defining membership functions for concepts like dark/bright. It can also be used for spatial filtering by defining
Qcce quality constrained co saliency estimation for common object detectionKoteswar Rao Jerripothula
Despite recent advances in joint processing of images,
sometimes it may not be as effective as single image
processing for object discovery problems. In this paper while
aiming for common object detection, we attempt to address
this problem by proposing a novel QCCE: Quality Constrained
Co-saliency Estimation method. The approach here is to iteratively
update the saliency maps through co-saliency estimation
depending upon quality scores, which indicate the degree of
separation of foreground and background likelihoods (the easier
the separation, the higher the quality of saliency map). In this
way, joint processing is automatically constrained by the quality
of saliency maps. Moreover, the proposed method can be applied
to both unsupervised and supervised scenarios, unlike other
methods which are particularly designed for one scenario only.
Experimental results demonstrate superior performance of the
proposed method compared to the state-of-the-art methods.
Improved Characters Feature Extraction and Matching Algorithm Based on SIFTNooria Sukmaningtyas
The document describes an improved SIFT feature extraction and matching algorithm based on the MSER algorithm. It first uses MSER instead of DOG to detect maximally stable elliptical regions, increasing stability and reducing the number of features. It then divides each elliptical region into fan-shaped subregions instead of square subregions, and constructs a new SIFT descriptor using Gaussian-weighted gradient information. Experimental results showed the new algorithm has affine invariance while maintaining other properties of SIFT, making it faster and better suited for real-time image processing.
This document discusses different methods of image segmentation: thresholding, edge-based segmentation, and region-based segmentation. It provides details on various thresholding techniques including basic global thresholding, Otsu's method, multiple thresholding, and variable thresholding. For edge-based segmentation, it mentions basic edge detection, the Marr-Hildreth edge detector, and watersheds. Finally, it covers region-based segmentation and provides an algorithm for region growing.
Scaling Transform Methods For Compressing a 2D Graphical image acijjournal
This document summarizes a research paper that proposes using 2D scaling transformations to compress grayscale images. It begins by defining scaling and different types of scaling transformations. It then describes how to represent 2D scaling mathematically using transformation matrices. The paper applies 2D scaling with different factors to compress the Lena test image and evaluates the compressed images using PSNR and MSE metrics. Scaling by factors of 2, 4, and 8 are tested, with higher scaling factors achieving better compression but lower image quality. In conclusions, the paper finds the proposed 2D scaling technique provides comparable or better performance than other image transformation methods for compression.
This document provides an overview of image filtering techniques in the spatial domain. It discusses smoothing filters using averaging and Gaussian weighting. It introduces first derivative filters like Sobel operators that detect edges, and second derivative filters like the Laplacian that are useful for sharpening. The Laplacian highlights edges by finding the second spatial derivative. Sharpening is done by subtracting the Laplacian from the original image. Variations are discussed.
ER Publication,
IJETR, IJMCTR,
Journals,
International Journals,
High Impact Journals,
Monthly Journal,
Good quality Journals,
Research,
Research Papers,
Research Article,
Free Journals, Open access Journals,
erpublication.org,
Engineering Journal,
Science Journals,
SinGAN - Learning a Generative Model from a Single Natural ImageJishnu P
SinGAN is a generative adversarial network (GAN) that can learn the distribution of a single natural image and generate new realistic samples from that image distribution. Unlike other GANs that require large datasets, SinGAN only needs a single image for training. It uses a multi-scale architecture with multiple generators and discriminators at different scales. SinGAN was shown to generate high quality samples for tasks like super resolution, image editing, and animation from a single image. It also has some failure cases like generating unrealistic samples at the boundaries.
Image segmentation techniques
More information on this research can be found in:
Hussein, Rania, Frederic D. McKenzie. “Identifying Ambiguous Prostate Gland Contours from Histology Using Capsule Shape Information and Least Squares Curve Fitting.” The International Journal of Computer Assisted Radiology and Surgery ( IJCARS), Volume 2 Numbers 3-4, pp. 143-150, December 2007.
Effect of Block Sizes on the Attributes of Watermarking Digital ImagesDr. Michael Agbaje
This work examines the effect of block sizes on attributes (robustness, capacity, time of watermarking, visibility and distortion) of watermarked digital images using Discrete Cosine Transform (DCT) function. The DCT function breaks up the image into various frequency bands and allows watermark data to be easily embedded. The advantage of this transformation is the ability to pack input image data into a few coefficients. The block size 8 x 8 is commonly used in watermarking. The work investigates the effect of using block sizes below and above 8 x 8 on the attributes of watermark. The attributes of robustness and capacity increase as the block size increases (62-70db, 31.5-35.9 bit/pixel). The time for watermarking reduces as the block size increases. The watermark is still visible for block sizes below 8 x 8 but invisible for those above it. Distortion decreases sharply from a high value at 2 x 2 block size to minimum at 8 x 8 and gradually increases with block size. The overall observation indicates that watermarked image gradually reduces in quality due to fading above 8 x 8 block size. For easy detection of image against piracy the block size 16 x 16 gives the best output result because it closely resembles the original image in terms of visual quality displayed despite the fact that it contains a hidden watermark.
This document describes a project that uses photometric stereo to reconstruct 3D surfaces from images taken under different lighting conditions on a computer screen. Photometric stereo uses variations in pixel intensities across images to estimate surface normals and reconstruct the 3D shape. The project creates a MATLAB program that performs photometric stereo in real-time by flashing different light patterns on a screen and capturing images with a webcam. By using singular value decomposition, the program can reconstruct surfaces without knowing the exact positions of the light sources, overcoming a limitation of traditional photometric stereo. The reconstruction contains noise but demonstrates photometric stereo in a less controlled environment. Further work could explore tradeoffs of the method and improve efficiency.
BM3D is a denoising algorithm that uses block matching and collaborative filtering in 3D transform domains. It finds similar 2D image fragments, stacks them into 3D groups, applies a 3D transformation, shrinks the transform coefficients to attenuate noise, and inverse transforms to estimate denoised fragments. BM3D improves on previous methods by exploiting inter-fragment correlation both within and between similar image blocks grouped together. It provides multiple estimates for each pixel by processing overlapping blocks, then fuses the estimates to produce the final denoised image.
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
The document proposes a new algorithm to reduce blocking artifacts in compressed images using a combination of the SAWS technique, Fuzzy Impulse Artifact Detection and Reduction Method (FIDRM), and Noise Adaptive Fuzzy Switching Median Filter (NAFSM). FIDRM uses fuzzy rules to detect noisy pixels, while NAFSM uses a median filter to correct pixels based on local information. Experimental results on test images show the proposed approach achieves better PSNR than other deblocking methods.
This document discusses methods for multichannel image filtering and classification. It describes peculiarities of multichannel radar images including different types of noise. It discusses possible approaches to multichannel data processing including component filtering and nonlinear vector filtering. Vector filtering is able to remove residual superimposing errors between images. The document also covers multichannel data classification techniques including using neural networks and processing real data to identify soil erosion states.
Bilateral filtering for gray and color imagesHarshal Ladhe
The document summarizes a research paper on bilateral filtering, which is an edge-preserving smoothing technique. Bilateral filtering smooths images while preserving edges by combining nearby pixel values based on both their geometric closeness and photometric similarity. It can smooth colors in a way that is perceptually tuned to human vision. In contrast to standard filters, bilateral filtering does not produce phantom colors along edges in color images. The paper introduces the concept of bilateral filtering and discusses its advantages over traditional filtering methods for edge-preserving smoothing of both gray-scale and color images.
Linear filters like averaging and Gaussian filters can remove grain noise by averaging pixel values in a neighborhood. Median filters are better at removing outliers without reducing sharpness by setting a pixel to the median value in its neighborhood. The document demonstrates applying averaging and median filters in Matlab to remove noise, and using morphological opening to estimate and subtract a background illumination to rectify it.
Spatial domain filtering involves modifying an image by applying a filter or kernel to pixels within a neighborhood region. There are two main types of spatial filters - smoothing/low-pass filters which blur an image, and sharpening/high-pass filters which enhance edges and details. Smoothing filters replace each pixel value with the average of neighboring pixels, reducing noise. Sharpening filters use derivatives of Gaussian kernels to highlight areas of rapid intensity change, increasing contrast along edges. The effects of filtering depend on the size and shape of the kernel, with larger kernels producing more blurring or sharpening.
Spatial filtering using image processingAnuj Arora
(1) Spatial filtering is defined as operations performed on pixels within a neighborhood of an image using a mask or kernel. (2) Filters can be used to blur/smooth an image by reducing noise or sharpen an image by enhancing edges. (3) Common linear filtering methods include averaging, Gaussian, and derivative filters which are implemented using various mask patterns to modify pixels in the filtered image.
This document provides an introduction to wavelet transforms. It begins with an outline of topics to be covered, including an overview of wavelet transforms, the limitations of Fourier transforms, the historical development of wavelets, the principle of wavelet transforms, examples of applications, and references. It then discusses the stationarity of signals and how Fourier transforms cannot show when frequency components occur over time. Short-time Fourier analysis is introduced as a solution, but it is noted that wavelet transforms provide a more flexible approach by allowing the window size to vary. The document proceeds to define what a wavelet is, discuss the historical development of wavelet theory, provide examples of popular mother wavelets, and explain the steps to compute a continuous wave
This document summarizes a student project on implementing lossless discrete wavelet transform (DWT) and inverse discrete wavelet transform (IDWT). It provides an overview of the project, which includes introducing DWT, reviewing literature on lifting schemes for faster DWT computation, and simulating a 2D (5,3) DWT. The results show DWT blocks decomposing signals into high and low pass coefficients. Applications mentioned are in medical imaging, signal denoising, data compression and image processing. The conclusion discusses the need for lossless transforms in medical imaging. Future work could extend this to higher level transforms and applications like compression and watermarking.
Digital image processing img smoothningVinay Gupta
The document discusses image smoothing and sharpening techniques in digital image processing. It begins by defining what a digital image is and the goals of digital image processing. Then it discusses various applications of digital image processing like image enhancement, medical visualization, and human-computer interfaces. Key techniques covered include image smoothing using spatial filters to average pixel values in a neighborhood and image sharpening using spatial filters based on spatial differentiation to highlight edges. Examples of the Hubble space telescope and facial recognition are also mentioned.
This document discusses using bilateral filtering to extract channel structures from 3D seismic data. It begins with an introduction to bilateral filtering and its advantages over traditional Gaussian filtering. The document then provides mathematical definitions of Gaussian and bilateral filtering. It applies bilateral filtering to examples of synthetic 1D data and a slice of 3D seismic volume data to demonstrate how it can extract channel edges while preserving features. The document concludes by discussing parameter selection and computational costs of bilateral filtering.
Performance Evaluation of 2D Adaptive Bilateral Filter For Removal of Noise F...CSCJournals
In this paper, we present the performance analysis of adaptive bilateral filter by pixel to noise ratio and mean square errors. It was evaluate changing the parameters of the adaptive filter half width values and standard deviations. In adaptive bilateral filter, the edge slope is enhanced by transforming the histogram via a range filter with adaptive offset and width. The variance of range filter can also be adaptive. The filter is applied to improve the sharpens of a gray level and color image by increasing the slope of the edges without producing overshoot or undershoots. The related graphs were plotted and the best filter parameters are obtained.
An Analysis of Energy Efficient Gaussian Filter ArchitecturesIRJET Journal
This document reviews and compares different energy efficient Gaussian filter architectures. It summarizes:
1) Existing architectures approximate kernel coefficients, change window size, modify architecture, or use approximate arithmetic to reduce computation and improve energy efficiency.
2) Fixed-point and digital approximated kernels reduce implementation complexity by approximating coefficients. Low complexity filters exploit neighboring pixel similarity. Approximate filters use approximate adders.
3) Reconfigurable designs like ES-GSF achieve quality-energy tradeoffs by computing pixel values using only significant kernel boundaries based on quality and energy constraints.
4) Simulations show approximated kernels reduce quality metrics like PSNR compared to accurate filters but provide improved energy efficiency, while reconfigurable
Many applications such as robot navigation, defense, medical and remote sensing perform
various processing tasks, which can be performed more easily when all objects in different images of the
same scene are combined into a single fused image. In this paper, we propose a fast and effective
method for image fusion. The proposed method derives the intensity based variations that is large and
small scale, from the source images. In this approach, guided filtering is employed for this extraction.
Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained.
Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
NMS and Thresholding Architecture used for FPGA based Canny Edge Detector for...idescitation
In this paper, an architecture designed for Non-
Maximal Suppression used in Canny edge detection algorithm
is presented in order to reduce memory requirements
significantly. The architecture also achieves decreased latency
and increased throughput with no loss in edge detection. The
new algorithm used has a low-complexity 8-bin non-uniform
gradient magnitude histogram to compute block-based
hysteresis thresholds that are used by the Canny edge detector.
Furthermore, the hardware architecture of the proposed
algorithm is presented in this paper and the architecture is
synthesized on the Xilinx Virtex 5 FPGA. The design
development is done in VHDL and simulated results are
obtained using modelsim 6.3 with Xilinx 12.2.
This document discusses GPU-based implementations of bilateral filtering for images. Bilateral filtering smooths images while preserving edges by combining pixel values based on both geometric closeness and photometric similarity. It can be applied to color images in a way that is tuned to human color perception. A naïve bilateral filtering implementation iterates over all pixels, but it is well-suited for parallel GPU implementations due to its iterative and local nature. The document provides mathematical definitions of domain filtering, range filtering, and bilateral filtering, and notes that bilateral filtering combines the benefits of both by enforcing both geometric and photometric locality. It describes using Gaussian functions to implement the filters and discusses parameters for controlling the degree of blurring and edge preservation.
This document discusses techniques for image segmentation and edge detection. It proposes a generalized boundary detection method called Gb that combines low-level and mid-level image representations in a single eigenvalue problem to detect boundaries. Gb achieves state-of-the-art results at low computational cost. Soft segmentation is also introduced to improve boundary detection accuracy with minimal extra computation. Common methods for edge detection are described, including gradient-based, texture-based, and projection profile-based approaches. Improved Harris and corner detection algorithms are presented to more accurately detect edges and corners. The output of Gb using soft segmentations as input is shown to correlate well with occlusions and whole object boundaries while capturing general boundaries.
Performance of Efficient Closed-Form Solution to Comprehensive Frontier Exposureiosrjce
This document discusses boundary detection techniques for images. It proposes a generalized boundary detection method (Gb) that combines low-level and mid-level image representations in a single eigenvalue problem to detect boundaries. Gb achieves state-of-the-art results at low computational cost. Soft segmentation and contour grouping methods are also introduced to further improve boundary detection accuracy with minimal extra computation. The document presents outputs of Gb on sample images and concludes that Gb effectively detects boundaries in a principled manner by jointly resolving constraints from multiple image interpretation layers in closed form.
This paper discusses techniques for digital image processing, including noise reduction, edge detection, and histogram equalization. Noise reduction techniques discussed include mean, Gaussian, and median filters to remove salt and pepper noise and Gaussian noise. Edge detection algorithms like Sobel and Laplacian are introduced to reduce image data while preserving object boundaries. Histogram equalization is used for image enhancement by spreading pixel values across the full intensity range for increased contrast. The goal is recognizing objects in images through these preprocessing steps.
Invariant Recognition of Rectangular Biscuits with Fuzzy Moment Descriptors, ...CSCJournals
In this paper a new approach for invariant recognition of broken rectangular biscuits is proposed using fuzzy membership-distance products, called fuzzy moment descriptors. The existing methods for recognition of flawed rectangular biscuits are mostly based on Hough transform. However these methods are prone to error due to noise and/or variation in illumination. Fuzzy moment descriptors are less sensitive to noise thus making it an effective approach invariant to the above stray external disturbances. Further, the normalization and sorting of the moment vectors make it a size and rotation invariant recognition process .In earlier studies fuzzy moment descriptors has successfully been applied in image matching problem. In this paper the algorithm is applied in recognition of flawed and non-flawed rectangular biscuits. In general the proposed algorithm has potential applications in industrial quality control.
The document discusses several functions of image processing, including increasing the contrast of planetary details and reducing noise. It describes techniques for increasing contrast such as unsharp masking and high-pass filters. It also discusses reducing noise through techniques like averaging multiple images, low-pass filters including mean, median and mode filters, and edge detection methods. Further, it covers topics such as texture analysis, texture segmentation, texture synthesis, and deriving shape from texture.
This document summarizes a method for enhancing fingerprint images using short-time Fourier transform (STFT) analysis. The key steps are:
1. Performing STFT analysis on overlapping windows of the fingerprint image to estimate the local ridge orientation, frequency, and region mask in each window.
2. Using the estimated orientation and frequency values to filter each window of the fingerprint image in the Fourier domain for enhancement.
3. Reconstructing the enhanced fingerprint image by combining the results from each analyzed window.
This paper proposes an incremental and adaptive method for 3D reconstruction from a single RGB camera. The key features are:
1) An incremental method for updating the cost volume as new frames are added, without needing to store hundreds of comparison images. This reduces processing time and memory usage.
2) A method for dynamically adapting the minimum and maximum depth limits of the cost volume based on estimated scene depth from a semi-dense reconstruction system. This achieves optimal depth resolution.
The algorithm provides dense 3D reconstruction of indoor environments with low computational and memory costs, making it suitable for robotic applications. It is tested on both simulated and real data and shown to outperform previous volumetric reconstruction methods.
Visual Quality for both Images and Display of Systems by Visual Enhancement u...IJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
International Journal of Modern Engineering Research (IJMER) covers all the fields of engineering and science: Electrical Engineering, Mechanical Engineering, Civil Engineering, Chemical Engineering, Computer Engineering, Agricultural Engineering, Aerospace Engineering, Thermodynamics, Structural Engineering, Control Engineering, Robotics, Mechatronics, Fluid Mechanics, Nanotechnology, Simulators, Web-based Learning, Remote Laboratories, Engineering Design Methods, Education Research, Students' Satisfaction and Motivation, Global Projects, and Assessment…. And many more.
Abstract: Many applications such as robot navigation, defense, medical and remote sensing performvarious processing tasks, which can be performed more easily when all objects in different images of the same scene are combined into a single fused image. In this paper, we propose a fast and effective method for image fusion. The proposed method derives the intensity based variations that is large and small scale, from the source images. In this approach, guided filtering is employed for this extraction. Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained. Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
EDGE DETECTION IN RADAR IMAGES USING WEIBULL DISTRIBUTIONcsitconf
Radar images can reveal information about the shape of the surface terrain as well as its
physical and biophysical properties. Radar images have long been used in geological studies to
map structural features that are revealed by the shape of the landscape. Radar imagery also has
applications in vegetation and crop type mapping, landscape ecology, hydrology, and
volcanology. Image processing is using for detecting for objects in radar images. Edge
detection; which is a method of determining the discontinuities in gray level images; is a very
important initial step in Image processing. Many classical edge detectors have been developed
over time. Some of the well-known edge detection operators based on the first derivative of the
image are Roberts, Prewitt, Sobel which is traditionally implemented by convolving the image
with masks. Also Gaussian distribution has been used to build masks for the first and second
derivative. However, this distribution has limit to only symmetric shape. This paper will use to
construct the masks, the Weibull distribution which was more general than Gaussian because it
has symmetric and asymmetric shape. The constructed masks are applied to images and we
obtained good results.
EDGE DETECTION IN RADAR IMAGES USING WEIBULL DISTRIBUTIONcscpconf
Radar images can reveal information about the shape of the surface terrain as well as its physical and biophysical properties. Radar images have long been used in geological studies to
map structural features that are revealed by the shape of the landscape. Radar imagery also has applications in vegetation and crop type mapping, landscape ecology, hydrology, and
volcanology. Image processing is using for detecting for objects in radar images. Edge detection; which is a method of determining the discontinuities in gray level images; is a very
important initial step in Image processing. Many classical edge detectors have been developed over time. Some of the well-known edge detection operators based on the first derivative of the image are Roberts, Prewitt, Sobel which is traditionally implemented by convolving the image with masks. Also Gaussian distribution has been used to build masks for the first and second derivative. However, this distribution has limit to only symmetric shape. This paper will use to construct the masks, the Weibull distribution which was more general than Gaussian because it has symmetric and asymmetric shape. The constructed masks are applied to images and we obtained good results.
This document describes an image fusion method using pyramidal decomposition. It proposes extracting fine details from input images using guided filtering and fusing the base layers of images across multiple exposures or focal points using a multiresolution pyramid approach. A weight map is generated considering exposure, contrast, and saturation to guide the fusion of base layers. The fused base layer is then combined with extracted fine details to produce a detail-enhanced fused image. The goal is to preserve details in both very dark and extremely bright regions of the input images. It is argued that this method can effectively fuse images from different exposures or focal points without introducing artifacts.
An automatic algorithm for object recognition and detection based on asift ke...Kunal Kishor Nirala
This document presents an automatic algorithm for object recognition and detection based on ASIFT keypoints. The algorithm combines affine scale invariant feature transform (ASIFT) and a region merging algorithm. ASIFT is used to extract keypoints from a training image of the object. These keypoints are then used instead of user markers in a region merging algorithm to recognize and detect the object with full boundary in other images. Experimental results show the method is efficient and accurate at recognizing and detecting objects.
This document discusses edge detection in images. It begins by defining edges as significant local changes in image intensity, often occurring at boundaries between regions. The chapter then covers:
- Common edge detection steps of filtering, enhancement via gradient calculation, and detection via thresholding.
- Edge detection operators like Roberts, Sobel, and Prewitt, which approximate the image gradient to find edges.
- Using second derivatives to find zero-crossings and better locate edges compared to first derivatives.
- Evaluating different edge detectors on examples and the impact of noise and filtering.
1. Comparison of Gaussian Filtering and Bilateral Filtering on
RGB Images
Aprameyo Roy
SC14B079
Avionics
Nikunj Gupta
SC14B105
Avionics
1 Abstract
This report presents the implementation of Bilateral Filtering - a filter which is used for
smoothening an image while conserving its edges. The comparison with other filters such
as the Box and the Gaussian filters has been elucidated. There has been rigorous analysis
of all the parameters of Bilateral Filtering and their effects on the test images with and
without noise added to it. Finally an analysis was performed on the parallel and serial
implementation of the Bilateral Filter and various factors such as speed and memory used
were compared for both the implementations.
2 Introduction
An image is a 2D array of pixels. Each of this pixel can be defined either by intensity(in case of Grayscale)
or a 3D vector(in case of RGB colour). Blurring is used in pre-processing steps, such as removal of small
details from an image prior to large object extraction. The shape of an object is due to its edges. So in
blurring, we reduce the edge content in an image and try to make the transitions between different pixel
intensities as smooth as possible.
(a) Original Image (b) Image only with edges
The concept of blurring involves making adjacent pixels look similar. So, the strategy which is
employed by most of the blur filters is that of replacing a particular pixel by an average of its neighbouring
pixels. The type of average which we will take is going to define different types of filters. Here, we would
be comparing the results of three very basic but effective filters, namely, box filter, Gaussian filter and
Bilateral filter. Also, the Bilateral filter has complex computations involved(basically because it uses two
different filters). So the paper discusses an effective parallel implementation of this filter and compares
its results with the normal implementation.
1
2. 3 Box Filter
Box(Mean) filtering is a simple, intuitive and easy to implement method of smoothing images, i.e.
reducing the amount of intensity variation between one pixel and the next. The idea of box filtering is
simply to replace each pixel value in an image with the mean (average) value of its neighbors, including
itself. This has the effect of eliminating pixel values which are unrepresentative of their surroundings
(very different in intensity). Box filtering is based around a kernel, which represents the shape and size
of the neighbourhood to be sampled when calculating the mean.
The kernel used looks like a box, hence the name of the filter. Due to its property of using equal weights
it can be implemented using a much simpler accumulation algorithm. The only degree of freedom in this
filter is the window(kernel) size whose effect is seen later. The equation in effect is:
BA[I]p = q∈S Bσ(p − q)Iq
Here, BA[I]p is the result at pixel p, Bσ(p − q) is the box function, which basically depends on the
distance between the two pixels(p and q) and Iq is the intensity at pixel q. Here the summation on pixel
q goes over the entire kernel(window).
As can be seen, the problem with such an approach is that it gives inappropriate blurring at the
edges(i.e. wherever there is sudden change in intensity). Further, a single pixel with a very unrepresen-
tative value can significantly affect the mean value of all the pixels in its neighborhood. Such a filter
gives blocky results and may cause axis-aligned streaks.
3.1 Effect of Kernel(window) Size
It can be seen that as the window size is increased, the results get worse, with image getting very
blurry for window of size around 10.
The common strategy to overcome these problems is the use of a window which has a smooth falloff.
Hence, we introduce Gaussian filter which uses an isotropic(i.e. circular) window.
2
3. 4 Gaussian Filter
Gaussian filter blurs an image by implementing the Gaussian function. Here, g(x) has mean = 0 and
variance = σ2
s .
g(x) = 1√
2πσ2
s
e
− x2
2σ2
s
Since image is a 2D array, we’ll be using two dimensional Gaussian, which is the product of two such
Gaussians, one per direction:
g(x, y) = 1
2πσ2
s
e
− x2+y2
2σ2
s
where x is the distance from the origin in the horizontal axis and y is the distance from the origin in
the vertical axis. The subscript s in σs has been used to emphasize on the fact that this parameter con-
trols the space over which the Gaussian function is spread. The effect of this parameter is discussed later.
Mathematically, applying a Gaussian blur to an image is the same as convolving the image with a
Gaussian function. Since the Fourier transform of a Gaussian is another Gaussian, applying a Gaussian
filter has the effect of reducing the image’s high-frequency components, therefore, a Gaussian filter is a
low pass filter.
It takes “weighted average of pixels”, and hence, this technique is much more efficient than the box filter.
The weights, in this case, is determined by the Gaussian function. The new pixel value is governed by
the following transformation:
GB[I]p = q∈S Gσs (||p − q||)Iq
Here, Gσs
(||p − q||) is the normalized Gaussian function with mean µ = 0 and variance = σ2
s . Therefore,
the control parameters here are the window size and the sigma for the Gaussian function. Changing the
two parameter values gives different results on the images.
4.1 Effect of Window Size
Here σs = 4.
It can be seen that Gaussian Filtering gives pretty good results, even when the window size is
increased. There is not much difference in output images for window size 8 and 12 as can be seen. This
3
4. is where the parameter σs plays a crucial role. Since σs(or its square, the variance) decides the area over
which the Gaussian is spread, however large window is taken beyond a certain value, the result will be
the same. Hence, the parameter σs gives some control so that the image doesn’t become too blurred
and indistinguishable from the original image and most of the important information from the original
image is still retained.
4.2 Effect of Spatial Parameter σs
With the window size kept a constant(say 5) and σs varied, following results are obtained:
The difference between the effect of window size and spatial parameter σs can be difficult to observe
or even explain. There is just a subtle difference, which is more clear from the MATLAB code(Appendix
B). Window size defines the area over which I’m currently focussing on, with distance between each pixel
as the input. Whereas, σs defines the variance or the spread of the Gaussian over this window. Hence,
the window is always a square, but the Gaussian function controls which pixels would actually contribute
in averaging the pixel of interest. Large variance(σs) gives more spread of the Gaussian over the window,
taking more pixels to average out the pixel of interest. It is observed that for certain σs, as window size
is increased beyond a certain value, the filter won’t give a different output. This is because the spread
of the Gaussian is fixed, so however large window is chosen, the output won’t be affected because of new
far away pixels.
If someone observes very closely around the eyes of the hawk in the picture, it can be seen that it
is blurred(except when σs is very small). Filtering of the edges can remove important information from
images, especially when the images being dealt with are related to weather or satellite data or something
of the sort. Edges form the basic shape of an image(as explained in the Introduction) and if the edges
are blurred, useful information which defines the image might get lost.
Intuitively, it can be seen that if there is another parameter which can control the shape of the ker-
nel(window), better results are expected. Bilateral Filtering is an example, where a third parameter(σr)
4
5. is employed. Bilteral Filtering is discussed in the following sections.
5 Bilateral Filter
The bilateral filter is a nonlinear filter that smooths a signal while preserving strong edges. As mentioned
earlier, it is able to preserve edges because of the third parameter, σr which is called the Range Parameter.
The technique is that it doesn’t average out the pixels around the edges and hence. the kernel shape
depends on the image content(i.e. the pixels intensity). Another Gaussian function with variance σr is
used, but the random variable here is the intensity(or RGB) difference between the current pixel and
it’s neighbouring pixels. Same idea as the Gaussian filter is used, with the only difference now being my
filter multiplied with another Gaussian factor.
BF[I]p = q∈S Gσs (||p − q||)Gσr (|Ip − Iq|)Iq
The first two factors are the same as defined in Gaussian filter. The additional factor Gσr takes dif-
ference in intensity of the pixels as input, and accordingly gives weights to the output. Therefore, the
combination of Gσs
and Gσr
considers effect of only those pixels which are close in space and in range.
An additional normalization factor(Wp) is required to take care of the amplitude changes which happen
due to the two Gaussians. Hence, the modified equation is:
BF[I]p = 1
Wp q∈S Gσs
(||p − q||)Gσr
(|Ip − Iq|)Iq
The effect of window size and σs are the same as observed in Gaussian filtering and won’t be discussed
further. The main point of interest here is how the range parameter σr produces edge retention and its
comparison with output of the Gaussian filter.
5
6. 5.1 Effect of Range Parameter σr
Following results are obtained for fixed values of window size(=5) and spatial parameter(σs = 2.5):
Same effect as of varying the spatial parameter σs is observed. As σr increases, the image gets more
blurred. But the important thing to note here is that the edges are retained, or the process atleast tries
to retain those points which have sudden jump in pixel intensity. This result is further discussed when
the Gaussian filter and Bilateral filter are compared.
6 Comparison of Gaussian Filtering and Bilateral Filtering
The main between the two type of filters lies in one using an additional parameter(σr) to control the
shape of the window. Keeping the window size and spatial parameter same for both the filters, output
of the filters is taken and observed
Here window size = 5 and σs = 3.5 .
It can be observed that for σr = 0.5, bilateral filter is able to retain the edges as well(notice closely the
beak and eyes). Gaussian filter causes blur to sharp edges as well, whereas depending on the value of
range parameter(σr), good results can be obtained from Bilateral filtering in terms of edge retention.
In case of Gaussian filter, only the spatial distance between the pixel matters. Even though this falls
of gradually given some σs, it is not effective in taking into account the intensity difference between these
pixels. In other words, the kernel shape is same everywhere which is nearly circular in 2 dimensions. On
the other hand, Bilateral filter takes the intensity difference into account, converting it into Gaussian
and is ultimately multiplied with the original filter to give different shaped kernels.
One can make an obvious remark that when range parameter σr is infinity, the Bilateral filter becomes
equal to the Gaussian filter.
6
7. Effect of Gaussian Filtering and Bilateral Filtering on Noisy Images
Common signal processing applications require removing noise from images or even other signals. The
most common type of noise in the environment is the Gaussian noise(or it can be shown that large
number of noises ultimately add up to form a gaussian noise according to Central Limit Theorem). Here
performance of both Gaussian and Bilateral filtering on noisy images is observed.
The MATLAB command imnoise is used to insert Gaussian noise of certain mean and variance to the
image. Results of the two filters are shown:
Both Bilateral and Gaussian filtering give good performance against noise. By varying the parameters
of these filters, better results can be obtained. It is interesting to see that in Bilateral Filtering sharper
edges are retained and thus better filtering is performed.
7 Parallel Implementation Of Bilateral Filter
7.1 Basic Introduction To General Purpose Graphics Processing Units
We are in a time where we require very high computational ability from our processors and in just a few
years the programmable graphics processor unit has evolved into very powerful computing workhorse.
All of the programs which have been implemented have been highly optimized for a sequential run but
the speeds of execution cannot match that of the program running in a massively parallel architecture.
With multiple cores driven by very high memory bandwidth, today’s GPUs offer incredible resources for
both graphics and non-graphics processing.
7
8. The above architecture is extremely efficient in compute-intensive, highly parallel computation and
that is exactly what image processing primarily needs. It sacrifices data caching and flow control features
almost entirely and all of the transistors are devoted to processing and there is enormous amounts of
brute computational power due to this.
7.1.1 CUDA by NVIDIATM
We employ CUDA- Compute Unified Device Architecture , a recent hardware and software archi-
tecture developed by NVIDIATM
for issuing and managing computations on the GPU.
8
9. The Programming Model is as follows:
Here a kernel is a C function that, when called, is executed N times in parallel by N different CUDA
threads, as opposed to only once like regular C functions.This is the basic idea behind parallelizing a
process.
The hierarchy is as follows :
GRID consists of multiple Blocks and a block consists of multiple Threads. The threads per block
and the blocks per grid have a maximum number which can be allocated according to the device which
is being used. The device is used to refer to the GPU whereas the HOST is the CPU. The CPU is used
to control the input and initialization of the GPU and then it is used for the computation and again
returns the data back to the CPU. This is the fundamental which is used in our project to make the
bilateral filtering operation to be much more computationally effective.
The other basic fundamental includes the cudaMemcpy() command which is used to transfer data
from the HOST memory to the DEVICE memory or vice-versa. This is the only Achilles heel in the
entire process of parallel computing as this instruction takes time , but if the process is computationally
intensive then the time taken for the transfer of data is insignificant to the amount of time taken to
compute , and thus the true power of multi core processors are visible.
9
10. 7.1.2 Current Specifications
GPU
Model: Nvidia - GeForce GTX- 980m
Dedicated Video Memory : 8192 MB - GDDR5
CUDA cores : 1536
Memory Bandwidth : 160.32 GB/sec
Memory Data Rate :5010 MHz
Graphics Clock: 1038 MHz
CPU
Model : Intel i7 6700
Memory : 8 GB 2133 MHz DDR4
Clock Speed : 3.40 GHz
7.2 Parallel Implementation Of Bilateral Filtering
Filtering images can often be a highly process where each pixel in the image is affected by a given filter.
Filtering each pixel is an operation which is independent of the application of the filter to other neighbor
pixels. Since an image consists of a very large number pixels this leads it to be a good candidate for
parallelization.
Bilateral filtering is in itself a very computationally heavy process and requires more time than any
standard filtering technique. Bilateral Filtering as discussed before is an edge preserving and smoothen-
ing filter which is twice as computationally costly as the Gaussian filter.
10
11. A sequential implementation of the filter allows Bilateral Filtering on each pixel on the image, using
a kernel radius/window to define the size of local neighborhood.A basic parallelization can be done by
taking the code for the individual pixel and use that as a basis for a kernel. This kernel is then launched
with parameters which generates a thread for each pixel in the image. The pseudo-code given below
would resemble what is used to call a kernel , where gridSize and blockSize are determined by image
height and width.
d_bilateral_filter<<< gridSize, blockSize>>>(res, width, height, e_d, radius);
7.3 Performance Analysis
We now analyze the massive boost in performance which is achieved due to the import of the Bilateral
Filter program from a sequential style to a parallel mode.
The following is the result obtained from the NSIGHTTM
debugger’s performance analysis tool for the
GPU. The program has been run for ”hawk.bmp” and a window of 5 with sigma = 2.5 and delta = .5.
7.3.1 CPU AND GPU
Figure 2: Performance Analysis Summary
Figure 3: Utilization Of CPU cores
Here we can see that the complete duration taken by the program to execute completely is 4.56
seconds which is still significantly less than what a single threaded CPU would have taken. However
the major chunk of this time has been eaten away by secondary processes like the debugger , wireless
services and system processes , however the time what we need to focus on is the time taken by the GPU
to compute and also the copying of data from HOST to DEVICE and vice-versa.
7.3.2 GPU
Here we see that the duration of the computation done by the GPU only requires a mere 1.167 seconds
to complete calculation. This is extremely fast and is slowed down further due to the transport of infor-
mation from DEVICE and HOST. Further improvement in performance with respect to serial execution
is visible when the image to be filtered has higher resolution ex. 1920x1080 or more.
11
12. Figure 4: Performance Analysis Summary(GPU)
Figure 5: Function Calls And Thread Usage)
8 References
http://dsp.stackexchange.com/questions/6289/understanding-the-parameters-for-a-bilateral-filter
TOMASI, C. and MANDUCHI, R. 1998. Bilateral Filtering for Gray and Color Images. In Proceedings
of the International Conference on Computer Vision
http://www.nvidia.com/object/cuda-home-new.html.
http://en.wikipedia.org/wiki/Gaussian-blur
Bilateral Filtering with CUDA Lasse Kljgaard Staal
http://people.csail.mit.edu/sparis/siggraph07-course/
12