This document proposes a new method for fusing multiple differently exposed images into a single high dynamic range image. The method uses Fast and Adaptive Bidimensional Empirical Mode Decomposition (FABEMD) to decompose the luminance component of each input image into intrinsic mode functions (IMFs). Local energy operators are used to detect features in each IMF and determine weighting factors. The weighted IMFs are combined to form the luminance of the fused image. Color values are taken from the "best exposed" pixels of the input images. Experimental results show the method captures features from all inputs while producing uniformly exposed fused images.
Image enhancement plays an important role in vision applications. Recently a lot of work has been performed in the field of image enhancement. Many techniques have already been proposed till now for enhancing the digital images. This paper has presented a comparative analysis of various image enhancement techniques. This paper has shown that the fuzzy logic and histogram based techniques have quite effective results over the available techniques. This paper ends up with suitable future directions to enhance fuzzy based image enhancement technique further. In the proposed technique, an approach is made to enhance the images other than low-contrast images as well by balancing the stretching parameter (K) according to the color contrast. Proposed technique is designed to restore the degraded edges resulted due to contrast enhancement as well.
This document proposes a new method for multifocus image fusion that operates based on categorizing image energy levels. It calculates the energy of gradient for input images to identify focused vs. blurred regions. The images are divided into low, mid, and high energy regions using thresholds on the average energy map. Pixels are then selected from the input images for each region using different fusion rules. Experimental results on book, clock, leaf, and wafer images show the proposed method produces clearer fused images without artifacts compared to other spatial and transform domain fusion methods.
This document summarizes a research paper that proposes a novel approach for enhancing digital images using morphological operators. The approach aims to improve contrast in images with poor lighting conditions. It uses two morphological operators based on Weber's law - the first employs blocked analysis while the second uses opening by reconstruction to define a multi-background. The performance of the proposed operators is evaluated on images with various backgrounds and lighting conditions. Key steps include dividing images into blocks, estimating minimum/maximum intensities in each block to determine background criteria, and applying contrast enhancement transformations based on the criteria. Opening by reconstruction is also used to approximate image background without modifying structures. Experimental results demonstrate the approach enhances images with poor lighting.
Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...Habibur Rahman
The document proposes a modified watershed algorithm for image segmentation. It applies adaptive masking and thresholding to each color channel before combining the results. The modified algorithm is compared to FCM, RG, and HKM using metrics like PSNR, MSE, PSNRRGB, and CQM on 10 images. Results show the proposed method ensures accuracy and quality while being faster than other algorithms, making it suitable for real-time use. It performs better than the other algorithms according to visual and quantitative analysis.
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...CSCJournals
Fractal image compressions of Color Standard Lena and Satellite imageries have been carried out for the variable size range block method. The image is partitioned by considering maximum and minimum size of the range block and transforming the RGB color image into YUV form. Affine transformation and entropy coding are applied to achieve fractal compression. The Matlab simulation has been carried out for three different cases of variable range block sizes. The image is reconstructed using iterative functions and inverse transforms. The results indicate that both color Lena and Satellite imageries with R max = 16 and R min = 8, shows higher Compression ratio (CR) and good Peak Signal to Noise Ratios (PSNR). For the color standard Lena image the achievable CR~13.9 and PSNR ~25.9 dB, for Satellite rural image of CR~ 16 and PSNR ~ 23 and satellite urban image CR~16.4 and PSNR~16.5. The results of the present analysis demonstrate that, for the fractal compression scheme with variable range method applied to both color and gray scale Lena and satellite imageries, show higher CR and PSNR values compared to fixed range block size of 4 and 4 iterations. The results are presented and discussed in the paper.
This document discusses various intensity transformation and spatial filtering techniques for digital image enhancement. It covers single pixel operations like negative image and contrast stretching. It also discusses neighborhood operations such as averaging and median filters. Finally, it discusses geometric spatial transformations like scaling, rotation and translation. The document provides details on basic intensity transformation functions including log, power law, and piecewise linear transformations. It also covers histogram processing techniques like histogram equalization, matching and local histogram processing. Spatial filtering and its mechanics are explained.
A version of watershed algorithm for color image segmentationHabibur Rahman
The document summarizes a master's thesis presentation on a new watershed algorithm for color image segmentation. The thesis addresses issues with existing watershed algorithms like over-segmentation and sensitivity to noise. The contributions of the thesis include an adaptive masking and thresholding mechanism to overcome over-segmentation and perform well on noisy images. The thesis is evaluated using five image quality assessment metrics on 20 classes of images, showing the proposed method performs better and has lower computational complexity than other algorithms. In conclusions, the adaptive watershed algorithm ensures accurate segmentation and is suitable for real-time applications.
This document provides an introduction to image segmentation. It discusses how image segmentation partitions an image into meaningful regions based on measurements like greyscale, color, texture, depth, or motion. Segmentation is often an initial step in image understanding and has applications in identifying objects, guiding robots, and video compression. The document describes thresholding and clustering as two common segmentation techniques and provides examples of segmentation based on greyscale, texture, motion, depth, and optical flow. It also discusses region-growing, edge-based, and active contour model approaches to segmentation.
Image enhancement plays an important role in vision applications. Recently a lot of work has been performed in the field of image enhancement. Many techniques have already been proposed till now for enhancing the digital images. This paper has presented a comparative analysis of various image enhancement techniques. This paper has shown that the fuzzy logic and histogram based techniques have quite effective results over the available techniques. This paper ends up with suitable future directions to enhance fuzzy based image enhancement technique further. In the proposed technique, an approach is made to enhance the images other than low-contrast images as well by balancing the stretching parameter (K) according to the color contrast. Proposed technique is designed to restore the degraded edges resulted due to contrast enhancement as well.
This document proposes a new method for multifocus image fusion that operates based on categorizing image energy levels. It calculates the energy of gradient for input images to identify focused vs. blurred regions. The images are divided into low, mid, and high energy regions using thresholds on the average energy map. Pixels are then selected from the input images for each region using different fusion rules. Experimental results on book, clock, leaf, and wafer images show the proposed method produces clearer fused images without artifacts compared to other spatial and transform domain fusion methods.
This document summarizes a research paper that proposes a novel approach for enhancing digital images using morphological operators. The approach aims to improve contrast in images with poor lighting conditions. It uses two morphological operators based on Weber's law - the first employs blocked analysis while the second uses opening by reconstruction to define a multi-background. The performance of the proposed operators is evaluated on images with various backgrounds and lighting conditions. Key steps include dividing images into blocks, estimating minimum/maximum intensities in each block to determine background criteria, and applying contrast enhancement transformations based on the criteria. Opening by reconstruction is also used to approximate image background without modifying structures. Experimental results demonstrate the approach enhances images with poor lighting.
Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...Habibur Rahman
The document proposes a modified watershed algorithm for image segmentation. It applies adaptive masking and thresholding to each color channel before combining the results. The modified algorithm is compared to FCM, RG, and HKM using metrics like PSNR, MSE, PSNRRGB, and CQM on 10 images. Results show the proposed method ensures accuracy and quality while being faster than other algorithms, making it suitable for real-time use. It performs better than the other algorithms according to visual and quantitative analysis.
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...CSCJournals
Fractal image compressions of Color Standard Lena and Satellite imageries have been carried out for the variable size range block method. The image is partitioned by considering maximum and minimum size of the range block and transforming the RGB color image into YUV form. Affine transformation and entropy coding are applied to achieve fractal compression. The Matlab simulation has been carried out for three different cases of variable range block sizes. The image is reconstructed using iterative functions and inverse transforms. The results indicate that both color Lena and Satellite imageries with R max = 16 and R min = 8, shows higher Compression ratio (CR) and good Peak Signal to Noise Ratios (PSNR). For the color standard Lena image the achievable CR~13.9 and PSNR ~25.9 dB, for Satellite rural image of CR~ 16 and PSNR ~ 23 and satellite urban image CR~16.4 and PSNR~16.5. The results of the present analysis demonstrate that, for the fractal compression scheme with variable range method applied to both color and gray scale Lena and satellite imageries, show higher CR and PSNR values compared to fixed range block size of 4 and 4 iterations. The results are presented and discussed in the paper.
This document discusses various intensity transformation and spatial filtering techniques for digital image enhancement. It covers single pixel operations like negative image and contrast stretching. It also discusses neighborhood operations such as averaging and median filters. Finally, it discusses geometric spatial transformations like scaling, rotation and translation. The document provides details on basic intensity transformation functions including log, power law, and piecewise linear transformations. It also covers histogram processing techniques like histogram equalization, matching and local histogram processing. Spatial filtering and its mechanics are explained.
A version of watershed algorithm for color image segmentationHabibur Rahman
The document summarizes a master's thesis presentation on a new watershed algorithm for color image segmentation. The thesis addresses issues with existing watershed algorithms like over-segmentation and sensitivity to noise. The contributions of the thesis include an adaptive masking and thresholding mechanism to overcome over-segmentation and perform well on noisy images. The thesis is evaluated using five image quality assessment metrics on 20 classes of images, showing the proposed method performs better and has lower computational complexity than other algorithms. In conclusions, the adaptive watershed algorithm ensures accurate segmentation and is suitable for real-time applications.
This document provides an introduction to image segmentation. It discusses how image segmentation partitions an image into meaningful regions based on measurements like greyscale, color, texture, depth, or motion. Segmentation is often an initial step in image understanding and has applications in identifying objects, guiding robots, and video compression. The document describes thresholding and clustering as two common segmentation techniques and provides examples of segmentation based on greyscale, texture, motion, depth, and optical flow. It also discusses region-growing, edge-based, and active contour model approaches to segmentation.
Abstract
Field of image processing has vast applications in medical, forensic, research etc., It includes various domains like enhancement,
classification, segmentation, etc., which are widely used for these applications. Image Enhancement is the pre processing step on
which the accuracy of the result lies. Image enhancement aims to improve the visual appearance of an image, without affecting
the original attributes (i.e.,) image contrast is adjusted and noise is removed to produce better quality image. Hence image
enhancement is one of the most important tasks in image processing. Enhancement is classified into two categories spatial domain
enhancement and frequency domain enhancement. Spatial domain enhancement acts upon pixel value whereas frequency domain
enhancement acts on the Fourier transform of the image. The enhancement techniques to be used depend on modality, climatic
and visual perspective etc., In this paper, we present a survey on various existing image enhancement techniques.
Keywords: Enhancement, Spatial domain enhancement, Frequency domain enhancement, Contrast, Modality.
This document summarizes a paper presented at the 2nd International Conference on Current Trends in Engineering and Management. The paper proposes using discrete wavelet transform techniques for pixel-based fusion of multi-focus images. It discusses registering the images and then applying pixel-level fusion methods like average, minimum and maximum approaches. It also introduces a wavelet-based fusion method that decomposes images into different frequency bands for fusion. The goal is to produce a single fused image that has the maximum information and focus from the input images.
An Inclusive Analysis on Various Image Enhancement TechniquesIJMER
The document discusses various techniques for image enhancement, which is the process of improving the visual quality of digital images. It describes several techniques including filtering with morphological operators, histogram equalization, noise removal using a Wiener filter, linear contrast enhancement, median filtering, unsharp mask filtering, contrast-limited adaptive histogram equalization, and decorrelation stretch. These techniques can be used to enhance images by reducing noise, increasing contrast, sharpening edges, and highlighting subtle differences to improve interpretability. The choice of enhancement technique depends on the image content and intended application.
Abstract: Many applications such as robot navigation, defense, medical and remote sensing performvarious processing tasks, which can be performed more easily when all objects in different images of the same scene are combined into a single fused image. In this paper, we propose a fast and effective method for image fusion. The proposed method derives the intensity based variations that is large and small scale, from the source images. In this approach, guided filtering is employed for this extraction. Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained. Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
This document summarizes a research paper that proposes a new technique for binarizing images captured of black/green boards using a mobile camera. It begins with an abstract that overviews binarizing degraded images from mobile-captured black/green board images to extract text with 92.589% accuracy. It then reviews existing binarization techniques in the literature and describes common global and local thresholding methods. The proposed technique enhances the input image, segments it into 3x3 parts, computes local thresholds using OTSU for each part, binarizes the parts, and joins them. Experimental results on a database of 50 mobile-captured board images show the technique achieves better accuracy than other algorithms according to evaluation metrics.
COM2304: Intensity Transformation and Spatial Filtering – II Spatial Filterin...Hemantha Kulathilake
At the end of this lecture, you should be able to;
describe the fundamentals of spatial filtering.
generating spatial filter masks.
identify smoothing via linear filters and non linear filters.
apply smoothing techniques for problem solving.
Quality Assessment of Gray and Color Images through Image Fusion TechniqueIJEEE
. Image fusion is an emerging trend in the digital image processing to enhance images. In image fusion two or more images can be fused (combined) to obtain an enhanced image. In the present work image fusion technology has been used to enhance a given input image. Image fusion is used here to combine two images which contains complementary information.
Image pre-processing involves operations on images to improve image data by suppressing distortions or enhancing features. There are four categories of pre-processing methods based on pixel neighborhood size used: pixel brightness transformations, geometric transformations, local neighborhood methods, and global image restoration. Pre-processing aims to correct degradations by using prior knowledge about the degradation, image acquisition device, or objects in the image. Common pre-processing methods include brightness and geometric transformations as well as brightness interpolation when re-sampling images.
//STEIM Workshop: A Vernacular of File FormatsRosa ɯǝukɯɐn
This document provides definitions and explanations of various file formats and compression techniques, including:
- Lossless vs lossy data compression
- Raster images and how color depth affects stored color values
- Image glitching techniques like imagebending and databending
- How bytes and binary files are structured
- Components of image files like headers, channels, and interleaved vs non-interleaved formats
- Explanations of formats like BMP, GIF, PNG, PSD, JPG, TIFF, and others
IMAGE ENHANCEMENT IN CASE OF UNEVEN ILLUMINATION USING VARIABLE THRESHOLDING ...ijsrd.com
Uneven illumination always affects the visual quality images which results in poor understanding about the content of the images. There is no accepted universal image enhancement algorithm or specific criteria which can fulfill user needs. The processed image may be very different with the original image in the visual effects, but it also may be similar to the original image [1]. It will be a developing tradition to integrate the advantage of various algorithms to practical application to image enhancements [2]. Zhang et al. [3] presents an adaptive image contrast enhancement method. The proposed method is based on a local gamma correction piloted by histogram analysis. In this paper , to avoid uneven Illuminance image is divided into different segments . It works locally to decrease contrast as if we perform enhancement techniques globally on portions which are already bright then this gives poor results. Enhancement techniques are applied only to those dark portions. We need accurate method that not only enhance the image but also preserve the information.
Abstract Image Segmentation plays a vital role in image processing. The research in this area is still relevant due to its wide applications. Image segmentation is a process of assigning a label to every pixel in an image such that pixels with same label share certain visual characteristics. Sometimes it becomes necessary to calculate the total number of colors from the given RGB image to quantize the image, to detect cancer and brain tumour. The goal of this paper is to provide the best algorithm for image segmentation. Keywords: Image segmentation, RGB
The document discusses various image enhancement techniques in digital image processing. It describes point operations like image negative, contrast stretching, thresholding, brightness enhancement, log transformation, and power law transformation. Contrast stretching expands the range of intensity levels and can be done by multiplying pixels with a constant, using a transfer function, or histogram equalization. Thresholding converts an image to binary by assigning pixel values above a threshold to one level and below to another. Log and power law transformations compress high intensity values and expand low values to enhance an image. Matlab code examples are provided for each technique.
Feature Extraction of an Image by Using Adaptive Filtering and Morpological S...IOSR Journals
Abstract: For enhancing an image various enhancement schemes are used which includes gray scale manipulation, filtering and Histogram Equalization, Where Histogram equalization is one of the well known image enhancement technique. It became a popular technique for contrast enhancement because it is simple and effective. The basic idea of Histogram Equalization method is to remap the gray levels of an image. Here using morphological segmentation we can get the segmented image. Morphological reconstruction is used to segment the image. Comparative analysis of different enhancement and segmentation will be carried out. This comparison will be done on the basis of subjective and objective parameters. Subjective parameter is visual quality and objective parameters are Area, Perimeter, Min and Max intensity, Avg Voxel Intensity, Std Dev of Intensity, Eccentricity, Coefficient of skewness, Coefficient of Kurtosis, Median intensity, Mode intensity. Keywords: Histogram Equalization, Segmentation, Morphological Reconstruction .
Feature selection is the fundamental step in image
registration. Various tasks such as feature extraction, detection
are based on feature based approach. In the current paper we are
going to discuss about our technique that is hybrid of Local affine
and thin plate spline. An automatic edge detection method to
achieve the correct edge map is put forward to dealing with
image registration with affine transformation for the better
image registration. Registration algorithms compute
transformations to set correspondence between the two images.
The purpose of this paper is to provide a comprehensive
comparison of the existing literature available on Image
registration methods with proposed technique
This document discusses image enhancement techniques in digital image processing. It defines image enhancement as modifying image attributes to make an image more suitable for a given task. The main techniques discussed are spatial domain enhancement methods like noise removal, contrast adjustment, and histogram equalization. Examples are provided to demonstrate the effects of these enhancement methods on images.
User Interactive Color Transformation between ImagesIJMER
Abstract: In this paper we present a process called color
transfer which can borrow one image’s color
characteristics from another. Most current colorization
algorithms either require a significant user effort or have
large computational time. Here focus on orthogonal color
space i.e. lαβ color space without correlation between the
axes is given. Here we have implemented two global color
transfer algorithms in lαβ color space using simple color
statistical information such as mean, standard deviation
and covariance between the pixels of image. Our approach
is the extension of Reinhard's. Our local color transfer
algorithm uses simple color statistical analysis to recolor
the target image according to selected color range in
source image. Target image’s color influence mask is
prepared. It is a mask that specifies what parts of target
image will be affected according to selected color range.
After that target image is recolored in lαβ color space
according to prepared color influence map. In the lαβ
color space luminance and chrominance information is
separate so it allows making image recoloring optional.
The basic color transformation uses stored color statistics
of source and target image. All the algorithms are
implemented in JAVA object oriented language. The main
advantage of proposed method over the existing one is it
allows the user to recolor a part of the image in a simple &
intuitive way, preserving other color intact & achieving
natural look.
Index Terms: color transfer, local color statistics, color
characteristics, orthogonal color space, color influence
map.
This document discusses different types of gray level transformations that are commonly used in image processing. It describes three main types of transformations: linear, logarithmic, and power-law transformations. Linear transformations include identity and negative transformations. Logarithmic transformations include log and inverse log transformations. Power-law transformations include nth power and nth root transformations which are also known as gamma transformations, where the gamma value determines whether darker or brighter images are produced. Examples of transformations with different gamma values are also shown.
This document discusses methods for enhancing spatial features in images. It describes spatial filtering and convolution, which involve applying kernels or filters to images to emphasize different spatial frequencies. Low-pass filters smooth images by reducing high spatial frequencies related to fine detail, while high-pass filters have the opposite effect of sharpening images. Edge enhancement filters aim to highlight linear features by increasing contrast around edges. Fourier analysis represents images as combinations of sine and cosine waves of different frequencies and can reveal features aligned in various directions.
Jpeg image compression using discrete cosine transform a surveyIJCSES Journal
Due to the increasing requirements for transmission of images in computer, mobile environments, the
research in the field of image compression has increased significantly. Image compression plays a crucial
role in digital image processing, it is also very important for efficient transmission and storage of images.
When we compute the number of bits per image resulting from typical sampling rates and quantization
methods, we find that Image compression is needed. Therefore development of efficient techniques for
image compression has become necessary .This paper is a survey for lossy image compression using
Discrete Cosine Transform, it covers JPEG compression algorithm which is used for full-colour still image
applications and describes all the components of it.
Stereo matching based on absolute differences for multiple objects detectionTELKOMNIKA JOURNAL
This article presents a new algorithm for object detection using stereo camera system. The problem to get an accurate object detion using stereo camera is the imprecise of matching process between two scenes with the same viewpoint. Hence, this article aims to reduce the incorrect matching pixel with four stages. This new algorithm is the combination of continuous process of matching cost computation, aggregation, optimization and filtering. The first stage is matching cost computation to acquire preliminary result using an absolute differences method. Then the second stage known as aggregation step uses a guided filter with fixed window support size. After that, the optimization stage uses winner-takes-all (WTA) approach which selects the smallest matching differences value and normalized it to the disparity level. The last stage in the framework uses a bilateral filter. It is effectively further decrease the error on the disparity map which contains information of object detection and locations. The proposed work produces low errors (i.e., 12.11% and 14.01% nonocc and all errors) based on the KITTI dataset and capable to perform much better compared with before the proposed framework and competitive with some newly available methods.
Stereo matching algorithm using census transform and segment tree for depth e...TELKOMNIKA JOURNAL
This article proposes an algorithm for stereo matching corresponding process that will be used in many applications such as augmented reality, autonomous vehicle navigation and surface reconstruction. Basically, the proposed framework in this article is developed through a series of functions. The final result from this framework is disparity map which this map has the information of depth estimation. Fundamentally, the framework input is the stereo image which represents left and right images respectively. The proposed algorithm in this article has four steps in total, which starts with the matching cost computation using census transform, cost aggregation utilizes segment-tree, optimization using winner-takes-all (WTA) strategy, and post-processing stage uses weighted median filter. Based on the experimental results from the standard benchmarking evaluation system from the Middlebury, the disparity map results produce an average low noise error at 9.68% for nonocc error and 18.9% for all error attributes. On average, it performs far better and very competitive with other available methods from the benchmark system.
Abstract
Field of image processing has vast applications in medical, forensic, research etc., It includes various domains like enhancement,
classification, segmentation, etc., which are widely used for these applications. Image Enhancement is the pre processing step on
which the accuracy of the result lies. Image enhancement aims to improve the visual appearance of an image, without affecting
the original attributes (i.e.,) image contrast is adjusted and noise is removed to produce better quality image. Hence image
enhancement is one of the most important tasks in image processing. Enhancement is classified into two categories spatial domain
enhancement and frequency domain enhancement. Spatial domain enhancement acts upon pixel value whereas frequency domain
enhancement acts on the Fourier transform of the image. The enhancement techniques to be used depend on modality, climatic
and visual perspective etc., In this paper, we present a survey on various existing image enhancement techniques.
Keywords: Enhancement, Spatial domain enhancement, Frequency domain enhancement, Contrast, Modality.
This document summarizes a paper presented at the 2nd International Conference on Current Trends in Engineering and Management. The paper proposes using discrete wavelet transform techniques for pixel-based fusion of multi-focus images. It discusses registering the images and then applying pixel-level fusion methods like average, minimum and maximum approaches. It also introduces a wavelet-based fusion method that decomposes images into different frequency bands for fusion. The goal is to produce a single fused image that has the maximum information and focus from the input images.
An Inclusive Analysis on Various Image Enhancement TechniquesIJMER
The document discusses various techniques for image enhancement, which is the process of improving the visual quality of digital images. It describes several techniques including filtering with morphological operators, histogram equalization, noise removal using a Wiener filter, linear contrast enhancement, median filtering, unsharp mask filtering, contrast-limited adaptive histogram equalization, and decorrelation stretch. These techniques can be used to enhance images by reducing noise, increasing contrast, sharpening edges, and highlighting subtle differences to improve interpretability. The choice of enhancement technique depends on the image content and intended application.
Abstract: Many applications such as robot navigation, defense, medical and remote sensing performvarious processing tasks, which can be performed more easily when all objects in different images of the same scene are combined into a single fused image. In this paper, we propose a fast and effective method for image fusion. The proposed method derives the intensity based variations that is large and small scale, from the source images. In this approach, guided filtering is employed for this extraction. Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained. Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
This document summarizes a research paper that proposes a new technique for binarizing images captured of black/green boards using a mobile camera. It begins with an abstract that overviews binarizing degraded images from mobile-captured black/green board images to extract text with 92.589% accuracy. It then reviews existing binarization techniques in the literature and describes common global and local thresholding methods. The proposed technique enhances the input image, segments it into 3x3 parts, computes local thresholds using OTSU for each part, binarizes the parts, and joins them. Experimental results on a database of 50 mobile-captured board images show the technique achieves better accuracy than other algorithms according to evaluation metrics.
COM2304: Intensity Transformation and Spatial Filtering – II Spatial Filterin...Hemantha Kulathilake
At the end of this lecture, you should be able to;
describe the fundamentals of spatial filtering.
generating spatial filter masks.
identify smoothing via linear filters and non linear filters.
apply smoothing techniques for problem solving.
Quality Assessment of Gray and Color Images through Image Fusion TechniqueIJEEE
. Image fusion is an emerging trend in the digital image processing to enhance images. In image fusion two or more images can be fused (combined) to obtain an enhanced image. In the present work image fusion technology has been used to enhance a given input image. Image fusion is used here to combine two images which contains complementary information.
Image pre-processing involves operations on images to improve image data by suppressing distortions or enhancing features. There are four categories of pre-processing methods based on pixel neighborhood size used: pixel brightness transformations, geometric transformations, local neighborhood methods, and global image restoration. Pre-processing aims to correct degradations by using prior knowledge about the degradation, image acquisition device, or objects in the image. Common pre-processing methods include brightness and geometric transformations as well as brightness interpolation when re-sampling images.
//STEIM Workshop: A Vernacular of File FormatsRosa ɯǝukɯɐn
This document provides definitions and explanations of various file formats and compression techniques, including:
- Lossless vs lossy data compression
- Raster images and how color depth affects stored color values
- Image glitching techniques like imagebending and databending
- How bytes and binary files are structured
- Components of image files like headers, channels, and interleaved vs non-interleaved formats
- Explanations of formats like BMP, GIF, PNG, PSD, JPG, TIFF, and others
IMAGE ENHANCEMENT IN CASE OF UNEVEN ILLUMINATION USING VARIABLE THRESHOLDING ...ijsrd.com
Uneven illumination always affects the visual quality images which results in poor understanding about the content of the images. There is no accepted universal image enhancement algorithm or specific criteria which can fulfill user needs. The processed image may be very different with the original image in the visual effects, but it also may be similar to the original image [1]. It will be a developing tradition to integrate the advantage of various algorithms to practical application to image enhancements [2]. Zhang et al. [3] presents an adaptive image contrast enhancement method. The proposed method is based on a local gamma correction piloted by histogram analysis. In this paper , to avoid uneven Illuminance image is divided into different segments . It works locally to decrease contrast as if we perform enhancement techniques globally on portions which are already bright then this gives poor results. Enhancement techniques are applied only to those dark portions. We need accurate method that not only enhance the image but also preserve the information.
Abstract Image Segmentation plays a vital role in image processing. The research in this area is still relevant due to its wide applications. Image segmentation is a process of assigning a label to every pixel in an image such that pixels with same label share certain visual characteristics. Sometimes it becomes necessary to calculate the total number of colors from the given RGB image to quantize the image, to detect cancer and brain tumour. The goal of this paper is to provide the best algorithm for image segmentation. Keywords: Image segmentation, RGB
The document discusses various image enhancement techniques in digital image processing. It describes point operations like image negative, contrast stretching, thresholding, brightness enhancement, log transformation, and power law transformation. Contrast stretching expands the range of intensity levels and can be done by multiplying pixels with a constant, using a transfer function, or histogram equalization. Thresholding converts an image to binary by assigning pixel values above a threshold to one level and below to another. Log and power law transformations compress high intensity values and expand low values to enhance an image. Matlab code examples are provided for each technique.
Feature Extraction of an Image by Using Adaptive Filtering and Morpological S...IOSR Journals
Abstract: For enhancing an image various enhancement schemes are used which includes gray scale manipulation, filtering and Histogram Equalization, Where Histogram equalization is one of the well known image enhancement technique. It became a popular technique for contrast enhancement because it is simple and effective. The basic idea of Histogram Equalization method is to remap the gray levels of an image. Here using morphological segmentation we can get the segmented image. Morphological reconstruction is used to segment the image. Comparative analysis of different enhancement and segmentation will be carried out. This comparison will be done on the basis of subjective and objective parameters. Subjective parameter is visual quality and objective parameters are Area, Perimeter, Min and Max intensity, Avg Voxel Intensity, Std Dev of Intensity, Eccentricity, Coefficient of skewness, Coefficient of Kurtosis, Median intensity, Mode intensity. Keywords: Histogram Equalization, Segmentation, Morphological Reconstruction .
Feature selection is the fundamental step in image
registration. Various tasks such as feature extraction, detection
are based on feature based approach. In the current paper we are
going to discuss about our technique that is hybrid of Local affine
and thin plate spline. An automatic edge detection method to
achieve the correct edge map is put forward to dealing with
image registration with affine transformation for the better
image registration. Registration algorithms compute
transformations to set correspondence between the two images.
The purpose of this paper is to provide a comprehensive
comparison of the existing literature available on Image
registration methods with proposed technique
This document discusses image enhancement techniques in digital image processing. It defines image enhancement as modifying image attributes to make an image more suitable for a given task. The main techniques discussed are spatial domain enhancement methods like noise removal, contrast adjustment, and histogram equalization. Examples are provided to demonstrate the effects of these enhancement methods on images.
User Interactive Color Transformation between ImagesIJMER
Abstract: In this paper we present a process called color
transfer which can borrow one image’s color
characteristics from another. Most current colorization
algorithms either require a significant user effort or have
large computational time. Here focus on orthogonal color
space i.e. lαβ color space without correlation between the
axes is given. Here we have implemented two global color
transfer algorithms in lαβ color space using simple color
statistical information such as mean, standard deviation
and covariance between the pixels of image. Our approach
is the extension of Reinhard's. Our local color transfer
algorithm uses simple color statistical analysis to recolor
the target image according to selected color range in
source image. Target image’s color influence mask is
prepared. It is a mask that specifies what parts of target
image will be affected according to selected color range.
After that target image is recolored in lαβ color space
according to prepared color influence map. In the lαβ
color space luminance and chrominance information is
separate so it allows making image recoloring optional.
The basic color transformation uses stored color statistics
of source and target image. All the algorithms are
implemented in JAVA object oriented language. The main
advantage of proposed method over the existing one is it
allows the user to recolor a part of the image in a simple &
intuitive way, preserving other color intact & achieving
natural look.
Index Terms: color transfer, local color statistics, color
characteristics, orthogonal color space, color influence
map.
This document discusses different types of gray level transformations that are commonly used in image processing. It describes three main types of transformations: linear, logarithmic, and power-law transformations. Linear transformations include identity and negative transformations. Logarithmic transformations include log and inverse log transformations. Power-law transformations include nth power and nth root transformations which are also known as gamma transformations, where the gamma value determines whether darker or brighter images are produced. Examples of transformations with different gamma values are also shown.
This document discusses methods for enhancing spatial features in images. It describes spatial filtering and convolution, which involve applying kernels or filters to images to emphasize different spatial frequencies. Low-pass filters smooth images by reducing high spatial frequencies related to fine detail, while high-pass filters have the opposite effect of sharpening images. Edge enhancement filters aim to highlight linear features by increasing contrast around edges. Fourier analysis represents images as combinations of sine and cosine waves of different frequencies and can reveal features aligned in various directions.
Jpeg image compression using discrete cosine transform a surveyIJCSES Journal
Due to the increasing requirements for transmission of images in computer, mobile environments, the
research in the field of image compression has increased significantly. Image compression plays a crucial
role in digital image processing, it is also very important for efficient transmission and storage of images.
When we compute the number of bits per image resulting from typical sampling rates and quantization
methods, we find that Image compression is needed. Therefore development of efficient techniques for
image compression has become necessary .This paper is a survey for lossy image compression using
Discrete Cosine Transform, it covers JPEG compression algorithm which is used for full-colour still image
applications and describes all the components of it.
Stereo matching based on absolute differences for multiple objects detectionTELKOMNIKA JOURNAL
This article presents a new algorithm for object detection using stereo camera system. The problem to get an accurate object detion using stereo camera is the imprecise of matching process between two scenes with the same viewpoint. Hence, this article aims to reduce the incorrect matching pixel with four stages. This new algorithm is the combination of continuous process of matching cost computation, aggregation, optimization and filtering. The first stage is matching cost computation to acquire preliminary result using an absolute differences method. Then the second stage known as aggregation step uses a guided filter with fixed window support size. After that, the optimization stage uses winner-takes-all (WTA) approach which selects the smallest matching differences value and normalized it to the disparity level. The last stage in the framework uses a bilateral filter. It is effectively further decrease the error on the disparity map which contains information of object detection and locations. The proposed work produces low errors (i.e., 12.11% and 14.01% nonocc and all errors) based on the KITTI dataset and capable to perform much better compared with before the proposed framework and competitive with some newly available methods.
Stereo matching algorithm using census transform and segment tree for depth e...TELKOMNIKA JOURNAL
This article proposes an algorithm for stereo matching corresponding process that will be used in many applications such as augmented reality, autonomous vehicle navigation and surface reconstruction. Basically, the proposed framework in this article is developed through a series of functions. The final result from this framework is disparity map which this map has the information of depth estimation. Fundamentally, the framework input is the stereo image which represents left and right images respectively. The proposed algorithm in this article has four steps in total, which starts with the matching cost computation using census transform, cost aggregation utilizes segment-tree, optimization using winner-takes-all (WTA) strategy, and post-processing stage uses weighted median filter. Based on the experimental results from the standard benchmarking evaluation system from the Middlebury, the disparity map results produce an average low noise error at 9.68% for nonocc error and 18.9% for all error attributes. On average, it performs far better and very competitive with other available methods from the benchmark system.
Development of stereo matching algorithm based on sum of absolute RGB color d...IJECEIAES
This article presents local-based stereo matching algorithm which comprises a devel- opment of an algorithm using block matching and two edge preserving filters in the framework. Fundamentally, the matching process consists of several stages which will produce the disparity or depth map. The problem and most challenging work for matching process is to get an accurate corresponding point between two images. Hence, this article proposes an algorithm for stereo matching using improved Sum of Absolute RGB Differences (SAD), gradient matching and edge preserving filters. It is Bilateral Filter (BF) to surge up the accuracy. The SAD and gradient matching will be implemented at the first stage to get the preliminary corresponding result, then the BF works as an edge-preserving filter to remove the noise from the first stage. The second BF is used at the last stage to improve final disparity map and increase the object boundaries. The experimental analysis and validation are using the Middlebury standard benchmarking evaluation system. Based on the results, the proposed work is capable to increase the accuracy and to preserve the object edges. To make the proposed work more reliable with current available methods, the quantitative measurement has been made to compare with other existing methods and it shows the proposed work in this article perform much better.
THE REDUCTION IN COMPUTATION TIME IN THE COLOR IMAGE DECOMPOSITION WITH THE B...IJCSEIT Journal
This paper presents two different approaches in color image decomposition domain with Bidimensional
Empirical Mode Decomposition (BEMD). The first approach applies the BEMD on each channel
separately and the second is based on interpolation of each channel in the sifting process. The comparison
of two approaches shows the same performance of each approach in terms of visual quality, but they do not
provide the same results in execution time which presents the most important criteron in real time
applications. It was shown that the second BEMD approach based on interpolation of each channel in the
sifting process, gives a gain in the point of view the execution time.
Towards better performance: phase congruency based face recognitionTELKOMNIKA JOURNAL
Phase congruency is an edge detector and measurement of the significant feature in the image. It is a robust method against contrast and illumination variation. In this paper, two novel techniques are introduced for developing alow-cost human identification system based on face recognition. Firstly, the valuable phase congruency features, the gradient-edges and their associate dangles are utilized separately for classifying 130 subjects taken from three face databases with the motivation of eliminating the feature extraction phase. By doing this, the complexity can be significantly reduced. Secondly, the training process is modified when a new technique, called averaging-vectors is developed to accelerate the training process and minimizes the matching time to the lowest value. However, for more comparison and accurate evaluation,three competitive classifiers: Euclidean distance (ED),cosine distance (CD), and Manhattan distance (MD) are considered in this work. The system performance is very competitive and acceptable, where the experimental results show promising recognition rates with a reasonable matching time.
The document presents a new efficient color image compression technique that aims to improve the quality of decompressed images while achieving higher compression ratios. It does this by compressing important edge parts of the image differently than non-edge background parts. Specifically, it applies low-quality lossy compression to non-edge parts and high-quality lossy compression to edge parts. The technique uses edge detection, adaptive thresholding based on local variance and mean, and discrete cosine transform followed by quantization and entropy encoding. Experimental results on various images show it achieves better compression ratios, lower bit rates, and higher peak signal to noise ratios compared to non-adaptive methods.
Matching algorithm performance analysis for autocalibration method of stereo ...TELKOMNIKA JOURNAL
Stereo vision is one of the interesting research topics in the computer vision field. Two cameras are used to generate a disparity map, resulting in the depth estimation. Camera calibration is the most important step in stereo vision. The calibration step is used to generate an intrinsic parameter of each camera to get a better disparity map. In general, the calibration process is done manually by using a chessboard pattern, but this process is an exhausting task. Self-calibration is an important ability required to overcome this problem. Self-calibration required a robust and good matching algorithm to find the key feature between images as reference. The purpose of this paper is to analyze the performance of three matching algorithms for the autocalibration process. The matching algorithms used in this research are SIFT, SURF, and ORB. The result shows that SIFT performs better than other methods.
Design and implementation of video tracking system based on camera field of viewsipij
The basic idea of this paper is to design and implement of video tracking system based on Camera Field of
View (CFOV), Otsu’s method was used to detect targets such as vehicles and people. Whereas most
algorithms were spent a lot of time to execute the process, an algorithm was developed to achieve it in a
little time. The histogram projection was used in both directional to detect target from search region,
which is robust to various light conditions in Charge Couple Device (CCD) camera images and saves
computation time.
Our algorithm based on background subtraction, and normalize cross correlation operation from a series
of sequential sub images can estimate the motion vector. Camera field of view (CFOV) was determined and
calibrated to find the relation between real distance and image distance. The system was tested by
measuring the real position of object in the laboratory and compares it with the result of computed one. So
these results are promising to develop the system in future.
This paper proposes an incremental and adaptive method for 3D reconstruction from a single RGB camera. The key features are:
1) An incremental method for updating the cost volume as new frames are added, without needing to store hundreds of comparison images. This reduces processing time and memory usage.
2) A method for dynamically adapting the minimum and maximum depth limits of the cost volume based on estimated scene depth from a semi-dense reconstruction system. This achieves optimal depth resolution.
The algorithm provides dense 3D reconstruction of indoor environments with low computational and memory costs, making it suitable for robotic applications. It is tested on both simulated and real data and shown to outperform previous volumetric reconstruction methods.
Wavelet based Image Coding Schemes: A Recent Survey ijsc
A variety of new and powerful algorithms have been developed for image compression over the years. Among them the wavelet-based image compression schemes have gained much popularity due to their overlapping nature which reduces the blocking artifacts that are common phenomena in JPEG compression and multiresolution character which leads to superior energy compaction with high quality reconstructed images. This paper provides a detailed survey on some of the popular wavelet coding techniques such as the Embedded Zerotree Wavelet (EZW) coding, Set Partitioning in Hierarchical Tree (SPIHT) coding, the Set Partitioned Embedded Block (SPECK) Coder, and the Embedded Block Coding with Optimized Truncation (EBCOT) algorithm. Other wavelet-based coding techniques like the Wavelet Difference Reduction (WDR) and the Adaptive Scanned Wavelet Difference Reduction (ASWDR) algorithms, the Space Frequency Quantization (SFQ) algorithm, the Embedded Predictive Wavelet Image Coder (EPWIC), Compression with Reversible Embedded Wavelet (CREW), the Stack-Run (SR) coding and the recent Geometric Wavelet (GW) coding are also discussed. Based on the review, recommendations and discussions are presented for algorithm development and implementation.
Development of depth map from stereo images using sum of absolute differences...nooriasukmaningtyas
This article proposes a framework for the depth map reconstruction using stereo images. Fundamentally, this map provides an important information which commonly used in essential applications such as autonomous vehicle navigation, drone’s navigation and 3D surface reconstruction. To develop an accurate depth map, the framework must be robust against the challenging regions of low texture, plain color and repetitive pattern on the input stereo image. The development of this map requires several stages which starts with matching cost calculation, cost aggregation, optimization and refinement stage. Hence, this work develops a framework with sum of absolute difference (SAD) and the combination of two edge preserving filters to increase the robustness against the challenging regions. The SAD convolves using block matching technique to increase the efficiency of matching process on the low texture and plain color regions. Moreover, two edge preserving filters will increase the accuracy on the repetitive pattern region. The results show that the proposed method is accurate and capable to work with the challenging regions. The results are provided by the Middlebury standard dataset. The framework is also efficiently and can be applied on the 3D surface reconstruction. Moreover, this work is greatly competitive with previously available methods.
Using A Application For A Desktop ApplicationTracy Huang
An Intentional Infliction of Emotional Distress claim has four elements that must be proven:
1. The defendant's conduct was intentional or reckless.
2. The conduct was outrageous and intolerable, exceeding all bounds of decency.
3. The defendant's conduct caused the plaintiff to suffer emotional distress.
4. The plaintiff's emotional distress was severe.
Here, Samuel Taylor would need to prove all four elements against his former employer:
1. The employer intentionally terminated Samuel and made disparaging comments, satisfying the intent requirement.
2. Firing an employee without cause after 20 years and making false statements that severely damaged his reputation could be considered outrageous conduct exceeding all bounds of dec
International Journal on Soft Computing ( IJSC )ijsc
This document provides a summary of various wavelet-based image coding schemes. It discusses the basics of image compression including transformation, quantization, and encoding. It then reviews the wavelet transform approach and several popular wavelet coding techniques such as EZW, SPIHT, SPECK, EBCOT, WDR, ASWDR, SFQ, EPWIC, CREW, SR and GW coding. These techniques exploit the multi-resolution properties of wavelets for superior energy compaction and compression performance compared to DCT-based methods like JPEG. The document provides details on how each coding scheme works and compares their features.
Many applications such as robot navigation, defense, medical and remote sensing perform
various processing tasks, which can be performed more easily when all objects in different images of the
same scene are combined into a single fused image. In this paper, we propose a fast and effective
method for image fusion. The proposed method derives the intensity based variations that is large and
small scale, from the source images. In this approach, guided filtering is employed for this extraction.
Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained.
Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
Spot Edge Detection in cDNA Microarray Images using Window based Bi-Dimension...idescitation
Ongoing Microarray is an increasingly playing a crucial role applied in the field
of medical and biological operations. The initiator of Microarray technology is M. Schena et
al. [1] and from past few years microarrays have begun to be used in many fields such as
biomedicine, mostly on cancer and Diabetic, and medical diagnoses. A Deoxyribonucleic
Acid (DNA) microarray is a collection of microscopic DNA spots attached to a solid surface,
such as glass, plastic or silicon chip forming an array. Processing of DNA microarray image
analysis includes three tasks: gridding, segmentation and intensity extraction and at the
stage of processing, the irregularities of shape and spot position which leads to generate
significant errors. This article presents a new spot edge detection method using Window
based Bi-dimensional Empirical Mode Decomposition. On separating spots form the
background area and to decreases the probability of errors and gives more accurate
information about the states of spots we are proposing a spot edge detection via WBEMD.
By using this method we can identify the spots with low density, which leads to increasing
the performance of cDNA microarray images.
Method of optimization of the fundamental matrix by technique speeded up rob...IJECEIAES
The purpose of determining the fundamental matrix (F) is to define the epipolar geometry and to relate two 2D images of the same scene or video series to find the 3D scenes. The problem we address in this work is the estimation of the localization error and the processing time. We start by comparing the following feature extraction techniques: Harris, features from accelerated segment test (FAST), scale invariant feature transform (SIFT) and speed-up robust features (SURF) with respect to the number of detected points and correct matches by different changes in images. Then, we merged the best chosen by the objective function, which groups the descriptors by different regions in order to calculate F. Then, we applied the standardized eight-point algorithm which also automatically eliminates the outliers to find the optimal solution F. The test of our optimization approach is applied on the real images with different scene variations. Our simulation results provided good results in terms of accuracy and the computation time of F does not exceed 900 ms, as well as the projection error of maximum 1 pixel, regardless of the modification.
This document discusses hybrid image fusion techniques for combining information from multiple images. It begins by defining image fusion and its applications in remote sensing to produce more informative images. It then describes two main categories of image fusion - spatial domain fusion, which operates directly on pixel values, and frequency domain fusion using transforms like wavelets. The key points are:
1. A hybrid technique is proposed that uses both spatial and frequency domain fusion, taking the sum of each.
2. In frequency domain fusion, discrete wavelet transforms are used to decompose images into approximation and detail coefficients, which are then fused using operations like mean, max, min.
3. The hybrid technique fuses the results of both spatial and frequency domain
Adaptive Image Contrast with Binarization Technique for Degraded Document Imagetheijes
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
Theoretical work submitted to the Journal should be original in its motivation or modeling structure. Empirical analysis should be based on a theoretical framework and should be capable of replication. It is expected that all materials required for replication (including computer programs and data sets) should be available upon request to the authors.
The International Journal of Engineering & Science would take much care in making your article published without much delay with your kind cooperation
Image fusion is the process of combining two or more images with specific objects with more precision. It is very common that when one object is focused remaining objects will be less highlighted. To get an image highlighted in all areas, a different means is necessary. This is done by the Image Fusion. In remote sensing, the increasing availability of Space borne images and synthetic aperture radar images gives a motivation to different kinds of image fusion algorithms. In the literature a number of time domain image fusion techniques are available. Few transform domain fusion techniques are proposed. In transform domain fusion techniques, the source images will be decomposed, then integrated into a single data and will be reconstructed back into time domain. In this paper, singular value decomposition as a tool to have transform domain data will be utilized for image fusion. In the literature, the quality assessment of fusion techniques is mainly by subjective tests. In this paper, objective quality assessment metrics are calculated for existing and proposed techniques. It has been found that the new image fusion technique outperformed the existing ones.
This document reviews techniques for multi-image morphing. It discusses early cross-dissolve morphing methods and their limitations. Mesh warping and field morphing are introduced as improved techniques that use control points and line mappings to better align images during transition. The document also summarizes point distribution, critical point filters, and other common morphing methods. It concludes by noting that effective morphing requires mechanisms for feature specification, warp generation, and transition control.
1. Exposure fusion by Fast and Adaptive Bidimensional Empirical Mode Decomposition
I. Lousis, K. Ioannidis, I. Andreadis
Laboratory of Electronics, Department of Electrical and Computer Engineering
Democritus University of Thrace, GR-67100, Xanthi, Greece
Email: {ilialous, kioannid, iandread}@ee.duth.gr
ABSTRACT
A new method for fusing a scene of two or more different exposed images is proposed. The process of Fast and Adaptive Bidimensional Empirical mode Decomposition is adopted in order to decompose the illumination component of the input images in their Intrinsic Mode Functions (IMFs). The features of each image can then be detected by applying local energy operators within the IMFs and the final fused image is derived as a collection of features from the input images. Finally, the color information is selected from the “best exposed” pixels of the input sequence. Experimental results show that this method captures all the features from the input images while the fused images display uniform pixel values in the entire image region.
KEY WORDS
High Dynamic Range Imaging, Exposure Fusion, Fast and Adaptive Bidimensional Empirical Mode Decomposition (FABEMD).
1. Introduction
Current imaging technology needs to deal with the inefficiency of most digital image sensors to capture the full dynamic range of a scene. The most frequent approach to this problem involves capturing multiple exposures of the same scene [1], and combining them into a single image. Due to different exposure times for each image, this approach captures different details of the same scene. The resulted multi-exposure images are fused into a single image of a higher dynamic range (HDR) followed by a tone mapping process in order to be displayed in a conventional low dynamic range devices such as computer displays.
Several tone mapping operators have been suggested in the past which can be categorized into two main classes. Global tone mapping operators [1] - [4] take into consideration the whole set of image pixels while a linear or nonlinear curve is produced. All pixels are remapped using the resulted curve in order that the image can be displayed in a low dynamic range device. Local tone mapping operators [5] - [7] transform local areas within the image based on specific measures while the pixels are remapped to the available dynamic range. Thus, each pixel could be remapped by a different remapping curve. Compared with the local operators, global operators outperform in terms of speed since the transformation curve is calculated in one single step. On the other hand, local operators produce smoother results to the human perception.
In addition, Mertens et al. [8] introduced an exposure fusion method where the input images are processed in order to extract quality measures like contrast, saturation and well-exposedness. Subsequently, the images are fused directly to a single high quality low dynamic range (LDR) skipping the intermediate step of an HDR image creation which eventually simplifies the entire process. Moreover, Goshtasby in [9] proposed a block level method in order to increase the efficiency of the spatial information exploitation. The required weights for transforming each pixel of the input images are calculated according to the entropy of each image block. Vonikakis et al. [10] proposed a fusion method whose basis relies on illumination estimation. Once illumination is estimated, membership-like functions are applied to define the required weights for the best exposed pixels from each image. Finally, Ahmed et al. [11] applied a modified version of bidimensional empirical mode decomposition (BEMD), namely fast and adaptive bidimensional empirical mode decomposition (FABEMD) [12], in order to achieve multi-focused image fusion.
The FABEMD was adopted since it can detect features such as edges and texture within an image. Such features play an important role in the human visual system and ultimately can be used for the recognition of the “well-exposed” regions within the image. This ensures that the final image will include all the unique features of the input images. Experimental results of this approach show that the proposed method produces comparable or even better results than the existing techniques.
In the proposed technique, the FABEMD method is used for multi-exposure image fusion. Initially, the YCbCr counterpart of each image is computed and the luminance vector is decomposed into multiple Intrinsic Mode Functions (IMFs) by the application of FABEMD method. In addition, based on local energy measurements, the required weights for each IMF are extracted. The summation of the weighted IMFs represents the Y component of the final HDR image. Finally, the color information of the final image is obtained from the “best exposed” pixels of the input images. This yields that the final image will have a purely natural representation of
2. Figure 1. Overview of the proposed method.
colors such as those of the original images, since there is no further processing of the color components. A brief demonstration of the method is presented in Fig. 1. Furthermore, the process is performed only in the luminance component of the images instead of all components as applied in fusion in the RGB domain. Indeed, this is essential in reducing the computationally cost and could be a useful approach in the design of future real time applications.
The rest of the paper is organized as follows. The EMD and the FABEMD methods are briefly described in Section 2. In Section 3, the proposed method is presented in details while in Section 4, experimental results are provided. Finally, conclusions are provided in Section 5.
2. EMD & FABED
Empirical Mode decomposition (EMD) was originally proposed by Huang, in order to overcome the drawbacks of the Hilbert Transform in non-linear and non-stationary signals. The combination of the EMD process with the Hilbert Transform forms the Hilbert-Huang Transform (HHT) [13], a powerful time-frequency data analysis tool. The EMD process decomposes a signal to its basis functions, called IMFs. An IMF represents a function that satisfies the following two conditions:
a. The number of extrema and the number of zero crossings must be equal or differ by at most one in the whole data.
b. The mean value of the envelope defined by the local maxima and the envelope of the local minima must be zero at every point.
If the above two conditions are not fulfilled, then the candidate IMF is processed by an iterative process, called sifting process. The sifting process is repeated until the candidate IMF fulfills the demanded conditions. The processing steps of the EMD process are:
i. Input the original data.
ii. Detect the local extrema (maxima and minima).
iii. Create the upper and lower envelopes by interpolating the maxima and the minima points respectively, and calculate the mean envelope.
iv. Subtract the mean envelope from the original data.
v. Check if the residue of the subtraction fulfills the requirements of an IMF. If so, store the IMF, subtract it from the original data and repeat the above steps for the residue, until the residue either appears one extrema or is a monotonic function. If not, treat the candidate IMF as the original data, and repeat the process until it fulfills the required conditions.
The flowchart of the EMD process is provided in Fig. 2.
Bidimensional empirical mode decomposition (BEMD) is the extension of the EMD process in two dimensional signals, such as images. In both EMD and BEMD, the resulted IMFs strictly depend on the interpolation method that is used in order to construct both the envelopes. Especially in the case of BEMD, traditional ways of interpolation include 2-d spline
3. interpolation, radial basis functions, finite elements
methods etc. All these methods suffer in terms of
accuracy in regions near the boundary of the image, since
few points are determined to support the interpolation
process. The effects of these drawbacks are unwanted
oscillations in these regions such as overshoots or
undershoots, which produce incorrect IMFs. Furthermore,
most of the classic interpolation methods are
computationally intense. Thus, their continuous
application during the sifting process renders the entire
algorithm time consuming.
Figure 2. Flowchart of the EMD process.
An efficient algorithm to decompose an image to its
IMFs was proposed by Bhuiyan [12] and was called Fast
and Adaptive Bidimensional Empirical Mode
Decomposition (FABEMD). Instead of applying the
classic interpolation methods, the FABEMD algorithm
calculates the required envelopes by using order statistic
filters. A MAX filter is applied for the upper envelope
construction, whereas for the lower envelope a MIN filter
is used. Once these filters are applied to the image,
smoothing operators are used in order to obtain the
desired envelope characteristics.
The most crucial processing step of the FABEMD
algorithm is to determine correctly the window size for
both the MAX and the MIN filter. The window size is
calculated as follows:
i. Detect the maxima and the minima within the image,
by comparing each element with its 3 × 3 neighbors.
If its value is higher than the value of the others then
consider it as a local maximum point and if its value
is higher than the value of the others then consider it
as a local minimum point.
ii. For each point determined as extrema, calculate the
Euclidean distance between itself and its nearest
neighbor. By the end of this step, there will have been
determined the maxima and minima maps M(x,y) and
N(x,y), respectively.
iii. The window size is calculated as
d M N
d M N
d M N
d M N
max max ,max
min max ,max
max min ,min
min min ,min
4
3
2
1
(1)
where, min and max denotes the minimum and maximum
operators, respectively and the order statistics filters are
considered as Type-1, Type-2, Type-3 and Type-4,
respectively.
The selection of the filter type relies on the
application of the algorithm. The method will produce a
bigger number of IMFs when using a filter of Type-1
while, on the contrary, the application of a Type-4 filter
will produce the minimum set of IMFs. Once the filtering
process is completed, the following operators are applied
in order to create smooth and continuous surfaces which
represent the required envelopes:
xy
xy
k l Z
E Ei
k l Z
E Ei
L k l
w w
L
U k l
w w
U
,
,
,
1
,
1
(2)
where, UEi, LEi are the upper and lower envelopes
respectively, Zxy denotes a square region w × w centered at
any point (x,y) of UEi and LEi, and UE, LE are the smoothed
upper and lower envelopes.
The sifting process is excluded and thus, a single
iteration is required. The entire process is repeated until
the limit of six extrema or less is reached. The initial
image can be expressed as:
I x y IMF x y Rx y
n
k
k , , ,
1
(3)
where, n denotes the total number of the produced IMFs
and R(x,y) is the residue signal. The flowchart of the
FABEMD process is demonstrated in Fig. 3.
Figure 3. Flowchart of the FABEMD algorithm.
4. 3. Exposure fusion via FABEMD
Exposure fusion comprises a set of techniques where
images with different exposure times are fused avoiding
the step of creating an HDR image [8]. The images are
directly fused into a high quality LDR image which
contains the best exposed details from the input image
sequence. A set of image fusion approaches includes the
decomposition of the original images into their
fundamental components and the fused image is
reconstructed by appropriate processing of the resulted
components. A representative approach for such type of
image fusion includes the application of wavelet
transform [14], [15]. Hariharan in [16] used the BEMD
algorithm in order to fuse images which were captured
from different sensors (thermal and visual images). The
method produces visually better results than the wavelet
decomposition approach as well as than similar image
fusion methods.
The main drawback that may occur by the application
of the BEMD method is that the process may result a
different number of IMFs for each image and thus, their
scales cannot be matched correctly. Looney et al. in [17]
used complex extensions of the EMD process in order to
decompose two images into the same number of IMFs.
The FABEMD process was proven to be efficient in
decomposing two different images into the same number
of IMFs by correctly choosing the window of the order
statistic filters. The FABEMD is applied to each RGB
channel of the input images and each set of IMFs is fused
to create the IMFs of the RGB channels of the fused
image. The nature of these images is suitable for applying
the FABEMD method to the RGB domain since they
contain the same color in the corresponding pixels, and
thus the color of the fused image will be exactly the color
of the initial images.
Nevertheless, in case of multiple exposure images,
each image contains different chromatic information in
multiple image regions, and thus fusion via FABEMD in
the RGB domain fails to produce satisfactory results. To
overcome this issue, the method initially transforms the
input images from the RGB to the YCbCr color space in
order to keep the computational burden at low levels and
exploit the luminance vector at the next processing step.
After the color space conversion, the FABEMD
process is applied to the Y-channel of each image. The
size of the order statistics filter is calculated by the
formula:
max{max{ },max{ },...,max{ }} 1 2 i w d d d (4)
where, i denotes the number of the input images and d
denotes the type of the applied filter.
The proposed method utilizes a Type-4 filter for the
decomposition process, as discussed Section 2. In
addition, since the same filter size is used for all input
images, the same number of IMFs, n, will be produced.
This yields that the Y-component of the fused image can
be reproduced by the weighted summation of the resulted
n IMFs. The resulted residues of each image are
processed as the corresponding IMFs. The IMFs of the
fused image are extracted by weighting the local amount
of information that each IMF carries:
k
i
i
k
i
i i
fused
IMF x y
IMF x y
1
1
,
,
(5)
where, (x,y) denotes the position of each pixel, k is the
number of the input images and αi are the coefficients that
are equal with the energy included in the 3 × 3
neighborhood of the (x,y) pixel.
The energy metric was selected due to its feature in
defining the contained information of a specific region.
The above process is repeated n+1 times, because there
are n-IMFs plus the residue of each image. In addition, in
order to increase the visual quality of the final fused
image, a local weighting approach was adopted meaning
that the applied weighting factor is proportional to the
calculated energy within the neighborhood of each pixel.
The applied region dimensions were selected to be 3 × 3
pixels for best optical results and speed issues.
The summation of the resulted weighted IMFs
corresponds to the Y component of the final image. The
maximum available dynamic range of the Y channel
requires that its values will be in the interval [16,236]. In
order to guarantee that the fused image will take
advantage of the full dynamic range, the following linear
operator is applied:
16 235 16
max min
min
Y Y
Y Y
Yout (6)
where, Yout and Y are the stretched values and the initial
values of the Y channel, respectively. Furthermore, Ymax
and Ymin are the maximum and minimum values of the
initial Y channel, respectively. The purpose of this
operator is to place the resulted values in the interval
[16,236]. Since the Y channel processing step is
completed, the appropriate color values are calculated. A
pixel is defined as well-exposed only if its luminance
value is approximately equal to 128, meaning that its
luminance is close to the median value of the values in the
interval [0,255]. The color information of each fused pixel
(Cb and Cr channels) is selected as the color information
(Cb and Cr channels) from the input pixel whose
luminance is closest to the 128 value. The chromatic
information for each fused pixel is calculated through the
following formula:
argmin( (128 - ), (128 ),..., (128 - )
( , ) ( , )
1 2 k
i
fused
i
i abs Y abs Y abs Y
C x y C x y
(7)
5. where, Ci corresponds to either the Cb or the Cr chromatic vector, x and y are the pixel coordinates, i is the index where the difference becomes minimum and argmin is the argument of minimum function. The same process is applied in both Cb and Cr chromatic vectors.
Finally, a YCbCr to RGB color space transformation is performed in order to display or save the final fused image to a compatible electronic equipment or an appropriate image format, respectively.
4. Experimental Results
In order to study the results of the proposed method, the produced images were quantitatively compared with other related approaches, namely with Mertens et al. method [8], Vonikakis et al. method [10], the Essential HDR [18] and the Photomatix [19]. Mertens et al. method is a classic exposure fusion algorithm. Vonikakis et al. algorithm is a very recent algorithm in the specific field and the Essential HDR and Photomatix are commercial software products. The latter provide the user the option to enhance further the produced image, by controlling the local contrast, color saturation and other parameters. In order to have an as objective comparison as possible, the images compared in this section are the default fused images, without further enhancement. The Root Mean Square Error (RMSE) of a well-exposed region (defined by observing the images) in the initial image sequence is calculated in each case, in order to have an objective comparison among them all. For example, Fig. 4 depicts the “Church” scene.
Results reveal that in RMSE terms, the proposed method produces comparable numerical results to the Vonikakis et al. approach, and outperforms the rest of the other methods. Moreover, it can be observed that the produced images lack of overexposed or underexposed regions. It is also clear that the proposed method perceives a pure natural representation since it does not involve any color processing. The advantage of the proposed method in preserving the original colors is better illustrated by observing Figs. 5 and 6.
More specifically, Fig. 5 depicts the “Venice” scene, where the proposed method produces similar results in terms of the RMSE metric with the corresponding results of Vonikakis et al. approach. However, it can be observed that the proposed method preserves the color information of the fused images better than all the other methods. This is due to the fact that the proposed method uses the color information of the initial images. No further processing is carried out and thus, the colors of the fused image are the same as the colors of the original images. This can be observed in the right column of Fig. 5. The proposed method captures the most vivid red color, which is closer to the original red color.
Table 1. Summary of the RMSE metric results.
Algorithm
Image scenes
Church
Venice
Coast
Essential HDR
49.27
76.27
145.07
Photomatix
47.18
95.05
114.89
Vonikakis et al.
16.87
65.05
113.75
Mertens et al.
82.13
96.82
96.08
Proposed method
19.08
75.44
91.19
Figure 4. Comparison for the “Church” scene.
6. Figure 5. Comparison for the “Venice” scene.
In addition, the advantage of the proposed method in perceiving the chromatic information is clearly demonstrated in Fig. 6. In this case, the proposed method outperforms all the other methods in RMSE metrics. Moreover, by observing the right column of the figure (zoomed areas), it is clear that the proposed method produces the best result as long as the colors contained are concerned. These areas contain information both from Original (1) and Original (3) images. This makes it really difficult to compare these regions in terms of RMSE, and thus no RMSE metrics are included.
7. Figure 6. Comparison for the “Coast” scene.
The performance of all the compared algorithms, as long as the RMSE metric is concerned, is summarized in Table 1. From the experimental results it is clear that the proposed method produces comparable or even better results than the rest of the methods. The resulted fused images display the best exposed image regions, since the spectral content of each input image is used. The main advantage of the proposed method is that it processes regions which display significant difference in edge or texture content, instead of fussing an entire image
8. sequence. Furthermore, compared to other similar methods, it has a clear advantage in preserving the color information, due to the fact that the colors are simply selected from the initial images, without further processing.
5. Conclusions
In this paper, a novel method of fusing multiple exposed images is presented. The proposed approach is based on the FABEMD algorithm which is used to decompose the luminance vector of the original images into a number of sub-images, called IMFs. The resulted images provide a useful distribution of information within the images. In contrary with the majority of the reported methods that use the illumination values to define whether a region is overexposed or underexposed, the proposed approach detects such regions due to the definition of spectral information. The calculated IMFs are then fused to extract the corresponding IMFs of the final image by using proportional weights to the local energy of the neighborhood of each pixel. Subsequently, the color information is selected from the “best exposed” pixels of the image sequence in order to create the final HDR image.
Experimental results reveal that the proposed method produces comparable or even better results than other similar methods in terms of RMSE while the produced images display a more realistic appearance of the colors. Furthermore, the most computationally intense part of the processing is being inducted in only one component (Y- channel) of the initial images.
References
[1] E. Reinhard, G. Ward, S. Pattanaik, and P. Debevec. High dynamic range imaging: Acquisition, display and image-based lighting (Amsterdam: Elsevier/Morgan Kaufmann, 1st edition, 2006).
[2] F. Drago, K. Myszkowski, T. Annen, and N. Chiba. Adaptive logarithmic mapping for displaying high contrast scenes, Computer Graphics Forum, 22(3), 2003, 419-426.
[3] C. Schlick, Quantization techniques for the visualization of high dynamic range pictures, P. Shirley, G. Sakas, and S. Muller (eds.), Photorealistic Rendering Techniques, New York: Springer-Verlag, 1994, 7-20.
[4] G. W. Larson, H. E. Rushmeier, and C. D. Piatko. A visibility matching tone reproduction operator for high dynamic range scenes. IEEE Trans. Vis. Comput. Graphics, 3(4), 1997, 291–306.
[5] F. Durand and J. Dorsey. Fast bilateral filtering for the display of high-dynamic range images, ACM Trans. Graphics, 21(3), 2002, 257–266.
[6] R. Fattal, D. Lischinski, and M. Werman, Gradient domain high dynamic range compression, ACM Trans. Graphics, 21(3), 2002, 249–256.
[7] L. Meylan, and S. Süsstrunk, High dynamic range image rendering with a retinex-based adaptive filter, IEEE Trans. Image Process., 15(9), 2006, 2820– 2830.
[8] T. Mertens, J. Kautz and F. Van Reeth, Exposure fusion, Proc. 15th Pacific Conf. on Computer Graphics and Applications, 2007, 382 – 390.
[9] A. A. Goshtasby, Fusion of multi-exposure images, Image and Vision Computing, 23(6), 2005, 611 – 618.
[10] V. Vonikakis, O. Bouzos and I. Andreadis, Multi- exposure image fusion based on illumination estimation, Proc. IASTED Int. Conf. on Signal and Image Processing and Applications, Crete, Greece, 2011, 135 – 142.
[11] M.U. Ahmed and D.P. Mandic, Image fusion based on fast and adaptive bidimensional empirical mode decomposition, Proc. 13th Conf. on Information Fusion, 2010, 1 – 6.
[12] S.M.A. Bhuiyan, R.R. Adhami, and J.F. Khan, Fast and adaptive bidimensional empirical mode decomposition using order-statistics filter based envelope estimation, EURASIP Journal on Advances in Signal Processing 164, 2008, 1-18.
[13] N. E. Huang, Z. Shen, S. R. Long, M. C. Wu, H. H. Shih, Q. Zheng, N.-C. Yen, C. C. Tung, and H. H. Liu, The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis, Proc. R. Soc. Lond. A, Math. Phys. Sci. 454(1971), 1998, 903 – 995.
[14] M. H. Malik, S. A. M. Gilani and Anwaar-ul-Haq, Wavelet based exposure fusion, Proc. of World Congress on Engineering, 2008, 688 – 693.
[15] T. Stathaki, Image fusion: Algorithms and applications (Academic Press, 2008).
[16] H. Hariharan, A. Gribok, M. Abidi, and A. Koschan, Image fusion and enhancement via empirical mode decomposition, Journal of Pattern Recognition Research, 1(1), 2006, 16 – 32.
[17] D. Looney and D. P. Mandic, Multiscale image fusion using complex extensions of EMD, IEEE Trans. Signal Process., 57(4), 2009, 1626 – 1630.
[18] http://www.imagingluminary.com
[19] http://www.hdrsoft.com/