This document presents an algorithm for image blur detection using 2D Haar wavelet transforms. The algorithm splits images into tiles and applies the 2D Haar wavelet transform to each tile to detect regions with pronounced changes. Tiles with similar changes are grouped into clusters. Images with large clusters are classified as sharp, while images with small clusters are classified as blurred. The algorithm is integrated with a skewed barcode scanning algorithm to filter out blurred images, improving scanning success rates. Experimental results found it performed comparably or better than two other blur detection algorithms and positively impacted skewed barcode scanning when integrated.
Frequency Domain Blockiness and Blurriness Meter for Image Quality AssessmentCSCJournals
Image and video compression introduces distortions (artefacts) to the coded image. The most prominent artefacts added are blockiness and blurriness. Many existing quality meters are normally distortion-specific. This paper proposes an objective quality meter for quantifying the combined blockiness and blurriness distortions in frequency domain. The model first applies edge detection and cancellation, then spatial masking to mimic the characteristics of the human visual system. Blockiness is then estimated by transforming image into frequency domain, followed by finding the ratio of harmonics to other AC components. Blurriness is determined by comparing the high frequency coefficients of the reference and coded images due to the fact that blurriness reduces the high frequency coefficients. Then, both blockiness and blurriness distortions are combined for a single quality metric. The meter is tested on blocky and blurred images from the LIVE image database, with a correlation coefficient of 95-96%.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Invariant Recognition of Rectangular Biscuits with Fuzzy Moment Descriptors, ...CSCJournals
In this paper a new approach for invariant recognition of broken rectangular biscuits is proposed using fuzzy membership-distance products, called fuzzy moment descriptors. The existing methods for recognition of flawed rectangular biscuits are mostly based on Hough transform. However these methods are prone to error due to noise and/or variation in illumination. Fuzzy moment descriptors are less sensitive to noise thus making it an effective approach invariant to the above stray external disturbances. Further, the normalization and sorting of the moment vectors make it a size and rotation invariant recognition process .In earlier studies fuzzy moment descriptors has successfully been applied in image matching problem. In this paper the algorithm is applied in recognition of flawed and non-flawed rectangular biscuits. In general the proposed algorithm has potential applications in industrial quality control.
Automatic rectification of perspective distortion from a single image using p...ijcsa
Perspective distortion occurs due to the perspective projection of 3D scene on a 2D surface. Correcting the distortion of a single image without losing any desired information is one of the challenging task in the field of Computer Vision. We consider the problem of estimating perspective distortion from a single still image of an unstructured environment and to make perspective correction which is both quantitatively accurate as well as visually pleasing. Corners are detected based on the orientation of the image. A method based on plane homography and transformation is used to make perspective correction. The algorithm infers frontier information directly from the images, without any reference objects or prior knowledge of the camera parameters. The frontiers are detected using geometric context based segmentation. The goal of this paper is to present a framework providing fully automatic and fast perspective correction.
MAGNETIC RESONANCE BRAIN IMAGE SEGMENTATIONVLSICS Design
Segmentation of tissues and structures from medical images is the first step in many image analysis applications developed for medical diagnosis. With the growing research on medical image segmentation, it is essential to categorize the research outcomes and provide researchers with an overview of the existing segmentation techniques in medical images. In this paper, different image segmentation methods applied on magnetic resonance brain images are reviewed. The selection of methods includes sources from image processing journals, conferences, books, dissertations and thesis. The conceptual details of the methods are explained and mathematical details are avoided for simplicity. Both broad and detailed categorizations of reviewed segmentation techniques are provided. The state of art research is provided with emphasis on developed techniques and image properties used by them. The methods defined are not always mutually independent. Hence, their inter relationships are also stated. Finally, conclusions are drawn summarizing commonly used techniques and their complexities in application.
Frequency Domain Blockiness and Blurriness Meter for Image Quality AssessmentCSCJournals
Image and video compression introduces distortions (artefacts) to the coded image. The most prominent artefacts added are blockiness and blurriness. Many existing quality meters are normally distortion-specific. This paper proposes an objective quality meter for quantifying the combined blockiness and blurriness distortions in frequency domain. The model first applies edge detection and cancellation, then spatial masking to mimic the characteristics of the human visual system. Blockiness is then estimated by transforming image into frequency domain, followed by finding the ratio of harmonics to other AC components. Blurriness is determined by comparing the high frequency coefficients of the reference and coded images due to the fact that blurriness reduces the high frequency coefficients. Then, both blockiness and blurriness distortions are combined for a single quality metric. The meter is tested on blocky and blurred images from the LIVE image database, with a correlation coefficient of 95-96%.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Invariant Recognition of Rectangular Biscuits with Fuzzy Moment Descriptors, ...CSCJournals
In this paper a new approach for invariant recognition of broken rectangular biscuits is proposed using fuzzy membership-distance products, called fuzzy moment descriptors. The existing methods for recognition of flawed rectangular biscuits are mostly based on Hough transform. However these methods are prone to error due to noise and/or variation in illumination. Fuzzy moment descriptors are less sensitive to noise thus making it an effective approach invariant to the above stray external disturbances. Further, the normalization and sorting of the moment vectors make it a size and rotation invariant recognition process .In earlier studies fuzzy moment descriptors has successfully been applied in image matching problem. In this paper the algorithm is applied in recognition of flawed and non-flawed rectangular biscuits. In general the proposed algorithm has potential applications in industrial quality control.
Automatic rectification of perspective distortion from a single image using p...ijcsa
Perspective distortion occurs due to the perspective projection of 3D scene on a 2D surface. Correcting the distortion of a single image without losing any desired information is one of the challenging task in the field of Computer Vision. We consider the problem of estimating perspective distortion from a single still image of an unstructured environment and to make perspective correction which is both quantitatively accurate as well as visually pleasing. Corners are detected based on the orientation of the image. A method based on plane homography and transformation is used to make perspective correction. The algorithm infers frontier information directly from the images, without any reference objects or prior knowledge of the camera parameters. The frontiers are detected using geometric context based segmentation. The goal of this paper is to present a framework providing fully automatic and fast perspective correction.
MAGNETIC RESONANCE BRAIN IMAGE SEGMENTATIONVLSICS Design
Segmentation of tissues and structures from medical images is the first step in many image analysis applications developed for medical diagnosis. With the growing research on medical image segmentation, it is essential to categorize the research outcomes and provide researchers with an overview of the existing segmentation techniques in medical images. In this paper, different image segmentation methods applied on magnetic resonance brain images are reviewed. The selection of methods includes sources from image processing journals, conferences, books, dissertations and thesis. The conceptual details of the methods are explained and mathematical details are avoided for simplicity. Both broad and detailed categorizations of reviewed segmentation techniques are provided. The state of art research is provided with emphasis on developed techniques and image properties used by them. The methods defined are not always mutually independent. Hence, their inter relationships are also stated. Finally, conclusions are drawn summarizing commonly used techniques and their complexities in application.
Texture based feature extraction and object trackingPriyanka Goswami
The project involved developing and implementing different texture analysis based extraction techniques like Local Binary Pattern (LBP), Local Derivative Pattern (LDP) and Local Ternary Pattern (LTP) in MATLAB and carrying out a comparative study by analyzing the effectiveness of each technique using a standard set of images (Yale data set). The most optimum technique is then applied to identify cloud patterns and track their motion (in pixel position changes) in time series images (acquired from weather satellites like GOES) using the Chi-Square Difference method.
Mislaid character analysis using 2-dimensional discrete wavelet transform for...IJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
Performance of Efficient Closed-Form Solution to Comprehensive Frontier Exposureiosrjce
IOSR Journal of Electronics and Communication Engineering(IOSR-JECE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of electronics and communication engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in electronics and communication engineering. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
This article aims at a new algorithm for tracking moving objects in the long term. We have tried to overcome some potential difficulties, first by a comparative study of the measuring methods of the difference and the similarity between the template and the source image. In the second part, an improvement of the best method allows us to follow the target in a robust way. This method also allows us to effectively overcome the problems of geometric deformation, partial occlusion and recovery after the target leaves the field of vision. The originality of our algorithm is based on a new model, which does not depend on a probabilistic process and does not require a data based detection in advance. Experimental results on several difficult video sequences have proven performance advantages over many recent trackers. The developed algorithm can be employed in several applications such as video surveillance, active vision or industrial visual servoing.
A NOVEL PROBABILISTIC BASED IMAGE SEGMENTATION MODEL FOR REALTIME HUMAN ACTIV...sipij
Automatic human activity detection is one of the difficult tasks in image segmentation application due to
variations in size, type, shape and location of objects. In the traditional probabilistic graphical
segmentation models, intra and inter region segments may affect the overall segmentation accuracy. Also,
both directed and undirected graphical models such as Markov model, conditional random field have
limitations towards the human activity prediction and heterogeneous relationships. In this paper, we have
studied and proposed a natural solution for automatic human activity segmentation using the enhanced
probabilistic chain graphical model. This system has three main phases, namely activity pre-processing,
iterative threshold based image enhancement and chain graph segmentation algorithm. Experimental
results show that proposed system efficiently detects the human activities at different levels of the action
datasets.
Gesture Recognition using Principle Component Analysis & Viola-Jones AlgorithmIJMER
Gesture recognition pertains to recognizing meaningful expressions of motion by a human,
involving the hands, arms, face, head, and/or body. It is of utmost importance in designing an intelligent
and efficient human–computer interface. The applications of gesture recognition are manifold, ranging
from sign language through medical rehabilitation to virtual reality. In this paper, we provide a survey on
gesture recognition with particular emphasis on hand gestures and facial expressions. Applications
involving wavelet transform and principal component analysis for face and hand gesture recognition on
digital images
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
FINGERPRINT MATCHING USING HYBRID SHAPE AND ORIENTATION DESCRIPTOR -AN IMPROV...IJCI JOURNAL
Fingerprint recognition is a promising factor for the Biometric Identification and authentication process.
Fingerprints are broadly used for personal identification due to its feasibility, distinctiveness, permanence,
accuracy and acceptability. This paper proposes a way to improve the Equal Error Rate (EER) in
fingerprint matching techniques in the domain of hybrid shape and orientation descriptor. This type of
fingerprint matching domain is popular due to capability of filtering false and strange minutiae pairings.
EER is calculated by using FMR and FNMR to check the performance of proposed technique.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Fuzzy Logic based Edge Detection Method for Image Processing IJECEIAES
Edge detection is the first step in image recognition systems in a digital image processing. An effective way to resolve many information from an image such depth, curves and its surface is by analyzing its edges, because that can elucidate these characteristic when color, texture, shade or light changes slightly. Thiscan lead to misconception image or vision as it based on faulty method. This work presentsa new fuzzy logic method with an implemention. The objective of this method is to improve the edge detection task. The results are comparable to similar techniques in particular for medical images because it does not take the uncertain part into its account.
Comparative Analysis of Common Edge Detection Algorithms using Pre-processing...IJECEIAES
Edge detection is the process of segmenting an image by detecting discontinuities in brightness. Several standard segmentation methods have been widely used for edge detection. However, due to inherent quality of images, these methods prove ineffective if they are applied without any preprocessing. In this paper, an image pre-processing approach has been adopted in order to get certain parameters that are useful to perform better edge detection with the standard edge detection methods. The proposed preprocessing approach involves median filtering to reduce the noise in image and then edge detection technique is carried out. Finally, Standard edge detection methods can be applied to the resultant pre-processing image and its Simulation results are show that our pre-processed approach when used with a standard edge detection method enhances its performance.
Performance Evaluation of Image Edge Detection Techniques CSCJournals
The success of an image recognition procedure is related to the quality of the edges marked. The
aim of this research is to investigate and evaluate edge detection techniques when applied to
noisy images at different scales. Sobel, Prewitt, and Canny edge detection algorithms are
evaluated using artificially generated images and comparison criteria: edge quality (EQ) and map
quality (MQ). The results demonstrated that the use of these criteria can be utilized as an aid for
further analysis and arbitration to find the best edge detector for a given image.
An efficient method for recognizing the low quality fingerprint verification ...IJCI JOURNAL
In this paper, we propose an efficient method to provide personal identification using fingerprint to get better accuracy even in noisy condition. The fingerprint matching based on the number of corresponding minutia pairings, has been in use for a long time, which is not very efficient for recognizing the low quality fingerprints. To overcome this problem, correlation technique is used. The correlation-based fingerprint verification system is capable of dealing with low quality images from which no minutiae can be extracted reliably and with fingerprints that suffer from non-uniform shape distortions, also in case of damaged and partial images. Orientation Field Methodology (OFM) has been used as a preprocessing module, and it converts the images into a field pattern based on the direction of the ridges, loops and bifurcations in the image of a fingerprint. The input image is then Cross Correlated (CC) with all the images in the cluster and the highest correlated image is taken as the output. The result gives a good recognition rate, as the proposed scheme uses Cross Correlation of Field Orientation (CCFO = OFM + CC) for fingerprint identification.
Shot Boundary Detection using Radon Projection MethodIDES Editor
The detection of shot boundaries provides a base for
nearly all video abstraction and high-level video segmentation
approaches. Therefore solving the problem of shot boundary
detection is one of the major prerequisites for revealing
higher level video content structure. As a crucial step in video
indexing and retrieval, accurate shot boundary detection plays
an important role to organize and summarize video content
into meaningful shots for further scene analysis. Many
algorithms have been proposed for detecting video shot
boundaries and classifying shots and shot transition types. In
this paper we propose a novel technique for shot boundary
detection using radon transform. We first removed the effect
of illumination using DCT and DWT. Then shot boundary is
detected using radon transform. Radon transform is based on
projection of image intensity along a radial line oriented at a
specific angle. Projection of image intensity for current frame
is different than that of projection of the previous frame, where
shot boundary is detected. In order to verify the performance
of algorithm, experiments have been carried out with news,
documentary and movie. Experimental result demonstrates
efficiency of proposed shot boundary detection technique.
Texture based feature extraction and object trackingPriyanka Goswami
The project involved developing and implementing different texture analysis based extraction techniques like Local Binary Pattern (LBP), Local Derivative Pattern (LDP) and Local Ternary Pattern (LTP) in MATLAB and carrying out a comparative study by analyzing the effectiveness of each technique using a standard set of images (Yale data set). The most optimum technique is then applied to identify cloud patterns and track their motion (in pixel position changes) in time series images (acquired from weather satellites like GOES) using the Chi-Square Difference method.
Mislaid character analysis using 2-dimensional discrete wavelet transform for...IJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
Performance of Efficient Closed-Form Solution to Comprehensive Frontier Exposureiosrjce
IOSR Journal of Electronics and Communication Engineering(IOSR-JECE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of electronics and communication engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in electronics and communication engineering. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
This article aims at a new algorithm for tracking moving objects in the long term. We have tried to overcome some potential difficulties, first by a comparative study of the measuring methods of the difference and the similarity between the template and the source image. In the second part, an improvement of the best method allows us to follow the target in a robust way. This method also allows us to effectively overcome the problems of geometric deformation, partial occlusion and recovery after the target leaves the field of vision. The originality of our algorithm is based on a new model, which does not depend on a probabilistic process and does not require a data based detection in advance. Experimental results on several difficult video sequences have proven performance advantages over many recent trackers. The developed algorithm can be employed in several applications such as video surveillance, active vision or industrial visual servoing.
A NOVEL PROBABILISTIC BASED IMAGE SEGMENTATION MODEL FOR REALTIME HUMAN ACTIV...sipij
Automatic human activity detection is one of the difficult tasks in image segmentation application due to
variations in size, type, shape and location of objects. In the traditional probabilistic graphical
segmentation models, intra and inter region segments may affect the overall segmentation accuracy. Also,
both directed and undirected graphical models such as Markov model, conditional random field have
limitations towards the human activity prediction and heterogeneous relationships. In this paper, we have
studied and proposed a natural solution for automatic human activity segmentation using the enhanced
probabilistic chain graphical model. This system has three main phases, namely activity pre-processing,
iterative threshold based image enhancement and chain graph segmentation algorithm. Experimental
results show that proposed system efficiently detects the human activities at different levels of the action
datasets.
Gesture Recognition using Principle Component Analysis & Viola-Jones AlgorithmIJMER
Gesture recognition pertains to recognizing meaningful expressions of motion by a human,
involving the hands, arms, face, head, and/or body. It is of utmost importance in designing an intelligent
and efficient human–computer interface. The applications of gesture recognition are manifold, ranging
from sign language through medical rehabilitation to virtual reality. In this paper, we provide a survey on
gesture recognition with particular emphasis on hand gestures and facial expressions. Applications
involving wavelet transform and principal component analysis for face and hand gesture recognition on
digital images
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
FINGERPRINT MATCHING USING HYBRID SHAPE AND ORIENTATION DESCRIPTOR -AN IMPROV...IJCI JOURNAL
Fingerprint recognition is a promising factor for the Biometric Identification and authentication process.
Fingerprints are broadly used for personal identification due to its feasibility, distinctiveness, permanence,
accuracy and acceptability. This paper proposes a way to improve the Equal Error Rate (EER) in
fingerprint matching techniques in the domain of hybrid shape and orientation descriptor. This type of
fingerprint matching domain is popular due to capability of filtering false and strange minutiae pairings.
EER is calculated by using FMR and FNMR to check the performance of proposed technique.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Fuzzy Logic based Edge Detection Method for Image Processing IJECEIAES
Edge detection is the first step in image recognition systems in a digital image processing. An effective way to resolve many information from an image such depth, curves and its surface is by analyzing its edges, because that can elucidate these characteristic when color, texture, shade or light changes slightly. Thiscan lead to misconception image or vision as it based on faulty method. This work presentsa new fuzzy logic method with an implemention. The objective of this method is to improve the edge detection task. The results are comparable to similar techniques in particular for medical images because it does not take the uncertain part into its account.
Comparative Analysis of Common Edge Detection Algorithms using Pre-processing...IJECEIAES
Edge detection is the process of segmenting an image by detecting discontinuities in brightness. Several standard segmentation methods have been widely used for edge detection. However, due to inherent quality of images, these methods prove ineffective if they are applied without any preprocessing. In this paper, an image pre-processing approach has been adopted in order to get certain parameters that are useful to perform better edge detection with the standard edge detection methods. The proposed preprocessing approach involves median filtering to reduce the noise in image and then edge detection technique is carried out. Finally, Standard edge detection methods can be applied to the resultant pre-processing image and its Simulation results are show that our pre-processed approach when used with a standard edge detection method enhances its performance.
Performance Evaluation of Image Edge Detection Techniques CSCJournals
The success of an image recognition procedure is related to the quality of the edges marked. The
aim of this research is to investigate and evaluate edge detection techniques when applied to
noisy images at different scales. Sobel, Prewitt, and Canny edge detection algorithms are
evaluated using artificially generated images and comparison criteria: edge quality (EQ) and map
quality (MQ). The results demonstrated that the use of these criteria can be utilized as an aid for
further analysis and arbitration to find the best edge detector for a given image.
An efficient method for recognizing the low quality fingerprint verification ...IJCI JOURNAL
In this paper, we propose an efficient method to provide personal identification using fingerprint to get better accuracy even in noisy condition. The fingerprint matching based on the number of corresponding minutia pairings, has been in use for a long time, which is not very efficient for recognizing the low quality fingerprints. To overcome this problem, correlation technique is used. The correlation-based fingerprint verification system is capable of dealing with low quality images from which no minutiae can be extracted reliably and with fingerprints that suffer from non-uniform shape distortions, also in case of damaged and partial images. Orientation Field Methodology (OFM) has been used as a preprocessing module, and it converts the images into a field pattern based on the direction of the ridges, loops and bifurcations in the image of a fingerprint. The input image is then Cross Correlated (CC) with all the images in the cluster and the highest correlated image is taken as the output. The result gives a good recognition rate, as the proposed scheme uses Cross Correlation of Field Orientation (CCFO = OFM + CC) for fingerprint identification.
Shot Boundary Detection using Radon Projection MethodIDES Editor
The detection of shot boundaries provides a base for
nearly all video abstraction and high-level video segmentation
approaches. Therefore solving the problem of shot boundary
detection is one of the major prerequisites for revealing
higher level video content structure. As a crucial step in video
indexing and retrieval, accurate shot boundary detection plays
an important role to organize and summarize video content
into meaningful shots for further scene analysis. Many
algorithms have been proposed for detecting video shot
boundaries and classifying shots and shot transition types. In
this paper we propose a novel technique for shot boundary
detection using radon transform. We first removed the effect
of illumination using DCT and DWT. Then shot boundary is
detected using radon transform. Radon transform is based on
projection of image intensity along a radial line oriented at a
specific angle. Projection of image intensity for current frame
is different than that of projection of the previous frame, where
shot boundary is detected. In order to verify the performance
of algorithm, experiments have been carried out with news,
documentary and movie. Experimental result demonstrates
efficiency of proposed shot boundary detection technique.
A novel predicate for active region merging in automatic image segmentationeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
A novel predicate for active region merging in automatic image segmentationeSAT Journals
Abstract Image segmentation is an elementary task in computer vision and image processing. This paper deals with the automatic image segmentation in a region merging method. Two essential problems in a region merging algorithm: order of merging and the stopping criterion. These two problems are solved by a novel predicate which is described by the sequential probability ratio test and the minimal cost criterion. In this paper we propose an Active Region merging algorithm which utilizes the information acquired from perceiving edges in color images in L*a*b* color space. By means of color gradient recognition method, pixels with no edges are clustered and considered alone to recognize some preliminary portion of the input image. The color information along with a region growth map consisting of completely grown regions are used to perform an Active region merging method to combine regions with similar characteristics. Experiments on real natural images are performed to demonstrate the performance of the proposed Active region merging method. Index Terms: Adaptive threshold generation, CIE L*a*b* color gradient, region merging, Sequential Probability Ratio Test (SPRT).
ROLE OF HYBRID LEVEL SET IN FETAL CONTOUR EXTRACTIONsipij
Image processing technologies may be employed for quicker and accurate diagnosis in analysis and
feature extraction of medical images. Here, existing level set algorithm is modified and it is employed for
extracting contour of fetus in an image. In traditional approach, fetal parameters are extracted manually
from ultrasound images. An automatic technique is highly desirable to obtain fetal biometric measurements
due to some problems in traditional approach such as lack of consistency and accuracy. The proposed
approach utilizes global & local region information for fetal contour extraction from ultrasonic images.
The main goal of this research is to develop a new methodology to aid the analysis and feature extraction.
Role of Hybrid Level Set in Fetal Contour Extractionsipij
Image processing technologies may be employed for quicker and accurate diagnosis in analysis and
feature extraction of medical images. Here, existing level set algorithm is modified and it is employed for
extracting contour of fetus in an image. In traditional approach, fetal parameters are extracted manually
from ultrasound images. An automatic technique is highly desirable to obtain fetal biometric measurements
due to some problems in traditional approach such as lack of consistency and accuracy. The proposed
approach utilizes global & local region information for fetal contour extraction from ultrasonic images.
The main goal of this research is to develop a new methodology to aid the analysis and feature extraction.
2-Dimensional Wavelet pre-processing to extract IC-Pin information for disarr...IOSR Journals
Abstract: Due to higher processing power to cost ratio, it is now possible to replace the manual detection methods used in the IC (Integrated Circuit) industry by Image-processing based automated methods, to detect a broken pin of an IC connected on a PCB during manufacturing, which will make the process faster, easier and cheaper. In this paper an accurate and fast automatic detection method is used where the top view camera shots of PCBs are processed using advanced methods of 2-dimensional discrete wavelet pre-processing before applying edge-detection. Comparison with conventional edge detection methods such as Sobel, Prewitt and Canny edge detection without 2-D DWT is also performed. Keywords :2-dimensional wavelets, Edge detection, Machine vision, Image processing, Canny.
Matching algorithm performance analysis for autocalibration method of stereo ...TELKOMNIKA JOURNAL
Stereo vision is one of the interesting research topics in the computer vision field. Two cameras are used to generate a disparity map, resulting in the depth estimation. Camera calibration is the most important step in stereo vision. The calibration step is used to generate an intrinsic parameter of each camera to get a better disparity map. In general, the calibration process is done manually by using a chessboard pattern, but this process is an exhausting task. Self-calibration is an important ability required to overcome this problem. Self-calibration required a robust and good matching algorithm to find the key feature between images as reference. The purpose of this paper is to analyze the performance of three matching algorithms for the autocalibration process. The matching algorithms used in this research are SIFT, SURF, and ORB. The result shows that SIFT performs better than other methods.
REGION CLASSIFICATION BASED IMAGE DENOISING USING SHEARLET AND WAVELET TRANSF...csandit
This paper proposes a neural network based region classification technique that classifies
regions in an image into two classes: textures and homogenous regions. The classification is
based on training a neural network with statistical parameters belonging to the regions of
interest. An application of this classification method is applied in image denoising by applying
different transforms to the two different classes. Texture is denoised by shearlets while
homogenous regions are denoised by wavelets. The denoised results show better performance
than either of the transforms applied independently. The proposed algorithm successfully
reduces the mean square error of the denoised result and provides perceptually good results.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
REGION CLASSIFICATION BASED IMAGE DENOISING USING SHEARLET AND WAVELET TRANSF...cscpconf
This paper proposes a neural network based region classification technique that classifies
regions in an image into two classes: textures and homogenous regions. The classification is
based on training a neural network with statistical parameters belonging to the regions of
interest. An application of this classification method is applied in image denoising by applying
different transforms to the two different classes. Texture is denoised by shearlets while
homogenous regions are denoised by wavelets. The denoised results show better performance
than either of the transforms applied independently. The proposed algorithm successfully
reduces the mean square error of the denoised result and provides perceptually good results.
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...Vladimir Kulyukin
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Forager Traffic Levels from Images in Solar-Powered, Electronic Beehive Monitoring
Many problems in information retrieval and related fields depend on a reliable measure of the distance or similarity between objects that, most frequently, are represented
as vectors. This paper considers vectors of bits. Such data structures implement entities as diverse as bitmaps that indicate the occurrences of terms and bitstrings indicating the presence
of edges in images. For such applications, a popular distance measure is the Hamming distance. The value of the Hamming distance for information retrieval applications is limited by the
fact that it counts only exact matches, whereas in information retrieval, corresponding bits that are close by can still be considered to be almost identical. We define a "Generalized
Hamming distance" that extends the Hamming concept to give partial credit for near misses, and suggest a dynamic programming algorithm that permits it to be computed efficiently.
We envision many uses for such a measure. In this paper we define and prove some basic properties of the :Generalized Hamming distance," and illustrate its use in the area of object
recognition. We evaluate our implementation in a series of experiments, using autonomous robots to test the measure's effectiveness in relating similar bitstrings.
Adapting Measures of Clumping Strength to Assess Term-Term SimilarityVladimir Kulyukin
Automated information retrieval relies heavily on statistical regularities that emerge as terms are deposited to produce text. This paper examines statistical patterns expected of a pair of
terms that are semantically related to each other. Guided by a conceptualization of the text generation process, we derive measures of how tightly two terms are semantically associated.
Our main objective is to probe whether such measures yields reasonable results. Specifically, we examine how the tendency of a content bearing term to clump, as quantified by previously
developed measures of term clumping, is influenced by the presence of other terms. This
approach allows us to present a toolkit from which a range of measures can be constructed.
As an illustration, one of several suggested measures is evaluated on a large text corpus built from an on-line encyclopedia.
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...Vladimir Kulyukin
V. Kulyukin & T. Zaman. "Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxed Pitch, Roll, and Yaw Camera A lignment Constraints." International Journal of Image Processing (IJIP), V olume (8) : Issue (5) : 2014, pp. 355-383.
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...Vladimir Kulyukin
The work presented in this article continues our investigation of such assisted navigation solutions where the
main emphasis is placed not on sensor sets or sensor fusion algorithms but on the ability of the travelers to interpret and
contextualize verbal route directions en route. This work contributes to our investigation of the research hypothesis that
we have formulated and partially validated in our previous studies: if a route is verbally described in sufficient and
appropriate amount of detail, independent VI travelers can use their O&M and problem solving skills to successfully
follow the route without any wearable sensors or sensors embedded in the environment.
In this investigation, we temporarily put aside the issue of how VI and blind travelers successfully interpret route
directions en route and tackle the question of how those route directions can be created, generated, and maintained by
online communities. In particular, we focus on the automation of path inference and present an algorithm that may be used
as part of the background computation of VGI sites to find new paths in the previous route directions written by online
community members, generate new route descriptions from them, and post them for subsequent community editing.
Wireless Indoor Localization with Dempster-Shafer Simple Support FunctionsVladimir Kulyukin
A mobile robot is localized in an indoor environment
using IEEE 802.11b wireless signals. Simple support
functions of the Dempster-Shafer theory are used to combine evidence
from multiple localization algorithms. Emperical results
are presented and discussed. Conclusions are drawn regarding
when the proposed sensor fusion methods may improve performance
and when they may not.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Seminar of U.V. Spectroscopy by SAMIR PANDASAMIR PANDA
Spectroscopy is a branch of science dealing the study of interaction of electromagnetic radiation with matter.
Ultraviolet-visible spectroscopy refers to absorption spectroscopy or reflect spectroscopy in the UV-VIS spectral region.
Ultraviolet-visible spectroscopy is an analytical method that can measure the amount of light received by the analyte.
Multi-source connectivity as the driver of solar wind variability in the heli...Sérgio Sacani
The ambient solar wind that flls the heliosphere originates from multiple
sources in the solar corona and is highly structured. It is often described
as high-speed, relatively homogeneous, plasma streams from coronal
holes and slow-speed, highly variable, streams whose source regions are
under debate. A key goal of ESA/NASA’s Solar Orbiter mission is to identify
solar wind sources and understand what drives the complexity seen in the
heliosphere. By combining magnetic feld modelling and spectroscopic
techniques with high-resolution observations and measurements, we show
that the solar wind variability detected in situ by Solar Orbiter in March
2022 is driven by spatio-temporal changes in the magnetic connectivity to
multiple sources in the solar atmosphere. The magnetic feld footpoints
connected to the spacecraft moved from the boundaries of a coronal hole
to one active region (12961) and then across to another region (12957). This
is refected in the in situ measurements, which show the transition from fast
to highly Alfvénic then to slow solar wind that is disrupted by the arrival of
a coronal mass ejection. Our results describe solar wind variability at 0.5 au
but are applicable to near-Earth observatories.
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...Scintica Instrumentation
Intravital microscopy (IVM) is a powerful tool utilized to study cellular behavior over time and space in vivo. Much of our understanding of cell biology has been accomplished using various in vitro and ex vivo methods; however, these studies do not necessarily reflect the natural dynamics of biological processes. Unlike traditional cell culture or fixed tissue imaging, IVM allows for the ultra-fast high-resolution imaging of cellular processes over time and space and were studied in its natural environment. Real-time visualization of biological processes in the context of an intact organism helps maintain physiological relevance and provide insights into the progression of disease, response to treatments or developmental processes.
In this webinar we give an overview of advanced applications of the IVM system in preclinical research. IVIM technology is a provider of all-in-one intravital microscopy systems and solutions optimized for in vivo imaging of live animal models at sub-micron resolution. The system’s unique features and user-friendly software enables researchers to probe fast dynamic biological processes such as immune cell tracking, cell-cell interaction as well as vascularization and tumor metastasis with exceptional detail. This webinar will also give an overview of IVM being utilized in drug development, offering a view into the intricate interaction between drugs/nanoparticles and tissues in vivo and allows for the evaluation of therapeutic intervention in a variety of tissues and organs. This interdisciplinary collaboration continues to drive the advancements of novel therapeutic strategies.
Cancer cell metabolism: special Reference to Lactate PathwayAADYARAJPANDEY1
Normal Cell Metabolism:
Cellular respiration describes the series of steps that cells use to break down sugar and other chemicals to get the energy we need to function.
Energy is stored in the bonds of glucose and when glucose is broken down, much of that energy is released.
Cell utilize energy in the form of ATP.
The first step of respiration is called glycolysis. In a series of steps, glycolysis breaks glucose into two smaller molecules - a chemical called pyruvate. A small amount of ATP is formed during this process.
Most healthy cells continue the breakdown in a second process, called the Kreb's cycle. The Kreb's cycle allows cells to “burn” the pyruvates made in glycolysis to get more ATP.
The last step in the breakdown of glucose is called oxidative phosphorylation (Ox-Phos).
It takes place in specialized cell structures called mitochondria. This process produces a large amount of ATP. Importantly, cells need oxygen to complete oxidative phosphorylation.
If a cell completes only glycolysis, only 2 molecules of ATP are made per glucose. However, if the cell completes the entire respiration process (glycolysis - Kreb's - oxidative phosphorylation), about 36 molecules of ATP are created, giving it much more energy to use.
IN CANCER CELL:
Unlike healthy cells that "burn" the entire molecule of sugar to capture a large amount of energy as ATP, cancer cells are wasteful.
Cancer cells only partially break down sugar molecules. They overuse the first step of respiration, glycolysis. They frequently do not complete the second step, oxidative phosphorylation.
This results in only 2 molecules of ATP per each glucose molecule instead of the 36 or so ATPs healthy cells gain. As a result, cancer cells need to use a lot more sugar molecules to get enough energy to survive.
Unlike healthy cells that "burn" the entire molecule of sugar to capture a large amount of energy as ATP, cancer cells are wasteful.
Cancer cells only partially break down sugar molecules. They overuse the first step of respiration, glycolysis. They frequently do not complete the second step, oxidative phosphorylation.
This results in only 2 molecules of ATP per each glucose molecule instead of the 36 or so ATPs healthy cells gain. As a result, cancer cells need to use a lot more sugar molecules to get enough energy to survive.
introduction to WARBERG PHENOMENA:
WARBURG EFFECT Usually, cancer cells are highly glycolytic (glucose addiction) and take up more glucose than do normal cells from outside.
Otto Heinrich Warburg (; 8 October 1883 – 1 August 1970) In 1931 was awarded the Nobel Prize in Physiology for his "discovery of the nature and mode of action of the respiratory enzyme.
WARNBURG EFFECT : cancer cells under aerobic (well-oxygenated) conditions to metabolize glucose to lactate (aerobic glycolysis) is known as the Warburg effect. Warburg made the observation that tumor slices consume glucose and secrete lactate at a higher rate than normal tissues.
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed Barcode Scanning
1. IPCV 2015
Image Blur Detection with 2D Haar Wavelet
Transform and Its Effect on Skewed Barcode
Scanning
Vladimir Kulyukin
Department of Computer Science
Utah State University
Logan, UT, USA
vladimir.kulyukin@usu.edu
Sarat Andhavarapu
Department of Computer Science
Utah State University
Logan, UT, USA
sarat.andhavarapu@aggiemail.usu.edu
Abstract—An algorithm is presented for image blur detection
with the 2D Haar Wavelet transform (2D HWT). The algorithm
classifies an image as blurred or sharp by splitting it into N x N
tiles, applying several iterations of the 2D HWT to each tile, and
grouping horizontally, vertically, and diagonally connected tiles
with pronounced changes into tile clusters. Images with large tile
clusters are classified as sharp. Images with small tile clusters are
classified as blurred. If need be, the blur extent can be estimated
as the ratio of the total area of the connected tile clusters and the
area of the image. When evaluated on a sample of five hundred
images, the algorithm performed on par or better than two other
blur detection algorithms found in the literature. The effect of
blur detection on skewed barcode scanning is investigated by
integrating the presented blur detection algorithm into a skewed
barcode scanning algorithm. The experimental results indicate
that blur detection had a positive effect on skewed barcode
scanning rates.
Keywords—computer vision; image blur detection; Haar
wavelets, 2D Haar wavelet transform, barcode scanning
I. Introduction
In our previous research [1, 2], we developed an algorithm for
in-place vision-based skewed barcode scanning with relaxed
pitch, roll, and yaw camera alignment constraints. The skewed
barcode scanning experiments were conducted on a set of 506
video recordings of common grocery products. Our
experiments showed that the scanning results were
substantially higher on sharp images than on blurred ones. A
limitation of that algorithm was that it did not filter the blurred
frames out of the barcode localization and scanning process.
The same limitation was experimentally discovered in
another algorithm that we developed for mobile vision-based
localization of skewed nutrition labels (NLs) on grocery
packages that maximizes specificity, i.e., the percentage of
true negative matches out of all possible negative matches [3].
The NL localization algorithm works on frames captured from
the smartphone camera’s video stream and localizes NLs
skewed up to 35-40 degrees in either direction from the
vertical axis of the captured frame.
The NL localization algorithm uses three image processing
methods: edge detection, line detection, and corner detection.
We experimentally discovered that the majority of false
negative matches were caused by blurred images. Both the
Canny edge detector [4] and dilate-erode corner detector [5]
used in the algorithm require rapid and contrasting changes to
identify key points and lines of interest. These data cannot be
readily retrieved from blurred images, which results in run-
time barcode scanning and NL localization failures.
Consequently, effective image blur detection methods will
likely improve both skewed barcode scanning and NL
localization rates.
Toward this end, in this paper, an algorithm is presented
for image blur detection based on the 2D HWT [6]. The
algorithm classifies an image as blurred or sharp by splitting it
into N x N tiles, applying several iterations of the 2D HWT to
each tile, and grouping the horizontally, vertically, and
diagonally connected tiles with pronounced changes into
clusters. Images with large clusters are classified as sharp
whereas images with small tile clusters are classified as
blurred. If need be, the blur extent can be estimated as the
ratio of the total area of the connected tile clusters and the area
of the image. The effect of blur detection on skewed barcode
scanning is investigated by integrating the blur detection
algorithm into our in-place vision-based skewed barcode
scanning with relaxed pitch, roll, and yaw camera alignment
constraints [1, 2]. The experimental results indicate that blur
detection improves skewed barcode scanning rates.
The remainder of our paper is organized as follows. In
Section II, we present and analyze related work. In Section III,
we outline the details of our image blur detection algorithm. In
Section IV, we present our experiments with the blur detection
algorithm and experimentally compare the algorithm’s
performance with two other blur detection algorithms found in
the literature [7, 8]. In Section V, the results of the
experiments are discussed. Section VI summarizes our
findings, presents our conclusions, and outlines some research
venues we would like to pursue in the future.
II. Related Work
A. Blur Detection
Mallat and Hwang [9] mathematically prove that signals carry
information via irregular structures and singularities. In
particular, they show that the local maxima of the wavelet
transform detect the locations of irregularities. For example,
2. IPCV 2015
the 2D wavelet transform maxima indicate the locations of
edges in images. The Fourier analysis [10], which has been
traditionally used in physics and mathematics to investigate
irregularities, is not always suitable to detecting the spatial
distribution of such irregularities.
According to Tong et al. [7], image blur detection methods
can be broadly classified as direct or indirect. Indirect methods
characterize image blur as a linear function 𝐼 𝐵 = 𝐵 ∙ 𝐼 𝑂 + 𝑁,
where 𝐼 𝑂 is the original image, 𝐵 is an unknown image blur
function, 𝑁 is a noise function, and 𝐼 𝐵 is the resulting image
after the introduction of the blur and noise.
Indirect methods consider 𝐵 unknown and use various
techniques to estimate it. Rooms et al. [11] propose a wavelet-
based method to estimate the blur of an image by looking at
the sharpness of the sharpest edges in the image. The Lipschitz
exponents [12] are computed for the sharpest edges and a
relation between the variance of a Gaussian point spread
function and the magnitude of the Lipschitz exponent is shown
to be dependent on the blur present in the image and not on the
image contents.
Venkatakrishnan et al. [13] show that the wavelet
transform modulus maxima (WTMM) detect all the
singularities of a function and describe strategies to measure
their regularity and propose an algorithm for characterizing
singularities of irregular signals. The researchers present a
method for measuring the Lipschitz exponents that uses the
area between the straight line satisfying specific properties and
the curve of the WTMM in a finite scale interval in the log-log
plot of scales versus WTMM as the objective function.
Pavlovic and Tekalp [14] propose a formulation of the
maximum likelihood (ML) blur identification based on
parametric modeling of the blur in the continuous spatial
coordinates. Unlike ML blur identification methods based on
discrete spatial domain blur models, their method finds the
ML estimate of the extent and the parameters of arbitrary
point spread functions that admit a closed form parametric
description in the continuous coordinates. Experiments show
significant results for the cases of 1D uniform motion blur, 2D
out-of-focus blur, and 2D truncated Gaussian blur at different
signal-to-noise ratios.
Panchapakesan et al. [15] present an indirect method for
image blur identification from vector quantizer encoder
distortion. The method takes a set of training images from all
candidate blur functions. These sets are used to train vector
quantizer encoders. The a-priori unknown blur function is
identified from a blurred image by choosing among the
candidate vector quantizer encoders the encoder with the
lowest distortion. The researchers investigated two training
methods: the generalized Lloyd algorithm and a non-iterative
discrete cosine transform (DCT)-based approach.
Direct methods estimate blur extent on the basis of some
distinctive features directly found in images such as edges,
corners, or discrete cosine transform (DCT) coefficients.
Marichal et al. [16] estimate image blur based on the
histogram computation of non-zero DCT coefficients
computed from MPEG or JPEG compressed images. The
proposed method takes into account the DCT information
from the entire image. A key assumption is that any edge type
will likely cross some 8 x 8 blocks at least once in the image.
The camera and motion blur is estimated through the
globalization among all DCT blocks.
Figure 1. Edge classification
Tong et al. [7] propose a direct method similar to the one
proposed in this paper in that it also uses the 2D Haar Wavelet
Transform (2D HWT). Their method is based on the
assumption that the introduction of blur has different effects
on the four main types of edges shown in Figure 1: Dirac, A-
Step, G-Step, and Roof. It is claimed that in blurred images
the Dirac and A-Step edges disappear while G-Step and Roof
edges lose their sharpness. The method classifies an image as
blurred on the basis of the presence or absence of Dirac and
A-Step edges and estimates the blur extent as the percentage
of G-Step and Roof edges present in the image.
The algorithm presented in this paper is also based on the
2D HWT. However, it does not extract any explicit
morphological features such as edges or corners from the
image. Instead, it uses the 2D HWT to detect regions with
pronounced changes and combines those regions into larger
segments without explicitly computing the causes of those
changes. Since the algorithm is based on the 2D HWT, the
regions are square tiles whose side is an integral power of 2.
This algorithm continues our investigation of vision-based
barcode and nutrition label scanning on mobile phones with
relaxed pitch, yaw, and roll constraints [1, 2]. As we have
previously reported, most false negatives in our experiments
were caused by blurred images. While newer models of
smartphones will likely have improved camera stability and
focus, software techniques to detect blurred images can still
make vision-based skewed barcode scanning and nutrition
information extraction more reliable and efficient. Since our
barcode scanning and nutrition information extraction
algorithms are cloud-based, eliminating blurred images from
processing will likely improve the network throughput and
decrease data plan consumption rates.
B. 2D Haar Transform
3. IPCV 2015
Our implementation of the 2D HWT is based on the approach
taken in [6] where the transition from 1D Haar wavelets to 2D
Haar wavelets is based on the products of basic wavelets in the
first dimension with basic wavelets in the second dimension.
For a pair of functions 𝑓1 and 𝑓2 their tensor product is defined
as (𝑓1 × 𝑓2)(𝑥, 𝑦) = 𝑓1(𝑥) ∙ 𝑓2(𝑥). Two 1D basic wavelet
functions are defined as follows:
𝜑[0,1[(𝑟) = {
1 𝑖𝑓 0 ≤ 𝑟 < 1,
0 𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒.
𝜓[0,1[(𝑟) =
{
1 𝑖𝑓 0 ≤ 𝑟 <
1
2
,
−1 𝑖𝑓
1
2
≤ 𝑟 < 1,
0 𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒.
The 2D Haar wavelets are defined as tensor products of
𝜑[0,1[(𝑟) and 𝜓[0,1[(𝑟): 𝛷0,0
(0)
(𝑥, 𝑦) = (𝜑[0,1[ × 𝜑[0,1[)(𝑥, 𝑦),
𝛹0,0
ℎ,(0)
(𝑥, 𝑦) = (𝜑[0,1[ × 𝜓[0,1[)(𝑥, 𝑦), 𝛹0,0
𝑣,(0)
(𝑥, 𝑦) =
(𝜓[0,1[ × 𝜑[0,1[)(𝑥, 𝑦), 𝛹0,0
𝑑,(0)
(𝑥, 𝑦) = (𝜓[0,1[ × 𝜓[0,1[)(𝑥, 𝑦).
The superscripts h, v, and d indicate the correspondence of
these wavelets with horizontal, vertical, and diagonal changes,
respectively. The horizontal wavelets detect horizontal (left to
right) changes in 2D data, the vertical wavelets detect vertical
(top to bottom) changes in 2D data, and the diagonal changes
detect diagonal changes in 2D data.
In practice, the basic 2D HWT is computed by applying a
1D wavelet transform of each row and then a 1D wavelet
transform of each column. Suppose we have a 2 x 2 pixel
image
[
𝑠0,0 𝑠0,1
𝑠1,0 𝑠1,1
] = [
11 9
7 5
].
Applying a 1D wavelet transform to each row results in the
following 2 x 2 matrix:
[
𝑠0,0 + 𝑠0,1
2
𝑠0,0 − 𝑠0,1
2
𝑠1,0 + 𝑠1,1
2
𝑠1,0 − 𝑠1,1
2 ]
=
[
11 + 9
2
11 − 9
2
7 + 5
2
7 − 5
2 ]
= [
10 1
6 1
].
Applying a 1D wavelet transform to each new column is
fetches us the result 2 x 2 matrix:
[
10+6
2
1+1
2
10−6
2
1−1
2
]
= [
8 1
2 0
].
The coefficients in the result matrix obtained after the
application of the 1D transform to the columns express the
original data in terms of the four tensor product wavelets
𝛷0,0
(0)
(𝑥, 𝑦), 𝛹0,0
ℎ,(0)
(𝑥, 𝑦), 𝛹0,0
𝑣,(0)
(𝑥, 𝑦), and 𝛹0,0
𝑑,(0)
(𝑥, 𝑦):
[
11 9
7 5
] = 8𝛷0,0
(0)
(𝑥, 𝑦) + 1𝛹0,0
ℎ,(0)
(𝑥, 𝑦) +
2𝛹0,0
𝑣,(0)
(𝑥, 𝑦) + 0𝛹0,0
𝑑,(0)
(𝑥, 𝑦).
The value 8 in the upper-left corner is the average value of
the original matrix: (11+9+7+5)/4= 8. The value 1 in the upper
right-hand corner is the horizontal change in the data from the
left average, (11+7)/2=9, to the right average, (9+5)/2=7,
which is equal 1 ∙ 𝛹0,0
ℎ,(0)
(𝑥, 𝑦) = 1 ∙ −2. The value 2 in the
bottom-left corner is the vertical change in the original data
from the upper average, (11+9)/2=10, to the lower average,
(7+5)/2=6, which is equal to 2 ∙ 𝛹0,0
𝑣,(0)
(𝑥, 𝑦) = 2 ∙ −2=−4.
The value 0 in the bottom-right corner is the change in the
original data from the average along the first diagonal (from
the top left corner to the bottom right corner), (11+5)/2=8, to
the average along the second diagonal (from the top right
corner to the bottom left corner), (9+7)/2=8, which is equal to
0 ∙ 𝛹0,0
𝑑,(0)
(𝑥, 𝑦). The decomposition operation can be
represented in terms of matrices:
[
11 9
7 5
] = 8 [
1 1
1 1
] + 1 [
1 − 1
1 − 1
] + 2 [
1 1
−1 − 1
] +
0 [
1 − 1
−1 1
].
III. Blur Detection Algorithm
The first stage of the blur detection algorithm is to find image
regions that have pronounced horizontal, vertical, or diagonal
changes. A captured frame is divided into N x N windows,
called tiles, where 𝑁 = 2 𝑘
, 𝑘 ∈ 𝑍. Figure 2 shows square tiles
of size 𝑁 = 64. The border pixels at the right and bottom
margins are discarded when captured frames are not evenly
divisible by N. A candidate tile must have a pronounced
change along at least one of the three directions: horizontal,
vertical, or diagonal. Whether a change is pronounced or not is
determined through a threshold.
Figure 2. Tile splitting
4. IPCV 2015
Each tile is processed by four iterations of the 2D Haar
transform. The number of iterations and the tile size are
parameters and can be made either smaller or larger. The
values reported in this paper were experimentally found to
work well in our domain of vision-based skewed barcode
scanning and nutrition information extraction [1, 2, 3].
Let HC be the horizontal change between the left half of
the tile and the right half of the tile. Let VC be the vertical
change between the upper half of the tile and the lower half of
the tile. Let DC be the change between the first diagonal (top
left to bottom right) and the second diagonal (top right to
bottom left). If at least one of the values HC, VC, and DC is
above the corresponding thresholds 𝐻𝐶 𝜃, 𝑉𝐶 𝜃, and 𝐷𝐶 𝜃,
respectively, the tile is marked as having a pronounced
change. Figure 3 (left) shows the tiles with pronounced
changes marked as squares.
Figure 3. Squre tiles with pronounced changes (left); Tile
clusters found by DFS (right)
Figure 4. Tiles with pronounced changes in a blurred
image
After the tiles with pronounced changes are found, the
depth-first search (DFS) is used to combine them into tile
clusters. The DFS starts with an unmarked tile with a
pronounced horizontal, vertical, or diagonal change and
proceeds by connecting to its immediate horizontal, vertical,
and diagonal tile neighbors if they have pronounced changes.
If such tiles are found, they are marked with the same cluster
number and the search continues recursively. The search stops
when no other tiles can be added to the current cluster. The
algorithm continues to look for another unmarked tile to which
it can apply the DFS. If no such tile is found, the algorithm
terminates. Figure 3 (right) shows five DFS-found tile
clusters.
Figure 5. DFS-found tile clusters
Figure 6. Pseudocode of the blur detection algorithm
After the iterative applications of the DFS have found the
tile clusters, two cluster-related rules are used to classify a
whole image as sharp or blurred. The first rule is the
percentage of the total area of the image covered by the
clusters. The second rule uses the number of the tiles in each
cluster to discard small clusters.
The first rule captures the intuition that in a sharp image
there are many tiles with pronounced changes. The second
rule discards small clusters whose size, i.e., the number of tiles
in the cluster, is below a given threshold. This cluster weeding
rule is currently based on a threshold of 5. In other words, any
1. DetectBlur(Img, N, NITER, 𝐻𝐶 𝜃, 𝑉𝐶 𝜃, 𝐷𝐶 𝜃, CSZ, A)
2. {
3. FOR each N x N tile T in image Img {
4. [AVRG, HC, VC, DC] = 2DHWT(T, NITER);
5. AH=Avrg(HC); AV=Avrg(VC); AD=Avrg(DC);
6. IF (AH > 𝐻𝐶 𝜃 or AV > 𝑉𝐶 𝜃 or AD > 𝐷𝐶 𝜃)
7. Mark T as having pronounced change;
8. }
9. FOR each N x N tile T in image Img {
10. IF ( T is not in any tile cluster )
11. Run DFS(T) to find and mark all tiles in the
12. same cluster;
13. }
14. TotalArea = 0;
15. FOR each tile cluster TC {
16. IF ( TC’s size > CSZ ) {
17. ClusterArea = TC’s size * N x N;
18. TotalArea += ClusterArea;
19. }
20. }
21. IF ( TotalArea/Area(Img) < A ) Return True;
22. ELSE Return False;
22. }
5. IPCV 2015
cluster with fewer than five tiles is rejected by the second rule
and does not contribute to the area of the image with
pronounced changes. Thus, in Figure 3, only two clusters are
left after the application of the second rule: the cluster with 18
tiles and the cluster with 6 tiles. Since both clusters are large,
the image is classified as sharp by the first rule. The three
singletons are discarded. On the other hand, all clusters found
in the blurred image of Figure 4 are shown in Figure 5. Since
all of them are small, they are discarded by the second rule.
Figure 6 gives the pseudocode of our blur detection
algorithm. The first argument to the DetectBlur function is
the image, the second argument is the size of the square tile
into which the image is split, as shown in Figure 2. In our
experiments presented in the next section, N=64. The next
argument, NITER, is the number of iterations of the 2DHWT
function runs in each square tile in line 4. In our current
implementation, NITER=4. Our Java source of the 2DHWT
function is available at [17]. This function returns an array of
four matrices: AVRG, HC, VC, and DC. The matrix AVRG
contains the average numbers after all iterations and the
matrices HC, VC, and DC contain the horizontal, vertical, and
diagonal wavelet coefficients, respectively.
If at least one of the averages of the HC, VC, or DC
matrices after all NITER iterations is above a corresponding
threshold, which is computed in lines 5 and 6, then the
appropriate tile is marked as having pronounced change. In
lines 9-13, the DFS is used to find all tile clusters in the image,
as shown in Figure 3 (right).
In lines 14-20, the two rules described above to compute
the total area occupied by the clusters greater than the value of
the cluster threshold parameter CSZ. In lines 21-22, if the
percentage of the total area occupied by the clusters with
pronounced changes is smaller than the threshold value
specified by the last parameter A, the image is classified as
blurred. If need be, the algorithm can be modified to return the
blur extent as the ratio of the total area occupied by the large
tile clusters and the total area of the image. The smaller the
ratio, the more blur exists in the image.
The outlined algorithm is based on the assumption that in
blurred images square tiles with pronounced changes do not
form large clusters but scatter across the image as singletons
or form clusters whose combined area is small relative to the
size of the image.
For example, Figure 4 is a blurred image where 64 x 64
tiles with pronounced changes are marked. Figure 5 shows the
tile clusters found by the iterative applications of the DFS to
the image in Figure 4. As can be seen in Figure 5, most
clusters are either singletons or are of size 2. The largest
cluster in the bottom left of the image is of size 4.
IV. Experiments
We took five hundred random RGB images from a set of 506
video recordings of common grocery products that we made
publicly available in our previous field investigations of
skewed barcode scanning [18]. The videos have a 1280 x 720
resolution and an average duration of 15 seconds. All videos
were recorded on an Android 4.2.2 Galaxy Nexus smartphone
in a supermarket in Logan, UT. All videos were taken by an
operator who held a grocery product in one hand and a
smartphone in the other. The videos covered four different
categories of products: bags, boxes, bottles, and cans.
Three human volunteers were recruited to classify each of
the five hundred images as blurred or sharp. An image was
classified as blurred if at least two volunteers classified it as
blurred. It was otherwise classified as sharp. The human
evaluation resulted in 167 blurred images and 333 sharp
images. These results were used as the ground truth.
We compared our algorithm with two other image blur
detection algorithms frequently cited in the literature [7, 8].
Since we could not find the source code of [7] publicly
available online, we implemented it ourselves in Python. Our
Python source code is available at [19]. We found a MATLAB
implementation of the other algorithm at [20] and used it for
the experiments.
Table 1 gives the numbers of true and false positives for all
three algorithms. The columns TPB and FPB give the numbers
of true and false positives, respectively, for the blurred
images. The columns TPS and FPS give the numbers of true
and false positives, respectively, for the sharp images. The
row Algo 1 gives the statistics for our algorithm implemented
in Java. The rows Algo 2 and Algo 3 give the statistics for the
Python implementation of [19] and the MATLAB
implementation of [8], respectively.
Table 1. True and false positives
Algorithm TPB FPB TPS FPS
Algo 1 163 4 254 79
Algo 2 167 0 183 150
Algo 3 81 86 268 65
To compare the performance numbers of each algorithm
with the ground truth, we used the relative difference
percentage, which is a unitless measure that compares two
quantities while taking into account their magnitudes. Table 2
gives the relative difference percentages computed as |𝑥– 𝑦|/
max(|𝑥|, |𝑦|) ∙ 100, where y is the humanly estimated number
of blurred or sharp images, i.e., the ground truth, and x is the
number of sharp or blurred images found by a given
algorithm. For example, for Algo 1, the first relative
difference is computed as |163– 167|/max(|163|, |167|) ∙
100=2.39, where 163 is the number of blurred images found
by Algo 1 and 167 is the number of blurred images found by
the human evaluators.
Table 2. Relative differences
Algorithm Blurred Sharp
Algo 1 2.39 23.72
Algo 2 0.00 45.05
Algo 3 51.50 19.52
Table 3. Effect of blur on barcode scanning I
Sample Sharp Blurred Barcode
in sharp
Barcodes
in blurred
1 15 15 12 1
2 13 17 11 0
3 16 14 12 0
We investigated the effect of image blur on skewed
barcode scanning. We chose three random samples of 30
6. IPCV 2015
images from the 500 images classified by the three human
evaluators. In each sample, 15 images were classified as
blurred and 15 as sharp. We integrated our blur detection
algorithm into our cloud-based barcode scanning algorithm
and estimated the effect of accurate image blur detection on
skewed barcode scanning. Tables 3 and 4 give the results of
our experiments.
In Table 3, the first column gives the sample numbers. The
column Sharp gives the number of sharp images classified as
sharp by our algorithm. The column Blurred gives the number
of images classified as blurred by our algorithm. The column
Barcode in sharp gives the number of barcodes correctly
scanned in the images classified as sharp. The column
Barcodes in blurred gives the number of barcodes correctly
scanned in the images classified by our algorithm as blurred.
Thus, in sample 1, all blurred and sharp images were classified
accurately. However, in the 15 sharp images, the barcode
scanner accurately scanned 12 barcodes whereas in the 15
blurred images, the barcode scanner accurately scanned only 1
barcode.
In sample 2, 13 out of 15 images were accurately classified
as sharp with 2 false negatives and 17 images were classified
as blurred with 2 false positives. In 11 images classified as
sharp, the barcodes were accurately scanned. No barcodes
were accurately scanned in the images classified as blurred.
In sample 3, 16 images were classified as sharp with 1
false positive and 14 images were classified as blurred with 1
false negative. Barcodes were successfully scanned in 12
images classified as sharp. No barcodes were scanned in the
images classified as blurred.
Table 4. Effect of blur on barcode scanning II
Sample Blurred Sharp Total Gain
1 1/15 12/15 13/30 0.37
2 0/17 11/13 11/30 0.50
3 0/16 12/14 12/30 0.46
Table 4 gives the results of the effect of blur detection on
skewed barcode scanning. The first column gives the numbers
of the random samples. The second column records the ratio
of accurately scanned barcodes in the images classified as
blurred. The third column records the ratio of accurately
scanned barcodes in the images classified as sharp. The fourth
column gives the ratio of recognized barcodes in all images.
The fifth column gives the gain measured as the difference
between the ratio of the accurately recognized barcodes only
in the sharp images and the ratio of the accurately recognized
barcodes in all images, which estimates the effect of blur
detection on barcode scanning. Thus, in sample 1, we increase
the barcode scanning rate by 37 percent if we eliminate
images classified as blurred from barcode scanning. In sample
2, if blurred images are eliminated from barcode scanning, we
gain 50 percent in barcode scanning rates. In sample 3, the
gain is 46 percent.
V. Results
In discussing the results of the experiments, we will again
refer to our algorithm as Algo 1, to the algorithm by Tong et
al. [7] as Algo 2, and to the algorithm by [8] as Algo 3. The
experiments indicate (see Table 1) that, on our sample of
images, in classifying images as blurred. Algo 1 performs as
well as Algo 2 and outperforms Algo 3. In classifying images
as sharp, Algo 1 performs as well as Algo 3 and outperforms
Algo 2.
Table 2 confirms the observations recorded in Table 1. In
image blur detection, there is almost no difference between
Algo 1 and Algo 2 in that these algorithms do not deviate from
the ground truth provided by the human evaluators on blurred
images. On the other hand, Algo 3 shows a significant
deviation from the ground truth on blurred images. On the
other hand, in classifying images as sharp, Algo 1 and Algo 2
deviate from the ground truth by approximately 20 points
while Algo 2 deviates from the ground truth by 45 points.
Tables 3 and 4 indicate that image blur detection has a
pronounced positive effect on skewed barcode scanning. In all
three random samples, the barcoding recognition gain was
above thirty percent. While we ran these experiments only
with our barcode scanning algorithm, we expect similar gains
with other vision-based barcode scanning algorithms.
Our experiments indicate that direct methods provide a
viable alternative to indirect methods. While indirect methods
may be more accurate, they tend to be more computationally
expensive due to complex matrix manipulations. Direct
methods may not be as precise as their indirect counterparts.
However, they compensate for it by increased efficiency,
which makes them more suitable for mobile and wearable
platforms.
Another observation that we would like to make is that in
working with our samples of images we could not observe the
blur effect on the edges observed by Tong et al. [7] in some
images. The edge blur effect observed by these researchers is
that the injection of blur in the images causes the Dirac and A-
Step edges disappear or turn into Roof and G-Step edges,
respectively and the G-Step and Roof edges to lose their
sharpness.
Instead of using the 2D HWT to detect edge types, the
algorithm proposed in this paper is based on the assumption
that in blurred images square tiles with pronounced changes
do not form larger clusters but scatter across the image as
singletons or form clusters that are small in size relative to the
overall size of the image.
Our approach is rooted in the research by Mallat and
Hwang [9] who show that the 2D HWT can detect the location
of irregularities in 2D images. In our algorithm, the 2D HWT
is used to detect the location of changes via square tiles
without explicitly identifying the causes of the detected
changes, e.g., edges or corners.
VI. Summary
We have presented an algorithm for direct image blur
detection with the 2D Haar Wavelet transform (2D HWT).
The algorithm classifies an image as blurred or sharp by
splitting it into N x N tiles, applying four iterations of the 2D
HWT to each tile, and grouping the horizontally, vertically,
and diagonally connected tiles with pronounced changes into
tile clusters. Images with large tile clusters are classified as
sharp. Images with small tile clusters are classified as blurred.
If necessary, the blur extent can be estimated as the ratio of the
total area of the large tile clusters and the area of the whole
image.
7. IPCV 2015
Our experiments on a sample of 500 images indicate that
our algorithm either performs on par or outperforms two other
blur detection algorithms found in the literature. The
experiments also indicate that image blur detection has a
pronounced positive effect on skewed barcode scanning. One
possible implication of the research presented in this paper is
that it may be possible to estimate blurriness in the images
without explicitly computing the explicit features in the image
that caused the blurriness (e.g., edges) or using involved
methods to find the best fitting blur function.
In our future work, we plan to investigate the effect of
image blur detection on vision-based nutrition label scanning
to improve the optical character recognition (OCR) rates. In
our previous work, we proposed to a greedy spellchecking
algorithm to correct OCR errors during nutrition label
scanning on smartphones [21]. The proposed algorithm, called
skip trie matching, uses a dictionary of strings stored in the
trie data structure to correct run-time OCR errors by skipping
misrecognized characters while going down specific trie paths.
We expect that eliminating blurred frames from the
processing stream, if done reliably, will improve the OCR
rates and will make it possible to use open source OCR
engines such as Tesseract (http://code.google.com/p/tesseract-
ocr) and GOCR (http://jocr.sourceforge.net) in vision-based
nutrition label scanning.
Acknowledgment
We would like to thank Tanwir Zaman and Sai Rekka for
volunteering their time to classify five hundred images as
blurred or sharp.
References
[1] Kulyukin, V. and Zaman, T. “Vision-based localization
and scanning of 1D UPC and EAN barcodes with relaxed
pitch, roll, and yaw camera alignment constraints.”
International Journal of Image Processing (IJIP), vol. 8,
issue 5, 2014, pp. 355-383.
[2] Kulyukin, V. and Zaman, T. “An Algorithm for in-place
vision-based skewed 1D barcode scanning in the cloud.”
In Proceedings of the 18th International Conference on
Image Processing and Pattern Recognition (IPCV 2014),
pp. 36-42, July 21-24, Las Vegas, NV, USA, CSREA
Press, ISBN: 1-60132-280-1.
[3] Kulyukin, V. and Blay, C. “An Algoritm for mobile
vision-based localization of skewed nutrition labels that
maximizes specificity.” In Proceedings of the 18th
International Conference on Image Processing and
Pattern Recognition (IPCV 2014), pp. 3-9, July 21-24,
2014, Las Vegas, NV, USA, CSREA Press, ISBN: 1-
60132-280-1.
[4] Canny, J.F. “A Computational approach to edge
detection.” IEEE Transactions on Pattern Analysis and
Machine Intelligence, vol. 8, 1986, pp. 679-698.
[5] Laganiere, R. OpenCV 2 Computer Vision Application
Programming Cookbook. Packt Publishing Ltd, 2011.
[6] Nievergelt, Y. Wavelets Made Easy. Birkäuser, Boston,
2000, ISBN-10: 0817640614.
[7] Tong, H., Li, M., Zhang, H., and Zhang, C. "Blur
detection for digital images using wavelet transform," In
Proceedings of the IEEE International Conference on
Multimedia and Expo, vol.1, pp. 27-30, June 2004.
doi: 10.1109/ICME.2004.1394114.
[8] Cretea,F., Dolmierea, T., Ladreta, P., Nicolas, M. “The
Blur effect: perception and estimation with a new no-
reference perceptual blur metric.” In Proceedings of
SPIE 6492, Human Vision and Electronic Imaging XII,
64920I, San Jose, CA, USA, January 28, 2007.
doi:10.1117/12.702790.
[9] Mallat, S. and Hwang, W. L. “Singularity detection and
processing with wavelets.” IEEE Transactions on
Information Theory, vol. 38, no. 2, March 1992, pp. 617-
643.
[10] Smith, J.O. Mathematics of the Discrete Fourier
Transform with Audio Applications, 2nd
Edition, W3K
Publishing, 2007, ISBN 978-0-9745607-4-8.
[11] Rooms, F., Pizurica, A., Philips, W. “Estimating image
blur in the wavelet domain.” In Proc. of IEEE Int. Conf.
on Acoustics and Signal Processing, vol. 4, pp.4190-
4195, IEEE, 2002.
[12] Wanqing, S., Qing, L., Yuming, W. “Tool wear detection
using Lipschitz exponent and harmonic wavelet.”
Mathematical Problems in Engineering, August 2013,
Article ID 489261,
http://dx.doi.org/10.1155/2013/489261.
[13] Venkatakrishnan, P., Sangeetha, S., and Sundar, M.
“Measurement of Lipschitz exponent using wavelet
transform modulus maxima.” International Journal of
Scientific & Engineering Research, vol. 3, issue 6, pp. 1-
4, June-2012, ISSN 2229-5518.
[14] Pavlovic, G., and Tekalp, M. “Maximum likelihood
parametric blur identification based on a continuous
spatial domain model.” IEEE Trans. on Image
Processing, vol. 1, issue 4, Oct. 1992, pp. 496-504.
[15] Panchapakesan, K., Sheppard, D.G., Marcellin, M.W.,
and Hunt, B.R. “Blur identification from vector quantizer
encoder distortion.” In Proc. of the 1998 International
Conference on Image Processing (ICIP 98), pp. 751-755,
4-7 Oct. 1998, Chicago, IL., USA.
[16] Marichal, X., Ma, W., and Zhang, H.J. “Blur
determination in the compressed domain using DCT
information.” in Proc. IEEE Int. Conf. Image Processing,
Oct. 1999, vol. 2, pp. 386–390.
[17] Java implementation of the 2DHWT procedure.
https://github.com/VKEDCO/java/tree/master/haar.
[18] Mobile supermarket barcode videos of grocery packages.
https://www.dropbox.com/sh/q6u70wcg1luxwdh/LPtUBd
wdY1.
[19] Python implementation of the blur detection algorithm
proposed in reference [7].
https://github.com/VKEDCO/PYPL/blob/master/haar_blu
r.
[20] MATLAB implementation of blur detection algorithm
proposed in reference [8].
http://www.mathworks.com/matlabcentral/fileexchange/2
4676-image-blur-metric.
[21] Kulyukin, V., Vanka, A., Wang, W. “Skip trie matching:
a greedy algorithm for real-time OCR error correction on
smartphones.” International Journal of Digital
Information and Wireless Communication (IJDIWC): vol.
3, issue 3, pp. 56-65, 2013. ISSN: 2225-658X.