The thirst of better and faster retrieval techniques has always fuelled to the research in content based image retrieval (CBIR). The paper presents innovative content based image retrieval (CBIR) techniques based on feature vectors as fractional coefficients of transformed images using Discrete Cosine, Walsh, Haar and Kekre’s transforms. Here the advantage of energy compaction of transforms in higher coefficients is taken to greatly reduce the feature vector size per image by taking fractional coefficients of transformed image. The feature vectors are extracted in fourteen different ways from the transformed image, with the first being considering all the coefficients of transformed image and then fourteen reduced coefficients sets (as 50%, 25%, 12.5%, 6.25%, 3.125%, 1.5625% ,0.7813%, 0.39%, 0.195%, 0.097%, 0.048%, 0.024%, 0.012% and 0.06% of complete transformed image) are considered as feature vectors. The four transforms are applied on gray image equivalents and the colour components of images to extract Gray and RGB feature sets respectively. Instead of using all coefficients of transformed images as feature vector for image retrieval, these fourteen reduced coefficients sets for gray as well as RGB feature vectors are used, resulting into better performance and lower computations. The proposed CBIR techniques are implemented on a database having 1000 images spread across 11 categories. For each proposed CBIR technique 55 queries (5 per category) are fired on the database and net average precision and recall are computed for all feature sets per transform. The results have shown performance improvement (higher precision and recall values) with fractional coefficients compared to complete transform of image at reduced computations resulting in faster retrieval. Finally Kekre’s transform surpasses all other discussed transforms in performance with highest precision and recall values for fractional coefficients (6.25% and 3.125% of all coefficients) and computation are lowered by 94.08% as compared to DCT.
This document summarizes a research paper that proposes a content-based image retrieval system using cascaded color and texture features. Color features are first extracted from images using statistical measures like mean, standard deviation, energy, entropy, skewness and kurtosis. Similarity to a query image is then measured using distance metrics. The top 150 most similar images are then analyzed to extract Haralick texture features. Similarity is again measured to retrieve the most relevant images. The paper finds that Canberra distance provides better retrieval results than other distance metrics like City Block and Minkowski.
Segmentation by Fusion of Self-Adaptive SFCM Cluster in Multi-Color Space Com...CSCJournals
This paper proposes a new, simple, and efficient segmentation approach that could find diverse applications in pattern recognition as well as in computer vision, particularly in color image segmentation. First, we choose the best segmentation components among six different color spaces. Then, Histogram and SFCM techniques are applied for initialization of segmentation. Finally, we fuse the segmentation results and merge similar regions. Extensive experiments have been taken on Berkeley image database by using the proposed algorithm. The results show that, compared with some classical segmentation algorithms, such as Mean-Shift, FCR and CTM, etc, our method could yield reasonably good or better image partitioning, which illustrates practical value of the method.
Extended Performance Appraise of Image Retrieval Using the Feature Vector as ...Waqas Tariq
This document summarizes the results of testing an image retrieval technique that uses row means of transformed image columns as a feature vector on a database of 1000 images across 11 categories. Seven different transforms were tested, including DCT, DST, Haar, Walsh, Kekre, Slant, and Hartley. For each transform, precision and recall were calculated with and without including the DC component in the feature vector. The technique was tested on both gray and color versions of the database. In general, the technique produced higher precision and recall than using the full transform or simple row means, especially when including the DC component. For gray images, the best performing transforms were DST, Haar, Hartley, DCT,
FAN search for image copy-move forgery-amalta 2014SondosFadl
1) The document proposes a fast fan search method for detecting copy-move image forgery. It divides images into blocks, extracts features from blocks, and uses a fan search algorithm to detect duplicated blocks more efficiently than previous methods.
2) Experimental results show the proposed method can detect copy-move forgery 75% faster than other methods, with 99% precision and 98% recall.
3) Future work will improve the method to detect duplications under geometric transformations like rotation and scaling.
Copy-Rotate-Move Forgery Detection Based on Spatial DomainSondosFadl
we propose a method which is efficient and fast for detecting Copy-Move regions even when the copied region was undergone rotation modify in spatial domain.
Andrey Kuznetsov and Vladislav Myasnikov - Using Efficient Linear Local Feat...AIST
The document proposes a new copy-move forgery detection algorithm using efficient linear local features. The key features of the algorithm are 100% recall, high calculation speed for real-time analysis, and 99.9% precision. It works by using a sliding window to analyze image fragments defined by a structural pattern. Hash values are calculated for each fragment based on linear local features and stored in a hash table. Duplicates are identified by finding hash values with a frequency greater than one. Experiments on satellite images show the algorithm has no false negatives and very low false positives, while being much faster than other methods.
FORGERY (COPY-MOVE) DETECTION IN DIGITAL IMAGES USING BLOCK METHODeditorijcres
AKHILESH KUMAR YADAV, DEENBANDHU SINGH, VIVEK KUMAR
Department of Computer Science and Engineering
Babu Banarasi Das University, Lucknow
akhi2232232@gmail.com, deenbandhusingh85@gmail.com, vivek.kumar0091@gmail.com
ABSTRACT- Digital images can be easily modified using powerful image editing software. Determining whether a manipulation is innocent of sharpening from those which are malicious, such as removing or adding parts to an image is the topic of this paper. In this paper we focus on detection of a special type of forgery-the Copy-Move forgery, in this part of the original image is copied moved to desired location in the same image and pasted. The proposed method compress images using DWT (discrete wavelet transform) and divided into blocks and choose blocks than perform feature vector calculation and lexicographical sorting and duplicated blocks are identified after sorting. This method is good at some manipulation/attack likes scaling, rotation, Gaussian noise, smoothing, JPEG compression etc.
INDEX TERMS- Copy-Move forgery, Wavelet Transform, Lexicographical Sorting, Region Duplication Detection.
This document summarizes a research paper that proposes a content-based image retrieval system using cascaded color and texture features. Color features are first extracted from images using statistical measures like mean, standard deviation, energy, entropy, skewness and kurtosis. Similarity to a query image is then measured using distance metrics. The top 150 most similar images are then analyzed to extract Haralick texture features. Similarity is again measured to retrieve the most relevant images. The paper finds that Canberra distance provides better retrieval results than other distance metrics like City Block and Minkowski.
Segmentation by Fusion of Self-Adaptive SFCM Cluster in Multi-Color Space Com...CSCJournals
This paper proposes a new, simple, and efficient segmentation approach that could find diverse applications in pattern recognition as well as in computer vision, particularly in color image segmentation. First, we choose the best segmentation components among six different color spaces. Then, Histogram and SFCM techniques are applied for initialization of segmentation. Finally, we fuse the segmentation results and merge similar regions. Extensive experiments have been taken on Berkeley image database by using the proposed algorithm. The results show that, compared with some classical segmentation algorithms, such as Mean-Shift, FCR and CTM, etc, our method could yield reasonably good or better image partitioning, which illustrates practical value of the method.
Extended Performance Appraise of Image Retrieval Using the Feature Vector as ...Waqas Tariq
This document summarizes the results of testing an image retrieval technique that uses row means of transformed image columns as a feature vector on a database of 1000 images across 11 categories. Seven different transforms were tested, including DCT, DST, Haar, Walsh, Kekre, Slant, and Hartley. For each transform, precision and recall were calculated with and without including the DC component in the feature vector. The technique was tested on both gray and color versions of the database. In general, the technique produced higher precision and recall than using the full transform or simple row means, especially when including the DC component. For gray images, the best performing transforms were DST, Haar, Hartley, DCT,
FAN search for image copy-move forgery-amalta 2014SondosFadl
1) The document proposes a fast fan search method for detecting copy-move image forgery. It divides images into blocks, extracts features from blocks, and uses a fan search algorithm to detect duplicated blocks more efficiently than previous methods.
2) Experimental results show the proposed method can detect copy-move forgery 75% faster than other methods, with 99% precision and 98% recall.
3) Future work will improve the method to detect duplications under geometric transformations like rotation and scaling.
Copy-Rotate-Move Forgery Detection Based on Spatial DomainSondosFadl
we propose a method which is efficient and fast for detecting Copy-Move regions even when the copied region was undergone rotation modify in spatial domain.
Andrey Kuznetsov and Vladislav Myasnikov - Using Efficient Linear Local Feat...AIST
The document proposes a new copy-move forgery detection algorithm using efficient linear local features. The key features of the algorithm are 100% recall, high calculation speed for real-time analysis, and 99.9% precision. It works by using a sliding window to analyze image fragments defined by a structural pattern. Hash values are calculated for each fragment based on linear local features and stored in a hash table. Duplicates are identified by finding hash values with a frequency greater than one. Experiments on satellite images show the algorithm has no false negatives and very low false positives, while being much faster than other methods.
FORGERY (COPY-MOVE) DETECTION IN DIGITAL IMAGES USING BLOCK METHODeditorijcres
AKHILESH KUMAR YADAV, DEENBANDHU SINGH, VIVEK KUMAR
Department of Computer Science and Engineering
Babu Banarasi Das University, Lucknow
akhi2232232@gmail.com, deenbandhusingh85@gmail.com, vivek.kumar0091@gmail.com
ABSTRACT- Digital images can be easily modified using powerful image editing software. Determining whether a manipulation is innocent of sharpening from those which are malicious, such as removing or adding parts to an image is the topic of this paper. In this paper we focus on detection of a special type of forgery-the Copy-Move forgery, in this part of the original image is copied moved to desired location in the same image and pasted. The proposed method compress images using DWT (discrete wavelet transform) and divided into blocks and choose blocks than perform feature vector calculation and lexicographical sorting and duplicated blocks are identified after sorting. This method is good at some manipulation/attack likes scaling, rotation, Gaussian noise, smoothing, JPEG compression etc.
INDEX TERMS- Copy-Move forgery, Wavelet Transform, Lexicographical Sorting, Region Duplication Detection.
Tracking and counting human in visual surveillance systemiaemedu
This document summarizes a proposed system for tracking and counting humans in a visual surveillance system. The system first performs background subtraction on input video frames to detect foreground objects. It applies this process to both grayscale and binary image formats, and selects the better-performing format. Features are then extracted from detected blobs, including size, centroid, and color. Tracking rectangles are drawn around whole detected objects to group separated body parts. Counting is done by detecting when pixel values change from background to foreground. Experimental results on videos with various challenges like shadows, illumination changes, and occlusions showed the system could accurately track and count humans, with near-perfect accuracy even for occluded or broken objects, though it introduced some minor errors.
The compression is a process of Image Processing which interested to change the information representation in order to reduce the stockage capacity and transmission time. In this work we propose a new image compression algorithm based on Haar wavelets by introducing a compression coefficient that controls the compression levels. This method reduces the complexity in obtaining the desired level of compression from the original image only and without using intermediate levels.
Object Shape Representation by Kernel Density Feature Points Estimator cscpconf
This paper introduces an object shape representation using Kernel Density Feature Points
Estimator (KDFPE). In this method we obtain the density of feature points within defined rings
around the centroid of the image. The Kernel Density Feature Points Estimator is then applied to
the vector of the image. KDFPE is invariant to translation, scale and rotation. This method of
image representation shows improved retrieval rate when compared to Density Histogram
Feature Points (DHFP) method. Analytic analysis is done to justify our method and we compared our results with object shape representation by the Density Histogram of Feature Points (DHFP) to prove its robustness.
Effect of Block Sizes on the Attributes of Watermarking Digital ImagesDr. Michael Agbaje
This work examines the effect of block sizes on attributes (robustness, capacity, time of watermarking, visibility and distortion) of watermarked digital images using Discrete Cosine Transform (DCT) function. The DCT function breaks up the image into various frequency bands and allows watermark data to be easily embedded. The advantage of this transformation is the ability to pack input image data into a few coefficients. The block size 8 x 8 is commonly used in watermarking. The work investigates the effect of using block sizes below and above 8 x 8 on the attributes of watermark. The attributes of robustness and capacity increase as the block size increases (62-70db, 31.5-35.9 bit/pixel). The time for watermarking reduces as the block size increases. The watermark is still visible for block sizes below 8 x 8 but invisible for those above it. Distortion decreases sharply from a high value at 2 x 2 block size to minimum at 8 x 8 and gradually increases with block size. The overall observation indicates that watermarked image gradually reduces in quality due to fading above 8 x 8 block size. For easy detection of image against piracy the block size 16 x 16 gives the best output result because it closely resembles the original image in terms of visual quality displayed despite the fact that it contains a hidden watermark.
This document provides a survey of single scalar point multiplication algorithms for elliptic curves over prime fields. It discusses the background of elliptic curve cryptography and point multiplication. Point multiplication is the dominant operation in ECC and can be computed using on-the-fly techniques or precomputation if the point is fixed. The efficiency of point multiplication depends on the recoding method used to represent the scalar and the composite elliptic curve operations employed. Various recoding methods and point multiplication algorithms are analyzed, including binary, signed binary using NAF representation, and window methods.
This document reviews different techniques for thinning images, including the Zhang and Suen algorithm and neural networks. It provides an overview of existing thinning approaches, such as iterative algorithms, and proposes a new approach using neural networks. The proposed approach aims to perform thinning invariant to rotations while being less sensitive to noise than existing methods. It evaluates techniques based on execution time, thinning rate, and other performance measures. The document concludes that neural networks may provide better results than existing techniques in terms of metrics like PSNR and MSE, while also reducing execution time for skeletonization.
This document provides an overview of digital image fundamentals and operations. It defines what a digital image is, how it is represented as a matrix, and common image types like RGB, grayscale, and binary. Pixels, resolution, neighborhoods, and basic relationships between pixels are discussed. The document also covers different types of image operations including point, local, and global operations as well as examples like arithmetic, logical, and geometric transformations. Finally, it introduces concepts of linear and nonlinear operations and announces the topic of the next lecture on image enhancement in the spatial domain.
An Approach for Image Deblurring: Based on Sparse Representation and Regulari...IRJET Journal
This document proposes an approach for image deblurring based on sparse representation and a regularized filter. The approach splits the blurred input image into patches, estimates sparse coefficients for each patch using dictionary learning, updates the dictionary, and estimates the deblur kernel. The deblur kernel is applied using Wiener deconvolution and further processed with a regularized filter to recover the original image. The approach was tested on MATLAB and evaluation metrics like RMSE, PSNR, and SSIM along with visual analysis showed it performed better deblurring compared to existing methods.
Abstract: Many applications such as robot navigation, defense, medical and remote sensing performvarious processing tasks, which can be performed more easily when all objects in different images of the same scene are combined into a single fused image. In this paper, we propose a fast and effective method for image fusion. The proposed method derives the intensity based variations that is large and small scale, from the source images. In this approach, guided filtering is employed for this extraction. Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained. Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
This document presents a new color image segmentation approach based on overlap wavelet transform (OWT). OWT extracts wavelet features to better separate different patterns in an image. The proposed method also uses morphological operators and 2D histogram clustering for effective segmentation. It is concluded that the proposed OWT method improves segmentation quality, is reliable, fast and computationally less complex than direct histogram clustering. When tested on various color spaces, the proposed segmentation scheme produced better results in RGB color space compared to others. The main advantages are its use of a single parameter and faster speed.
This document describes an implementation of artistic style learning using a neural network approach. The algorithm is based on the VGG convolutional neural network and uses the NVidia CUDA platform to speed up computations. The implementation reconstructs content using loss functions on convolutional layer outputs and represents style using Gram matrices of filter correlations. A new image is generated by minimizing a combined loss of content and style. Experiments show improved results over the original algorithm by using softplus activations and average pooling.
Fast Segmentation of Sub-cellular OrganellesCSCJournals
Segmentation and counting sub-cellular structure is a very challenging problem even for medical experts. A fast and efficient method for segmentation and counting of sub-cellular structure is proposed. The proposed method uses a hybrid combination of several image processing techniques and is effective in segmenting the sub-cellular structures in a fast and effective manner.
PERFORMANCE EVALUATION OF DIFFERENT TECHNIQUES FOR TEXTURE CLASSIFICATION cscpconf
This document compares different techniques for texture classification, including wavelet transforms and co-occurrence matrices. It finds that the Haar wavelet technique is the most efficient in terms of time complexity and classification accuracy, except when images are rotated. The co-occurrence matrix method has higher time requirements but excellent classification results, except for rotated images where accuracy is greatly reduced due to its dependence on pixel values. Overall, the Haar wavelet proves to be the best method for texture classification based on the performance assessment parameters of time complexity and classification accuracy.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Background Estimation Using Principal Component Analysis Based on Limited Mem...IJECEIAES
Given a video of 푀 frames of size ℎ × 푤. Background components of a video are the elements matrix which relative constant over 푀 frames. In PCA (principal component analysis) method these elements are referred as “principal components”. In video processing, background subtraction means excision of background component from the video. PCA method is used to get the background component. This method transforms 3 dimensions video (ℎ × 푤 × 푀) into 2 dimensions one (푁 × 푀), where 푁 is a linear array of size ℎ × 푤 . The principal components are the dominant eigenvectors which are the basis of an eigenspace. The limited memory block Krylov subspace optimization then is proposed to improve performance the computation. Background estimation is obtained as the projection each input image (the first frame at each sequence image) onto space expanded principal component. The procedure was run for the standard dataset namely SBI (Scene Background Initialization) dataset consisting of 8 videos with interval resolution [146 150, 352 240], total frame [258,500]. The performances are shown with 8 metrics, especially (in average for 8 videos) percentage of error pixels (0.24%), the percentage of clustered error pixels (0.21%), multiscale structural similarity index (0.88 form maximum 1), and running time (61.68 seconds).
Change Detection of Water-Body in Synthetic Aperture Radar ImagesCSCJournals
Change detection is the art of quantifying the changes in the Synthetic Aperture Radar (SAR) images that have happened over a period of time. Remote sensing has been the parental technique to perform change detection analysis. This paper empirically investigates the impact of applying the combination of texture features for different classification techniques to separate water body from non-water body. At first, the images are classified using unsupervised Principle Component Analysis (PCA) based K-means clustering for dimension reduction. Then the texture features like Energy, Entropy, Contrast , Inverse Differential Moment , Directional Moment and the Median are extracted using Gray Level Co-occurrence Matrix (GLCM) and these features are utilized in Linear Vector Quantization (LVQ) and Support Vector Machine (SVM) classifiers. This paper aims to apply a combination of the texture features in order to significantly improve the accuracy of detection. The utility of detection analysis, influences management and policy decision making for long-term construction projects by predicting the preventable losses.
Teresa Clotilde Ojeda Sánchez: Sesión de Aprendizaje 05 de Unidad Didáctica 03 del Área de Personal Social – 1er. Grado de Primaria 2015: Mi nombre me da identidad
Este documento presenta una lección sobre cómo manejar conflictos de manera constructiva. Expone un caso de conflicto entre niños durante el recreo y ofrece materiales para que los estudiantes analicen la situación, comprendan la diferencia entre conflicto y violencia, y desarrollen estrategias para resolver conflictos futuros de una manera que mejore la convivencia.
El documento describe una sesión para que un grupo de estudiantes de sexto grado elaboren normas de convivencia democrática. La sesión incluye actividades como analizar situaciones que dificultan la convivencia, proponer soluciones, elegir normas y plasmarlas en carteles para colocar en el aula. El objetivo es que los estudiantes participen en establecer acuerdos que promuevan buenas relaciones y condiciones para el aprendizaje.
Las normas de convivencia no son un fin en sí mismas sino un medio para regular las relaciones entre personas y respetar los derechos de todos. El documento describe cómo los estudiantes participaron en la elaboración de normas de convivencia democráticas para el aula a través de identificar situaciones problemáticas comunes y establecer acuerdos para mejorar la convivencia.
Teresa Clotilde Ojeda Sánchez: el Ministerio de Educación del Perú (MINEDU) pone a disposición del personal docente el documento: Sesión de Aprendizaje 02 de Unidad Didáctica 01 del Área de Personal Social – Cuarto grado de Primaria 2015: "Planificamos para organizarnos y aprender mejor”
Tracking and counting human in visual surveillance systemiaemedu
This document summarizes a proposed system for tracking and counting humans in a visual surveillance system. The system first performs background subtraction on input video frames to detect foreground objects. It applies this process to both grayscale and binary image formats, and selects the better-performing format. Features are then extracted from detected blobs, including size, centroid, and color. Tracking rectangles are drawn around whole detected objects to group separated body parts. Counting is done by detecting when pixel values change from background to foreground. Experimental results on videos with various challenges like shadows, illumination changes, and occlusions showed the system could accurately track and count humans, with near-perfect accuracy even for occluded or broken objects, though it introduced some minor errors.
The compression is a process of Image Processing which interested to change the information representation in order to reduce the stockage capacity and transmission time. In this work we propose a new image compression algorithm based on Haar wavelets by introducing a compression coefficient that controls the compression levels. This method reduces the complexity in obtaining the desired level of compression from the original image only and without using intermediate levels.
Object Shape Representation by Kernel Density Feature Points Estimator cscpconf
This paper introduces an object shape representation using Kernel Density Feature Points
Estimator (KDFPE). In this method we obtain the density of feature points within defined rings
around the centroid of the image. The Kernel Density Feature Points Estimator is then applied to
the vector of the image. KDFPE is invariant to translation, scale and rotation. This method of
image representation shows improved retrieval rate when compared to Density Histogram
Feature Points (DHFP) method. Analytic analysis is done to justify our method and we compared our results with object shape representation by the Density Histogram of Feature Points (DHFP) to prove its robustness.
Effect of Block Sizes on the Attributes of Watermarking Digital ImagesDr. Michael Agbaje
This work examines the effect of block sizes on attributes (robustness, capacity, time of watermarking, visibility and distortion) of watermarked digital images using Discrete Cosine Transform (DCT) function. The DCT function breaks up the image into various frequency bands and allows watermark data to be easily embedded. The advantage of this transformation is the ability to pack input image data into a few coefficients. The block size 8 x 8 is commonly used in watermarking. The work investigates the effect of using block sizes below and above 8 x 8 on the attributes of watermark. The attributes of robustness and capacity increase as the block size increases (62-70db, 31.5-35.9 bit/pixel). The time for watermarking reduces as the block size increases. The watermark is still visible for block sizes below 8 x 8 but invisible for those above it. Distortion decreases sharply from a high value at 2 x 2 block size to minimum at 8 x 8 and gradually increases with block size. The overall observation indicates that watermarked image gradually reduces in quality due to fading above 8 x 8 block size. For easy detection of image against piracy the block size 16 x 16 gives the best output result because it closely resembles the original image in terms of visual quality displayed despite the fact that it contains a hidden watermark.
This document provides a survey of single scalar point multiplication algorithms for elliptic curves over prime fields. It discusses the background of elliptic curve cryptography and point multiplication. Point multiplication is the dominant operation in ECC and can be computed using on-the-fly techniques or precomputation if the point is fixed. The efficiency of point multiplication depends on the recoding method used to represent the scalar and the composite elliptic curve operations employed. Various recoding methods and point multiplication algorithms are analyzed, including binary, signed binary using NAF representation, and window methods.
This document reviews different techniques for thinning images, including the Zhang and Suen algorithm and neural networks. It provides an overview of existing thinning approaches, such as iterative algorithms, and proposes a new approach using neural networks. The proposed approach aims to perform thinning invariant to rotations while being less sensitive to noise than existing methods. It evaluates techniques based on execution time, thinning rate, and other performance measures. The document concludes that neural networks may provide better results than existing techniques in terms of metrics like PSNR and MSE, while also reducing execution time for skeletonization.
This document provides an overview of digital image fundamentals and operations. It defines what a digital image is, how it is represented as a matrix, and common image types like RGB, grayscale, and binary. Pixels, resolution, neighborhoods, and basic relationships between pixels are discussed. The document also covers different types of image operations including point, local, and global operations as well as examples like arithmetic, logical, and geometric transformations. Finally, it introduces concepts of linear and nonlinear operations and announces the topic of the next lecture on image enhancement in the spatial domain.
An Approach for Image Deblurring: Based on Sparse Representation and Regulari...IRJET Journal
This document proposes an approach for image deblurring based on sparse representation and a regularized filter. The approach splits the blurred input image into patches, estimates sparse coefficients for each patch using dictionary learning, updates the dictionary, and estimates the deblur kernel. The deblur kernel is applied using Wiener deconvolution and further processed with a regularized filter to recover the original image. The approach was tested on MATLAB and evaluation metrics like RMSE, PSNR, and SSIM along with visual analysis showed it performed better deblurring compared to existing methods.
Abstract: Many applications such as robot navigation, defense, medical and remote sensing performvarious processing tasks, which can be performed more easily when all objects in different images of the same scene are combined into a single fused image. In this paper, we propose a fast and effective method for image fusion. The proposed method derives the intensity based variations that is large and small scale, from the source images. In this approach, guided filtering is employed for this extraction. Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained. Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
This document presents a new color image segmentation approach based on overlap wavelet transform (OWT). OWT extracts wavelet features to better separate different patterns in an image. The proposed method also uses morphological operators and 2D histogram clustering for effective segmentation. It is concluded that the proposed OWT method improves segmentation quality, is reliable, fast and computationally less complex than direct histogram clustering. When tested on various color spaces, the proposed segmentation scheme produced better results in RGB color space compared to others. The main advantages are its use of a single parameter and faster speed.
This document describes an implementation of artistic style learning using a neural network approach. The algorithm is based on the VGG convolutional neural network and uses the NVidia CUDA platform to speed up computations. The implementation reconstructs content using loss functions on convolutional layer outputs and represents style using Gram matrices of filter correlations. A new image is generated by minimizing a combined loss of content and style. Experiments show improved results over the original algorithm by using softplus activations and average pooling.
Fast Segmentation of Sub-cellular OrganellesCSCJournals
Segmentation and counting sub-cellular structure is a very challenging problem even for medical experts. A fast and efficient method for segmentation and counting of sub-cellular structure is proposed. The proposed method uses a hybrid combination of several image processing techniques and is effective in segmenting the sub-cellular structures in a fast and effective manner.
PERFORMANCE EVALUATION OF DIFFERENT TECHNIQUES FOR TEXTURE CLASSIFICATION cscpconf
This document compares different techniques for texture classification, including wavelet transforms and co-occurrence matrices. It finds that the Haar wavelet technique is the most efficient in terms of time complexity and classification accuracy, except when images are rotated. The co-occurrence matrix method has higher time requirements but excellent classification results, except for rotated images where accuracy is greatly reduced due to its dependence on pixel values. Overall, the Haar wavelet proves to be the best method for texture classification based on the performance assessment parameters of time complexity and classification accuracy.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Background Estimation Using Principal Component Analysis Based on Limited Mem...IJECEIAES
Given a video of 푀 frames of size ℎ × 푤. Background components of a video are the elements matrix which relative constant over 푀 frames. In PCA (principal component analysis) method these elements are referred as “principal components”. In video processing, background subtraction means excision of background component from the video. PCA method is used to get the background component. This method transforms 3 dimensions video (ℎ × 푤 × 푀) into 2 dimensions one (푁 × 푀), where 푁 is a linear array of size ℎ × 푤 . The principal components are the dominant eigenvectors which are the basis of an eigenspace. The limited memory block Krylov subspace optimization then is proposed to improve performance the computation. Background estimation is obtained as the projection each input image (the first frame at each sequence image) onto space expanded principal component. The procedure was run for the standard dataset namely SBI (Scene Background Initialization) dataset consisting of 8 videos with interval resolution [146 150, 352 240], total frame [258,500]. The performances are shown with 8 metrics, especially (in average for 8 videos) percentage of error pixels (0.24%), the percentage of clustered error pixels (0.21%), multiscale structural similarity index (0.88 form maximum 1), and running time (61.68 seconds).
Change Detection of Water-Body in Synthetic Aperture Radar ImagesCSCJournals
Change detection is the art of quantifying the changes in the Synthetic Aperture Radar (SAR) images that have happened over a period of time. Remote sensing has been the parental technique to perform change detection analysis. This paper empirically investigates the impact of applying the combination of texture features for different classification techniques to separate water body from non-water body. At first, the images are classified using unsupervised Principle Component Analysis (PCA) based K-means clustering for dimension reduction. Then the texture features like Energy, Entropy, Contrast , Inverse Differential Moment , Directional Moment and the Median are extracted using Gray Level Co-occurrence Matrix (GLCM) and these features are utilized in Linear Vector Quantization (LVQ) and Support Vector Machine (SVM) classifiers. This paper aims to apply a combination of the texture features in order to significantly improve the accuracy of detection. The utility of detection analysis, influences management and policy decision making for long-term construction projects by predicting the preventable losses.
Teresa Clotilde Ojeda Sánchez: Sesión de Aprendizaje 05 de Unidad Didáctica 03 del Área de Personal Social – 1er. Grado de Primaria 2015: Mi nombre me da identidad
Este documento presenta una lección sobre cómo manejar conflictos de manera constructiva. Expone un caso de conflicto entre niños durante el recreo y ofrece materiales para que los estudiantes analicen la situación, comprendan la diferencia entre conflicto y violencia, y desarrollen estrategias para resolver conflictos futuros de una manera que mejore la convivencia.
El documento describe una sesión para que un grupo de estudiantes de sexto grado elaboren normas de convivencia democrática. La sesión incluye actividades como analizar situaciones que dificultan la convivencia, proponer soluciones, elegir normas y plasmarlas en carteles para colocar en el aula. El objetivo es que los estudiantes participen en establecer acuerdos que promuevan buenas relaciones y condiciones para el aprendizaje.
Las normas de convivencia no son un fin en sí mismas sino un medio para regular las relaciones entre personas y respetar los derechos de todos. El documento describe cómo los estudiantes participaron en la elaboración de normas de convivencia democráticas para el aula a través de identificar situaciones problemáticas comunes y establecer acuerdos para mejorar la convivencia.
Teresa Clotilde Ojeda Sánchez: el Ministerio de Educación del Perú (MINEDU) pone a disposición del personal docente el documento: Sesión de Aprendizaje 02 de Unidad Didáctica 01 del Área de Personal Social – Cuarto grado de Primaria 2015: "Planificamos para organizarnos y aprender mejor”
Este documento presenta una sesión sobre el derecho a un nombre. La sesión comienza con una discusión sobre la importancia de los nombres y el Documento Nacional de Identidad. Luego, los estudiantes trabajan en parejas para completar los datos personales de dos niños ficticios en formularios de DNI. Finalmente, reflexionan sobre la relevancia de tener un nombre y un DNI registrado.
Este documento presenta una sesión de aprendizaje sobre metas personales para mejorar la convivencia. Los estudiantes reflexionan sobre sus propias dificultades de convivencia y establecen compromisos individuales para cumplir metas que los ayuden a relacionarse mejor con los demás en el aula. El profesor guía la discusión y ayuda a los estudiantes a definir metas alcanzables que luego monitorearán en las próximas semanas.
Teresa Clotilde Ojeda Sánchez: Sesión de Aprendizaje 03 de Unidad Didáctica 02 del Área de Personal Social - Tercer grado de Primaria 2015: "Nos conocemos y nos aceptamos como somos"
En esta sesión, los niños y las niñas aprenderán a identificar lo bueno y positivo de cada uno de ellos, para así tener mayor confianza en sí mismos y aceptarse como son; en la medida en que se sepan reconocer sus fortalezas, podrán afirmar su identidad.
Los niños utilizarán números cardinales del 1 al 5 y números ordinales del 1° al 5° en diversas situaciones. Jugarán un juego llamado "Danza de puntos" en parejas donde lanzarán un dado y anotarán los puntos obtenidos, representándolos de forma gráfica y simbólica para verificar el aprendizaje de estos conceptos numéricos. Al final, reflexionarán sobre cómo usaron los números cardinales y ordinales en el juego.
En esta sesión, los estudiantes participarán en un juego llamado "La carrera de la amistad" para identificar su orden de llegada usando los números ordinales del 1° al 15°. Después del juego, los estudiantes ordenarán tarjetas con los números ordinales y pegarán sus nombres debajo del número que corresponde a su posición de llegada. Finalmente, practicarán la pronunciación de los números ordinales y escribirán los primeros 15 números ordinales en su cuaderno.
El documento describe una sesión de lectura sobre el Documento Nacional de Identidad (DNI) con estudiantes de primer grado. La sesión incluye leer el DNI en grupo y de forma individual, identificar los datos personales como el nombre y fecha de nacimiento, y discutir para qué sirve el DNI.
eresa Clotilde Ojeda Sánchez: el Ministerio de Educación del Perú (MINEDU) pone a disposición del personal docente el documento: Sesión de Aprendizaje 03 de Unidad Didáctica 01 del Área de Comunicación – Segundo grado de Primaria 2015: "Escribimos normas de convivencia"
Los números ordinales indican la posición o el orden de algo en una secuencia y se forman añadiendo al número cardinal la terminación -esimo, -ésimo o -igésimo dependiendo de su valor.
Los números ordinales se usan para ordenar objetos, sucesos o personas agregando el símbolo "o" al final de un número cardinal. Por ejemplo, 1o significa primero y 2o significa segundo. El documento instruye escribir un número ordinal a cada patito siguiendo el orden en que entran al agua.
Este documento presenta una unidad didáctica y sus sesiones de aprendizaje para el primer grado de primaria. La unidad se enfoca en que los estudiantes se conozcan a sí mismos y a sus familias. Incluye recomendaciones para los docentes sobre cómo implementar las sesiones de una manera flexible y adaptada a cada aula. El objetivo es desarrollar las competencias de los estudiantes en las áreas de comunicación, personal social y matemáticas.
Este documento presenta una unidad didáctica y sesiones de aprendizaje para el primer grado de primaria. La unidad se titula "Conocemos el espacio natural y social en el que vivimos" e integra las áreas de comunicación, ciencias sociales y ciencia. Incluye 31 sesiones de aprendizaje que abordan temas como el espacio donde viven los estudiantes, plantas, animales y cómo cuidar el medio ambiente. También presenta una unidad didáctica para matemáticas sobre el mismo tema.
Este documento presenta la planificación de una sesión de aprendizaje sobre los derechos y deberes de los niños para estudiantes de 4to grado. La sesión consta de tres partes: 1) Introducción, donde los estudiantes comparten características personales y conocimientos previos sobre el tema; 2) Desarrollo, que incluye la lectura sobre derechos fundamentales y la elaboración de organizadores gráficos; y 3) Cierre, con evaluación y reflexión sobre los aprendizajes y la elaboración de un afiche.
El documento instruye identificar y dibujar personas y objetos en diferentes posiciones, incluyendo encerrar niños según su posición de color y identificar y encerrar la posición de autos enumerados. También indica dibujar objetos específicos en cada piso de un edificio según su posición enumerada.
Este documento describe los pasos para realizar una asamblea en el aula con el fin de elegir a un delegado estudiantil. Incluye instrucciones sobre los materiales necesarios, la lectura de un cuento para analizar, establecer los criterios de selección, proponer candidatos, votar y juramentar al delegado elegido. El objetivo es que los estudiantes aprendan sobre la democracia y participen en el proceso de elección de su representante.
Esta unidad didáctica tiene como objetivo principal que los estudiantes conozcan y difundan sus derechos a través de diferentes actividades y productos. La unidad presenta 31 sesiones de aprendizaje en las áreas de Comunicación, Ciencia y Ambiente, y Personal Social. Los productos incluyen afiches, un periódico mural y una maqueta sobre los derechos de los niños.
Comprehensive Performance Comparison of Cosine, Walsh, Haar, Kekre, Sine, Sla...CSCJournals
This document presents a comparison of various image transforms for content-based image retrieval (CBIR) using fractional coefficients of transformed images. It proposes CBIR techniques that extract features from images transformed using discrete cosine, Walsh, Haar, Kekre, discrete sine, slant, and discrete Hartley transforms. Features are extracted from the gray-scale and individual color planes of images. Fractional coefficients, representing percentages of the full transformed image, are used to reduce feature vector size and speed up retrieval times compared to using the full transform. The techniques are tested on a database of 1000 images from 11 categories, and results show the Kekre transform achieves the best precision and recall, outperforming other transforms when using 6
Performance Improvement of Vector Quantization with Bit-parallelism HardwareCSCJournals
Vector quantization is an elementary technique for image compression; however, searching for the nearest codeword in a codebook is time-consuming. In this work, we propose a hardware-based scheme by adopting bit-parallelism to prune unnecessary codewords. The new scheme uses a “Bit-mapped Look-up Table” to represent the positional information of the codewords. The lookup procedure can simply refer to the bitmaps to find the candidate codewords. Our simulation results further confirm the effectiveness of the proposed scheme.
Performance Evaluation of Object Tracking Technique Based on Position VectorsCSCJournals
This document presents a novel algorithm for object tracking in video frames based on position vectors. The algorithm first extracts position vectors for the object in the first frame. It then tracks the object across subsequent frames by cropping each new frame into blocks based on shifted position vectors and performing block matching between frames using feature vectors extracted by discrete wavelet transform (DWT) or dual tree complex wavelet transform (DTCWT). Experimental results on video sequences show the algorithm using DTCWT achieves higher tracking precision (95%) compared to DWT (92%). The algorithm is computationally efficient and can track multiple moving and still objects.
Performance Comparison of Density Distribution and Sector mean of sal and cal...CSCJournals
In this paper we have proposed two different approaches for feature vector generation with absolute difference as similarity measuring parameter. Sal-cal vectors density distribution and Individual sector mean of complex Walsh transform. The cross over point performance of overall average of precision and recall for both approaches on all applicable sectors sizes are compared. The complex Walsh transform is conceived by multiplying sal components by j= ã-1. The density distribution of real (cal) and imaginary (sal) values and individual mean of Walsh sectors in all three color planes are considered to design the feature vector. The algorithm proposed here is worked over database of 270 images spread over 11 different classes. Overall Average precision and recall is calculated for the performance evaluation and comparison of 4, 8, 12 & 16 Walsh sectors. The overall average of cross over points of precision and recall is of all methods for both approaches are compared. The use of Absolute difference as similarity measure always gives lesser computational complexity and Individual sector mean approach of feature vector has the best retrieval.
Performance Analysis of Image Enhancement Using Dual-Tree Complex Wavelet Tra...IJERD Editor
Resolution enhancement (RE) schemes which are not based on wavelets has one of the major
drawbacks of losing high frequency contents which results in blurring. The discrete wavelet- transform-based
(DWT) Resolution Enhancement scheme generates artifacts (due to a DWT shift-variant property). A wavelet-
Domain approach based on dual-tree complex wavelet transform (DT-CWT) & nonlocal means (NLM) is
proposed for RE of the satellite images. A satellite input image is decomposed by DT-CWT (which is nearly
shift invariant) to obtain high-frequency sub bands. Here the Lanczos interpolator is used to interpolate the highfrequency
sub bands & the low-resolution (LR) input image. The high frequency sub bands are passed through
an NLM filter to cater for the artifacts generated by DT-CWT (despite of it’s nearly shift invariance). The
filtered high-frequency sub bands and the LR input image are combined by using inverse DTCWT to obtain a
resolution-enhanced image. Objective and subjective analyses show superiority of the new proposed technique
over the conventional and state-of-the-art RE techniques.
This document discusses content-based image retrieval (CBIR) systems. It covers the types of image databases and queries used, as well as common image features and distance measures for determining matches, such as color histograms, texture, shape, and objects/relationships. Relevance feedback and term weighting are described for refining search results. Specific CBIR systems are summarized, including QBIC, Blobworld, and Andy Berman's FIDS system which uses triangle inequalities for efficient retrieval. Building recognition using consistent line clusters is presented as an example of object-oriented feature extraction.
Content Based Video Retrieval in Transformed Domain using Fractional Coeffici...CSCJournals
With the development of multimedia and growing database there is huge demand of video retrieval systems. Due to this, there is a shift from text based retrieval systems to content based retrieval systems. Selection of extracted features play an important role in content based video retrieval. Good features selection also allows the time and space costs of the retrieval process to be reduced. Different methods[1,2,3] have been proposed to develop video retrievals systems to achieve better performance in terms of accuracy.
The proposed technique uses transforms to extract the features. The used transforms are Discrete Cosine, Walsh, Haar, Kekre, Discrete Sine, Slant and Discrete Hartley transforms. The benefit of energy compaction of transforms in higher coefficients is taken to reduce the feature vector size by taking fractional coefficients[5] of transformed frames of video. Smaller feature vector size results in less time for comparison of feature vectors resulting in faster retrieval of images. The feature vectors are extracted and coefficients sets are considered as feature vectors (100%, 6.25%, 3.125%, 1.5625%, 0.7813%, 0.39%, 0.195%, 0.097%, 0.048%, 0.024%, 0.012%, 0.006% and 0.003% of complete transformed coefficients). The database consists of 500 videos spread across 10 categories.
Effect of Similarity Measures for CBIR using Bins ApproachCSCJournals
This paper elaborates on the selection of suitable similarity measure for content based image retrieval. It contains the analysis done after the application of similarity measure named Minkowiski Distance from order first to fifth. It also explains the effective use of similarity measure named correlation distance in the form of angle ‘cosè’ between two vectors. Feature vector database prepared for this experimentation is based on extraction of first four moments into 27 bins formed by partitioning the equalized histogram of R, G and B planes of image into three parts. This generates the feature vector of dimension 27. Image database used in this work includes 2000 BMP images from 20 different classes. Three feature vector databases of four moments namely Mean, Standard deviation, Skewness and Kurtosis are prepared for three color intensities (R, G and B) separately. Then system enters in the second phase of comparing the query image and database images which makes of set of similarity measures mentioned above. Results obtained using all distance measures are then evaluated using three parameters PRCP, LSRR and Longest String. Results obtained are then refined and narrowed by combining the three different results of three different colors R, G and B using criterion 3. Analysis of these results with respect to similarity measures describes the effectiveness of lower orders of Minkowiski distance as compared to higher orders. Use of Correlation distance also proved its best for these CBIR results.
FACE RECOGNITION USING DIFFERENT LOCAL FEATURES WITH DIFFERENT DISTANCE TECHN...IJCSEIT Journal
A face recognition system using different local features with different distance measures is proposed in this
paper. Proposed method is fast and gives accurate detection. Feature vector is based on Eigen values,
Eigen vectors, and diagonal vectors of sub images. Images are partitioned into sub images to detect local
features. Sub partitions are rearranged into vertically and horizontally matrices. Eigen values, Eigenvector
and diagonal vectors are computed for these matrices. Global feature vector is generated for face
recognition. Experiments are performed on benchmark face YALE database. Results indicate that the
proposed method gives better recognition performance in terms of average recognized rate and retrieval
time compared to the existing methods.
This document proposes using machine learning techniques to predict COVID-19 infections based on chest x-ray images. Specifically, it involves using discrete wavelet transform to extract space-frequency features from chest x-rays, reducing the dimensionality of features using Shannon entropy, and then training standard machine learning classifiers like logistic regression, support vector machine, decision tree, and convolutional neural network on the extracted features to classify images as COVID-19 positive or negative. The document provides background on the proposed techniques of discrete wavelet transform, entropy, and various machine learning models.
This document proposes a method for change detection in images that combines Change Vector Analysis, K-Means clustering, Otsu thresholding, and mathematical morphology. It involves detecting intensity changes using CVA, segmenting the difference image using K-Means, calculating a threshold with Otsu's method, applying the threshold and morphological operations, and comparing results to other change detection techniques. Experimental results on medical and other images show the proposed method achieves satisfactory change detection with fewer errors compared to other methods.
The document reports on the results of three image processing projects. The first project implemented Lloyd-Max quantization to reduce image file sizes and Retinex theory to compensate for uneven illumination. The second project used principal component analysis to compute eigenfaces for face recognition. The third project performed linear discriminant analysis and tensor-based linear discriminant analysis for binary classification and visual object recognition. Illumination compensation subtracted an estimated illumination plane from image intensities to reduce shadows. Eigenfaces were the principal components of a training set of face images. Tensor-based linear discriminant analysis treated images as higher-order tensors to outperform conventional LDA.
The document discusses content-based image retrieval (CBIR) which involves retrieving desired images from a large collection based on automatically extracted visual features like color, texture, and shape. It describes using exact Legendre moments to represent images and support vector machines (SVM) to classify images. The algorithm trains each class independently against other classes and constructs hyperplanes to classify new images based on which planes an image's features satisfy. The method achieved over 96% accuracy on a database with features up to order 5 and 18 training images per class.
Multimodal Medical Image Fusion Based On SVDIOSR Journals
Image fusion is a promising process in the field of medical image processing, the idea behind is to
improve the content of medical image by combining two or more multimodal medical images. In this paper a
novel fusion framework based on singular value decomposition - based image fusion algorithm is proposed.
SVD is an image adaptive transform, it transforms the matrix of the given image into product USVT
, which
allows to refactor a digital image into three matrices called tensors. The proposed algorithm picks out
informative image patches of source images to constitute the fused image by processing the divided subtensors
rather than the whole tensor and a novel sigmoid-function-like coefficient-combining scheme is applied to
construct the fused result. Experimental results show that the proposed algorithm is an alternative image fusion
approach.
A broad ranging open access journal Fast and efficient online submission Expe...ijceronline
nternational Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
The document discusses a method for 3D object recognition from 2D images using centroidal representation. It involves several steps: filtering and binarizing the image, detecting edges, calculating the object center point, extracting features around the centroid, and creating mathematical models using wavelet transforms and autoregression. Centroidal samples represent distances from the center to the boundary every 45 degrees. Wavelet transforms and autoregression are used to create scale and position invariant representations of the object for recognition.
This document presents a novel iris recognition technique that uses fractional energies of transformed iris images to extract features and identify individuals. Various transforms like cosine, Walsh, Haar, Kekre, and Hartley transforms are applied to iris images to generate transformed images. Feature vectors are then extracted from the transformed images by selecting the higher energy coefficients, which represent most of the image information. Performance is evaluated using genuine acceptance rate on different percentages of higher energies, from 100% down to 96%. The technique is tested on a database of 384 iris images from 64 individuals. Results show that cosine and Walsh transforms achieve the best genuine acceptance rate of 85% when using 99% of the fractional energies. Considering fractional energies improves performance by reducing computations compared to
This document describes a content-based image retrieval system that uses 2-D discrete wavelet transform with texture features. It proposes using DWT to reduce image dimensions before extracting texture features from images using gray level co-occurrence matrix. Texture features and Euclidean distance are then used to retrieve similar images from a database. The system is tested on a dataset of 1000 images from 10 classes and achieves an average retrieval accuracy of 89.8%.
Content Based Image Retrieval Using 2-D Discrete Wavelet TransformIOSR Journals
This document proposes a content-based image retrieval system using 2D discrete wavelet transform and texture features. The system decomposes images using 2D DWT, extracts texture features from low frequency coefficients using GLCM, and retrieves similar images by calculating Euclidean distances between feature vectors. Experimental results on Wang's database show the proposed approach achieves 89.8% average retrieval accuracy.
This document presents a study on medial axis transformation (MAT) based skeletonization of image patterns using image processing techniques. It discusses how the MAT of an image can be extracted by first computing the Euclidean distance transform of the binary image. Local maxima in the distance transform image correspond to the MAT. Several performance evaluation metrics for analyzing skeletonized images are also introduced, such as connectivity number, thinness measurement and sensitivity. The technique is demonstrated on sample images and results show it can effectively extract the skeleton with good computational speed.
Similar to Performance Comparison of Image Retrieval Using Fractional Coefficients of Transformed Image Using DCT, Walsh, Haar and Kekre’s Transform (20)
A review of the growth of the Israel Genealogy Research Association Database Collection for the last 12 months. Our collection is now passed the 3 million mark and still growing. See which archives have contributed the most. See the different types of records we have, and which years have had records added. You can also see what we have for the future.
Introduction to AI for Nonprofits with Tapp NetworkTechSoup
Dive into the world of AI! Experts Jon Hill and Tareq Monaur will guide you through AI's role in enhancing nonprofit websites and basic marketing strategies, making it easy to understand and apply.
How to Add Chatter in the odoo 17 ERP ModuleCeline George
In Odoo, the chatter is like a chat tool that helps you work together on records. You can leave notes and track things, making it easier to talk with your team and partners. Inside chatter, all communication history, activity, and changes will be displayed.
How to Fix the Import Error in the Odoo 17Celine George
An import error occurs when a program fails to import a module or library, disrupting its execution. In languages like Python, this issue arises when the specified module cannot be found or accessed, hindering the program's functionality. Resolving import errors is crucial for maintaining smooth software operation and uninterrupted development processes.
This presentation includes basic of PCOS their pathology and treatment and also Ayurveda correlation of PCOS and Ayurvedic line of treatment mentioned in classics.
This slide is special for master students (MIBS & MIFB) in UUM. Also useful for readers who are interested in the topic of contemporary Islamic banking.
हिंदी वर्णमाला पीपीटी, hindi alphabet PPT presentation, hindi varnamala PPT, Hindi Varnamala pdf, हिंदी स्वर, हिंदी व्यंजन, sikhiye hindi varnmala, dr. mulla adam ali, hindi language and literature, hindi alphabet with drawing, hindi alphabet pdf, hindi varnamala for childrens, hindi language, hindi varnamala practice for kids, https://www.drmullaadamali.com
Thinking of getting a dog? Be aware that breeds like Pit Bulls, Rottweilers, and German Shepherds can be loyal and dangerous. Proper training and socialization are crucial to preventing aggressive behaviors. Ensure safety by understanding their needs and always supervising interactions. Stay safe, and enjoy your furry friends!
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...Dr. Vinod Kumar Kanvaria
Exploiting Artificial Intelligence for Empowering Researchers and Faculty,
International FDP on Fundamentals of Research in Social Sciences
at Integral University, Lucknow, 06.06.2024
By Dr. Vinod Kumar Kanvaria
How to Manage Your Lost Opportunities in Odoo 17 CRMCeline George
Odoo 17 CRM allows us to track why we lose sales opportunities with "Lost Reasons." This helps analyze our sales process and identify areas for improvement. Here's how to configure lost reasons in Odoo 17 CRM
A Strategic Approach: GenAI in EducationPeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
Performance Comparison of Image Retrieval Using Fractional Coefficients of Transformed Image Using DCT, Walsh, Haar and Kekre’s Transform
1. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 142
Performance Comparison of Image Retrieval Using Fractional
Coefficients of Transformed Image Using DCT, Walsh, Haar
and Kekre’s Transform
Dr. H. B. Kekre hbkekre@yahoo.com
Senior Professor, Computer Engineering,
MPSTME, SVKM’S NMIMS University,
Mumbai, 400056, India
Sudeep D. Thepade sudeepthepade@gmail.com
Ph.D. Research Scholar and Asst. Professor,
Computer Engineering,
MPSTME, SVKM’S NMIMS University,
Mumbai, 400056, India
Akshay Maloo akshaymaloo@gmail.com
Student, Computer Engineering,
MPSTME, SVKM’S NMIMS University
Mumbai, 400056, India
Abstract
The thirst of better and faster retrieval techniques has always fuelled to the research in
content based image retrieval (CBIR). The paper presents innovative content based image
retrieval (CBIR) techniques based on feature vectors as fractional coefficients of transformed
images using Discrete Cosine, Walsh, Haar and Kekre’s transforms. Here the advantage of
energy compaction of transforms in higher coefficients is taken to greatly reduce the feature
vector size per image by taking fractional coefficients of transformed image. The feature
vectors are extracted in fourteen different ways from the transformed image, with the first
being considering all the coefficients of transformed image and then fourteen reduced
coefficients sets (as 50%, 25%, 12.5%, 6.25%, 3.125%, 1.5625% ,0.7813%, 0.39%, 0.195%,
0.097%, 0.048%, 0.024%, 0.012% and 0.06% of complete transformed image) are considered
as feature vectors. The four transforms are applied on gray image equivalents and the colour
components of images to extract Gray and RGB feature sets respectively. Instead of using all
coefficients of transformed images as feature vector for image retrieval, these fourteen
reduced coefficients sets for gray as well as RGB feature vectors are used, resulting into
better performance and lower computations. The proposed CBIR techniques are implemented
on a database having 1000 images spread across 11 categories. For each proposed CBIR
technique 55 queries (5 per category) are fired on the database and net average precision
and recall are computed for all feature sets per transform. The results have shown
performance improvement (higher precision and recall values) with fractional coefficients
compared to complete transform of image at reduced computations resulting in faster retrieval.
Finally Kekre’s transform surpasses all other discussed transforms in performance with
highest precision and recall values for fractional coefficients (6.25% and 3.125% of all
coefficients) and computation are lowered by 94.08% as compared to DCT.
Keywords: CBIR, Discrete Cosine Transform (DCT), Walsh Transform, Haar Transform, Kekre’s
Transform, Fractional Coefficients, Feature Vector.
1. INTRODUCTION
The computer systems have been posed with large number of challenges to store/transmit
and index/manage large numbers of images effectively, which are being generated from a
variety of sources. Storage and transmission is taken care by Image compression with
significant advancements been made [1,4,5].Image databases deal with the challenge of
2. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 143
image indexing and retrieval[2,6,7,10,11], which has become one of the promising and
important research area for researchers from a wide range of disciplines like computer vision,
image processing and database areas. The thirst of better and faster image retrieval
techniques is till appetising to the researchers working in some of important applications for
CBIR technology like art galleries [12,14], museums, archaeology [3], architecture design
[8,13], geographic information systems [5], weather forecast [5,22], medical imaging [5,18],
trademark databases [21,23], criminal investigations [24,25], image search on the Internet
[9,19,20].
1.1 Content Based Image Retrieval
In literature the term content based image retrieval (CBIR) has been used for the first time by
Kato et.al.[4], to describe his experiments into automatic retrieval of images from a database
by colour and shape feature. The typical CBIR system performs two major tasks [16,17]. The
first one is feature extraction (FE), where a set of features, called feature vector, is generated
to accurately represent the content of each image in the database. The second task is
similarity measurement (SM), where a distance between the query image and each image in
the database using their feature vectors is used to retrieve the “closest” images [16,17,26].
For CBIR feature extraction the two main approaches are feature extraction in spatial domain
[5] and feature extraction in transform domain [1]. The feature extraction in spatial domain
includes the CBIR techniques based on histograms [5], BTC [2,16,23], VQ [21,25,26]. The
transform domain methods are widely used in image compression, as they give high energy
compaction in transformed image[17,24]. So it is obvious to use images in transformed
domain for feature extraction in CBIR [1]. Transform domain results in energy compaction in
few elements, so large number of the coefficients of transformed image can be neglected to
reduce the size of feature vector [1]. Reducing the size feature vector using fractional
coefficients of transformed image and till getting the improvement in performance of image
retrieval is the theme of the work presented here. Many current CBIR systems use average
Euclidean distance [1,2,3,8-14,23]on the extracted feature set as a similarity measure. The
direct Average Euclidian Distance (AED) between image P and query image Q can be given
as equation 1, where Vpi and Vqi are the feature vectors of image P and Query image Q
respectively with size ‘n’.
∑=
−=
n
i
VqiVpi
n
AED
1
2
)(
1
(1)
2. DISCRETE COSINE TRANSFORM
The discrete cosine transform (DCT) [1,10,21,22,24] is closely related to the discrete Fourier
transform. It is a separable linear transformation; that is, the two-dimensional transform is
equivalent to a one-dimensional DCT performed along a single dimension followed by a one-
dimensional DCT in the other dimension. The definition of the two-dimensional DCT for an
input image A and output image B is
10
10
,
2
)12(
cos
2
)12(
cos
−≤≤
−≤≤++
= ∑∑ Nq
Mp
N
qn
M
pm
AB mnqppq
ππ
αα (2)
−≤≤
=
=
11,2
0,1
MpM
pM
pα (3)
−≤≤
=
=
11,2
0,1
NqN
qN
qα (4)
where M and N are the row and column size of A, respectively. If you apply the DCT to real
data, the result is also real. The DCT tends to concentrate information, making it useful for
image compression applications and also helping in minimizing feature vector size in CBIR
[23]. For full 2-Dimensional DCT for an NxN image the number of multiplications required are
N2(2N) and number of additions required are N2(2N-2).
3. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 144
3. WALSH TRANSFORM
Walsh transform matrix [1,11,18,19,26,30] is defined as a set of N rows, denoted Wj, for j = 0,
1, .... , N - 1, which have the following properties:
• Wj takes on the values +1 and -1.
• Wj[0] = 1 for all j.
• WjxWk
T
=0, for j ≠ k and WjxWk
T
=N, for j=k.
• Wj has exactly j zero crossings, for j = 0, 1, ...., N-1.
• Each row Wj is even or odd with respect to its midpoint.
Walsh transform matrix is defined using a Hadamard matrix of order N. The Walsh transform
matrix row is the row of the Hadamard matrix specified by the Walsh code index, which must
be an integer in the range [0, ..., N - 1]. For the Walsh code index equal to an integer j, the
respective Hadamard output code has exactly j zero crossings, for j = 0, 1,..., N - 1. The step
of the algorithm to generate Walsh matrix from hadamard matrix by reordering hadamard
matrix is given below [30].
Step 1 : Let H be the hadamard matix of size NxN and W be the expected Walsh
Matrix of same size
Step 2 : Let seq=0, cseq=0, seq(0)=0, seq(1)=1, i=0
Step 3 : Repeat steps 3 to 12 till i <= log2(N)-2
Step 4 : s=size(seq)
Step 5 : Let j=1, Repeat steps 6 and 7 till j<=s(1)
Step 6 : cseq(j)=2*seq(j)
Step 7 : j=j+1
Step 8 : Let p=1, k=2*s(2) repeat steps 9 to 11 until k<=s(2)+1
Step 9 : cseq(k)=cseq(p)+1
Step 10 : p=p+1 and k=k-1
Step 11 : seq=cseq,
Step 12 : i=i+1
Step 13 : Let seq=seq+1
Step 14 : Let x and y indicate the rows and columns of ‘seq’
Step 15 : Let i=0 Repeat steps 16 and 17 till i<= y-1
Step 16 : q=seq(i)
Step 17 : i=i+1
Step 18 : Let i=0, repeat steps 19 to 22 till i<=s1-1
Step 19 : for j=0, repeat steps 20 and 21 till j<=s1-1
Step 20 : W(i,j)=H(seq(i),j)
Step 21 : j=j+1
Step 22 : i=i+1
For the full 2-Dimensional Walsh transform applied to image of size NxN, the number of
additions required are 2N
2
(N-1) and absolutely no multiplications are needed in Walsh
transform [1].
4. HAAR TRANSFORM
This sequence was proposed in 1909 by Alfréd Haar [28]. Haar used these functions to give
an example of a countable orthonormal system for the space of square-integral functions on
the real line. The study of wavelets, and even the term "wavelet", did not come until much
later [29,31].The Haar wavelet is also the simplest possible wavelet. The technical
disadvantage of the Haar wavelet is that it is not continuous, and therefore not differentiable.
This property can, however, be an advantage for the analysis of signals with sudden
transitions, such as monitoring of tool failure in machines. The Haar wavelet's mother wavelet
function (t) can be described as:
(5)
4. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 145
and its scaling function φ(t) can be described as:
(6)
5. KEKRE’S TRANSFORM
Kekre’s transform matrix is the generic version of Kekre’s LUV color space matrix
[1,8,12,13,15,22]. Kekre’s transform matrix can be of any size NxN, which need not have to
be in powers of 2 (as is the case with most of other transforms). All upper diagonal and
diagonal values of Kekre’s transform matrix are one, while the lower diagonal part except the
values just below diagonal is zero.
Generalized NxN Kekre’s transform matrix can be given as:
−+−
+−
+−
=
1)1(..000
11..000
11..120
11..111
11..111
NN
N
N
KNxN
MMMMMM
(7)
The formula for generating the term Kxy of Kekre’s transform matrix is:
(8)
For taking Kekre’s transform of an NxN image, the number of required
multiplications are 2N(N-2) and number of additions required are N(N2
+N-2).
DCT Walsh Haar Kekre’s
Number of Additions 2N
2
(N-1) 2N
2
(N-1) 2N
2
log2(N) N[N(N+1)-2]
Number of Multiplications N
2
(2N) 0 0 2N(N-2)
Total Additions for transform of
256x256 image
301,858,816 33,423,360 1,048,576 17,882,624
Computations Comparison
For 256x256 image
100 % 11.07% 0.35% 5.92%
[Here one multiplication is considered as eight additions for second last row
computations and DCT computations are considered to be 100% for
comparison in last row]
TABLE 1: Computational Complexity for applying transforms to image of size NxN [1]
5. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 146
1.a. Flowchart of proposed CBIR Technique
1.b. Feature Extraction for Proposed CBIR Techniques
FIGURE 1: Proposed CBIR Techniques using fractional Coefficients of Transformed Images
6. PROPOSED CBIR-GRAY TECHNIQUES
Figure 1.a gives the flowchart of proposed CBIR technique for feature extraction and query
execution. Figure 1.b explains the feature sets extraction used to extract feature sets for
proposed CBIR techniques using fractional coefficients of transformed images.
6.1 Feature Extraction for feature vector ‘T-Gray’
Here the feature vector space of the image of size NxN has NxN number of elements. This is
obtained using following steps of T-Gray
Extract Red, Green and Blue components of the colour image.
i. Take average of Red, Green and Blue components of respective pixels
to get gray image.
6. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 147
ii. Apply the Transform ‘T’ on gray image to extract feature vector.
iii. The result is stored as the complete feature vector ‘T-Gray’ for the
respective image.
Thus the feature vector database for DCT, Walsh, Haar and Kekre’s transform are generated
as DCT-Gray, Walsh-Gray, Haar-Gray, Kekre’s-Gray respectively. Here the size of feature
vector is NxN for every transform.
6.2 Feature Vector Database ‘Fractional T-Gray’
The fractional coefficients of transformed image as shown in figure 1, are considered to form
‘fractional T-Gray’ feature vector databases. Here first 50% of coefficients from upper
triangular part of feature vector ‘T-Gray’ are considered to prepare the feature vector
database ‘50%-T-Gray’ for every image as shown in figure 1. Thus DCT-Gray, Walsh-Gray,
Haar-Gray, Kekre’s-Gray feature databases are used to obtain new feature vector databases
as 50%-DCT-Gray, 50%-Walsh-Gray, 50%-Haar-Gray, 50%-Kekre’s-Gray respectively. Then
per image first 25% number of coefficients (as shown in figure 1) form the feature vectors
database DCT-Gray, Walsh-Gray, Haar-Gray, Kekre’s-Gray are stored separately as feature
vector databases as 25%-DCT-Gray, 25%-Walsh-Gray, 25%-Haar-Gray, 25%-Kekre’s-Gray
respectively. Then for each image in the database as shown in figure 1, fractional feature
vector set for DCT-Gray, Walsh-Gray, Haar-Gray, Kekre’s-Gray using 25%, 12.5%, 6.25%,
3.125%, 1.5625% ,0.7813%, 0.39%, 0.195%, 0.097%, 0.048%, 0.024%, 0.012% and 0.06%
of total coefficients are formed.
6.3 Query Execution for ‘T-Gray’ CBIR
Here the feature set of size NxN for the query image is extracted using transform ‘T’. This
feature set is compared with each entry from the feature database using Euclidian distance as
similarity measure. Thus DCT, Walsh, Haar, Kekre’s transform based feature sets are
extracted from query image and are compared respectively with DCT-Gray and Walsh-Gray
feature sets using average Euclidian distance to find the best match in the database.
6.4 Query Execution for ‘Fractional T-Gray’ CBIR
For 50%-T-Gray query execution, only 50% number of coefficients of upper triangular part of
‘T’ transformed query image (with NxN coefficients) are considered for the CBIR and are
compared with ‘50%-T-Gray’ database feature set for Euclidian distance computations. Thus
DCT, Walsh, Haar, Kekre’s transform based feature sets are extracted from the query image
and are compared respectively with 50%-DCT-Gray, 50%-Walsh-Gray, 50%-Haar-Gray, 50%-
Kekre’s-Gray feature sets to find average Euclidian distances. For 25%, 12.5%, 6.25%,
3.125%, 1.5625% ,0.7813%, 0.39%, 0.195%, 0.097%, 0.048%, 0.024%, 0.012% and 0.06%
T-Gray based query execution, the feature set of the respective percentages are considered
from the ‘T’ transformed NxN image as shown in figure 1, to be compared with the respective
percentage T-Gray feature set database to find average Euclidian distance.
7. PROPOSED CBIR-RGB TECHNIQUES
7.1 Feature Extraction for feature vector ‘T-RGB’
Here the feature vector space of the image of size NxNx3 has NxNx3 number of elements.
This is obtained using following steps of T-RGB
i. Extract Red, Green and Blue components of the color image.
ii. Apply the Transform ‘T’ on individual color planes of image to extract feature vector.
iii. The result is stored as the complete feature vector ‘T-RGB’ for the respective image.
Thus the feature vector database for DCT, Walsh, Haar, Kekre’s transform is generated as
DCT-RGB, Walsh-RGB, Haar-RGB, Kekre’s-RGB respectively. Here the size of feature
database is NxNx3.
7.2 Query Execution for ‘T-RGB’ CBIR
Here the feature set of NxNx3 for the query image is extracted using transform ‘T’ applied on
the red, green and blue planes of query image. This feature set is compared with other
feature sets in feature database using Euclidian distance as similarity measure. Thus DCT,
Walsh, Haar, Kekre’s transform based feature sets are extracted for query image and are
7. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 148
compared respectively with DCT-RGB, Walsh-RGB, Haar-RGB, Kekre’s-RGB feature sets to
find Euclidian distance.
7.3 CBIR using ‘Fractional-T-RGB’
As explained in section 6- 6.4 and section 7 – 7.2 the ‘T-RGB’ feature extraction and query
execution are extended to get 50%,25%, 12.5%, 6.25%, 3.125%, 1.5625% ,0.7813%, 0.39%,
0.195%, 0.097%, 0.048%, 0.024%, 0.012% and 0.006% of T-RGB image retrieval techniques.
8. IMPLEMENTATION
The implementation of the three CBIR techniques is done in MATLAB 7.0 using a computer
with Intel Core 2 Duo Processor T8100 (2.1GHz) and 2 GB RAM. The CBIR techniques are
tested on the image database [15] of 1000 variable size images spread across 11 categories
of human being, animals, natural scenery and manmade things. The categories and
distribution of the images is shown in table 2.
Category Tribes Buses Beaches Dinosaurs Elephants Roses
No.of Images 85 99 99 99 99 99
Category Horses Mountains Airplanes Monuments Sunrise
No.of Images 99 61 100 99 61
TABLE 2: Image Database: Category-wise Distribution
FIGURE 2: Sample Database Images
[Image database contains total 1000 images with 11 categories]
Figure 2 gives the sample database images from all categories of images including scenery,
flowers, buses, animals, aeroplanes, monuments, and tribal people. To assess the retrieval
effectiveness, we have used the precision and recall as statistical comparison parameters
[1,2] for the proposed CBIR techniques. The standard definitions of these two measures are
given by following equations.
(5)
8. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 149
retrievedimagesofnumberTotal
retrievedimagesrelevantofNumber
ecision
____
____
Pr =
databaseinimagesreleventofnumberTotal
retrievedimagesrelevantofNumber
call
______
____
Re = (6)
9. RESULTS AND DISCUSSION
For testing the performance of each proposed CBIR technique, per technique 55 queries (5
from each category) are fired on the database of 1000variable size generic images spread
across 11 categories. The query and database image matching is done using average
Euclidian distance. The average precision and average recall are computed by grouping the
number of retrieved images sorted according to ascending average Euclidian distances with
the query image. In all transforms, the average precision and average recall values for CBIR
using fractional coefficients are higher than CBIR using full set of coefficients. The CBIR-RGB
techniques are giving higher values of crossover points than CBIR-Gray techniques indicating
better performance. The crossover point of precision and recall of the CBIR techniques acts
as one of the important parameters to judge their performance [1,2,19,20].
Figure 3 shows the precision-recall crossover points plotted against number of retrieved
images for proposed image retrieval techniques using DCT. Uniformly in all image retrieval
techniques based on gray DCT and colour DCT features 0.012% fractional feature set
(1/8192
th
of total coefficients) based image retrieval gives highest precision and recall values.
Figure 3.a gives average precision/recall values plotted against number of retrieved images
for all DCT-Gray image retrieval techniques. Precision/recall values for DCT-RGB image
retrieval techniques are plotted in figure 3.b.
Figures 4.a and 4.b respectively shows the graphs of precision/recall values plotted against
number of retrieved images for Walsh-Gray and Walsh-RGB based image retrieval
techniques. Here 1/4096th
fractional coefficients (0.024% of total Walsh transformed
coefficients) based image retrieval gives the highest precision/recall crossover values
specifying the best performance using Walsh transform.
3.a. DCT-Gray based CBIR
9. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 150
3.b. DCT-RGB based CBIR
FIGURE 3: Crossover Point of Precision and Recall for DCT based CBIR.
4.a. Walsh-Gray based CBI
4.b. Walsh-RGB based CBIR
FIGURE 4: Crossover Point of Precision and Recall for Walsh T. based CBIR
Figures 5.a and 5.b respectively shows the graphs of precision/recall values plotted against
number of retrieved images for Haar-Gray and Haar-RGB based image retrieval techniques.
Here 1/4096th
fractional coefficients (0.024% of total Haar transformed coefficients) based
image retrieval gives the highest precision/recall crossover values specifying the best
performance when using Haar transform.
Figure 6.a gives average precision/recall values plotted against number of retrieved images
for all DCT-Gray image retrieval techniques. Precision/recall values for DCT-RGB image
retrieval techniques are plotted in figure 6.b. Here 1/32
th
fractional coefficients (3.125% of
10. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 151
total Kekre’s transformed coefficients) based image retrieval gives the highest precision/recall
crossover values specifying the best performance when using Kekre’s transform.
5.a. Haar-Gray based CBIR
5.b. Haar-RGB based CBIR
FIGURE 5: Crossover Point of Precision and Recall for Haar T. based CBIR
6.a. Kekre’s-Gray based CBIR
11. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 152
6.b. Kekre’s-RGB based CBIR
FIGURE 6: Crossover Point of Precision and Recall for Kekre’s T. based CBIR
Figure 7 shows the performance comparison of all the four transformsfor
proposed CBIR techniques. Figure 7.a is indicating the crossover points of DCT-Gray, Walsh-
Gray, Haar-Gray, Kekre’s-Gray CBIR for all considered feature vectors (percentage of
coefficients of transformed gray images). Herefor upto 0.78 % of coefficients Kekre’s
transform performs better than all discussed transforms after which Haar transform
outperforms other transformsupto 0.012 % ofcoefficients and finally for 0.006 % of coefficients
DCT gives highest crossover point value, as the energy compaction in Haar and DCT
transform is better than other transform. For Kekre’s-Gray CBIR the performance improves
with decreasing feature vector size from 100% to 0.195% and then drops indicating 0.195%
as best fractional coefficients. In DCT-Gray CBIR the performance is improved till 0.012% and
then drops. Overall in all, CBIR using Kekre’s transform with 3.125 % of fractional coefficients
gives the best performance for Gray-CBIR techniques discussed here.
7.a. Transform Comparison in Gray based CBIR
12. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 153
7.b. Transform Comparison in Color based CBIR
FIGURE 7: Performance Comparison of Fractional Walsh-CBIR and Fractional DCT-CBIR
Figure 7.b indicates the performance comparison of DCT-RGB, Walsh- RGB, Haar- RGB,
Kekre’s- RGB CBIR with different percentage of fractional coefficients. Here Kekre’s-RGB
CBIR outperforms all other transforms till 0.097% of coefficients as feature vector then Walsh-
RGB CBIR takes over till 0.024% then DCT-RGB performs best for 0.012% of coefficients. In
Walsh-RGB and Haar-RGB CBIR the feature vector with 0.024% of coefficients gives best
performance, in DCT-RGB CBIR 0.012% of coefficients shows highest crossover value of
average precision and average recall and Kekre’s transform gives the best performance when
6.25% of coefficients are considered. In all, CBIR using Kekre’s transform with 6.25 % of
fractional coefficients gives the best performance for RGB-CBIR techniques discussed here.
10.CONCLUSION
In the information age where the size of image databases is growing exponentially more
precise retrieval techniques are needed, for finding relatively similar images. Computational
complexity and retrieval efficiency are the key objectives in the image retrieval system.
Nevertheless it is very difficult to reduce the computations and improve the performance of
image retrieval technique.
Here the performance of image retrieval is improved using fractional coefficients of
transformed images at reduced computational complexity. In all transforms (DCT, Walsh,
Haar and Kekre’s), the average precision and average recall values for CBIR using fractional
coefficients are higher than CBIR using full set of coefficients. Hence the feature vector size
for image retrieval could be greatly reduced, which ultimately will result in faster query
execution in CBIR with better performance. In all Kekre’s transform with fractional coefficients
(3.125 % in Gray and 6.25 % in RGB) gives best performance with highest crossover points
of average precision and average recall. Feature extraction using Kekre’s transform is also
computationally lighter as compared to DCT or Walsh transform. Thus feature extraction in
lesser time is possible with increased performance.
Finally the conclusion that the fractional coefficients gives better discrimination capability in
CBIR than the complete set of transformed coefficients and image retrieval with better
performance at much faster rate can be done from the proposed techniques and
experimentation done.
13. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 154
11.REFERENCES
1. H.B.Kekre, Sudeep D. Thepade, “Improving the Performance of Image Retrieval using
Partial Coefficients of Transformed Image”, International Journal of Information Retrieval
(IJIR), Serials Publications, Volume 2, Issue 1, 2009, pp. 72-79(ISSN: 0974-6285)
2. H.B.Kekre, Sudeep D. Thepade, “Image Retrieval using Augmented Block Truncation
Coding Techniques”, ACM International Conference on Advances in Computing,
Communication and Control (ICAC3-2009), pp. 384-390, 23-24 Jan 2009, Fr.
ConceicaoRodrigous College of Engg., Mumbai. Is uploaded on online ACM portal.
3. H.B.Kekre, Sudeep D. Thepade, “Scaling Invariant Fusion of Image Pieces in Panorama
Making and Novel Image Blending Technique”, International Journal on Imaging (IJI),
www.ceser.res.in/iji.html, Volume 1, No. A08, pp. 31-46, Autumn 2008.
4. Hirata K. and Kato T. “Query by visual example – content-based image retrieval”, In
Proc. of Third International Conference on Extending Database Technology, EDBT’92,
1992, pp 56-71
5. H.B.Kekre, Sudeep D. Thepade, “Rendering Futuristic Image Retrieval System”, National
Conference on Enhancements in Computer, Communication and Information
Technology, EC2IT-2009, 20-21 Mar 2009, K.J.Somaiya College of Engineering,
Vidyavihar, Mumbai-77.
6. Minh N. Do, Martin Vetterli, “Wavelet-Based Texture Retrieval Using Generalized
Gaussian Density and Kullback-Leibler Distance”, IEEE Transactions On Image
Processing, Volume 11, Number 2, pp.146-158, February 2002.
7. B.G.Prasad, K.K. Biswas, and S. K. Gupta, “Region –based image retrieval using
integrated color, shape, and location index”, International Journal on Computer Vision
and Image Understanding Special Issue: Colour for Image Indexing and Retrieval,
Volume 94, Issues 1-3, April-June 2004, pp.193-233.
8. H.B.Kekre, Sudeep D. Thepade, “Creating the Color Panoramic View using Medley of
Grayscale and Color Partial Images ”, WASET International Journal of Electrical,
Computer and System Engineering (IJECSE), Volume 2, No. 3, Summer 2008. Available
online at www.waset.org/ijecse/v2/v2-3-26.pdf.
9. Stian Edvardsen, “Classification of Images using color, CBIR Distance Measures and
Genetic Programming”, Ph.D. Thesis, Master of science in Informatics, Norwegian
university of science and Technology, Department of computer and Information science,
June 2006.
10. H.B.Kekre, Tanuja Sarode, Sudeep D. Thepade, “DCT Applied to Row Mean and
Column Vectors in Fingerprint Identification”, In Proceedings of International Conference
on Computer Networks and Security (ICCNS), 27-28 Sept. 2008, VIT, Pune.
11. Zhibin Pan, Kotani K., Ohmi T., “Enhanced fast encoding method for vector quantization
by finding an optimally-ordered Walsh transform kernel”, ICIP 2005, IEEE International
Conference, Volume 1, pp I - 573-6, Sept. 2005.
12. H.B.kekre, Sudeep D. Thepade, “Improving ‘Color to Gray and Back’ using Kekre’s LUV
Color Space”, IEEE International Advanced Computing Conference 2009 (IACC’09),
Thapar University, Patiala, INDIA, 6-7 March 2009. Is uploaded and available online at
IEEE Xplore.
13. H.B.Kekre, Sudeep D. Thepade, “Image Blending in Vista Creation using Kekre's LUV
Color Space”, SPIT-IEEE Colloquium and International Conference, Sardar Patel
Institute of Technology, Andheri, Mumbai, 04-05 Feb 2008.
14. H.B.Kekre, Sudeep D. Thepade, “Color Traits Transfer to Grayscale Images”, In Proc.of
IEEE First International Conference on Emerging Trends in Engg. & Technology,
(ICETET-08), G.H.Raisoni COE, Nagpur, INDIA. Uploaded on online IEEE Xplore.
15. http://wang.ist.psu.edu/docs/related/Image.orig (Last referred on 23 Sept 2008)
16. H.B.Kekre, Sudeep D. Thepade, “Using YUV Color Space to Hoist the Performance of
Block Truncation Coding for Image Retrieval”, IEEE International Advanced Computing
Conference 2009 (IACC’09), Thapar University, Patiala, INDIA, 6-7 March 2009.
17. H.B.Kekre, Sudeep D. Thepade, ArchanaAthawale, Anant Shah, PrathmeshVerlekar,
SurajShirke,“Energy Compaction and Image Splitting for Image Retrieval using Kekre
Transform over Row and Column Feature Vectors”, International Journal of Computer
Science and Network Security (IJCSNS),Volume:10, Number 1, January 2010, (ISSN:
1738-7906) Available at www.IJCSNS.org.
18. H.B.Kekre, Sudeep D. Thepade, ArchanaAthawale, Anant Shah, PrathmeshVerlekar,
SurajShirke,“Walsh Transform over Row Mean and Column Mean using Image
14. Dr. H. B. Kekre, Sudeep D. Thepade & Akshay Maloo
International Journal of Image Processing (IJIP) Volume (4): Issue (2) 155
Fragmentation and Energy Compaction for Image Retrieval”, International Journal on
Computer Science and Engineering (IJCSE),Volume 2S, Issue1, January 2010, (ISSN:
0975–3397). Available online at www.enggjournals.com/ijcse.
19. H.B.Kekre, Sudeep D. Thepade,“Image Retrieval using Color-Texture Features
Extracted from Walshlet Pyramid”, ICGST International Journal on Graphics, Vision and
Image Processing (GVIP), Volume 10, Issue I, Feb.2010, pp.9-18, Available online
www.icgst.com/gvip/Volume10/Issue1/P1150938876.html
20. H.B.Kekre, Sudeep D. Thepade,“Color Based Image Retrieval using Amendment Block
Truncation Coding with YCbCrColor Space”, International Journal on Imaging (IJI),
Volume 2, Number A09, Autumn 2009, pp. 2-14. Available online at
www.ceser.res.in/iji.html (ISSN: 0974-0627).
21. H.B.Kekre, Tanuja Sarode, Sudeep D. Thepade,“Color-Texture Feature based Image
Retrieval using DCT applied on Kekre’s Median Codebook”, International Journal on
Imaging (IJI), Volume 2, Number A09, Autumn 2009,pp. 55-65. Available online at
www.ceser.res.in/iji.html (ISSN: 0974-0627).
22. H.B.Kekre, Sudeep D. Thepade, “Image Retrieval using Non-Involutional Orthogonal
Kekre’s Transform”, International Journal of Multidisciplinary Research and Advances in
Engineering (IJMRAE), Ascent Publication House, 2009, Volume 1, No.I, pp 189-203,
2009. Abstract available online at www.ascent-journals.com (ISSN: 0975-7074)
23. H.B.Kekre, Sudeep D. Thepade, “Boosting Block Truncation Coding using Kekre’s LUV
Color Space for Image Retrieval”, WASET International Journal of Electrical, Computer
and System Engineering (IJECSE), Volume 2, Number 3, pp. 172-180, Summer 2008.
Available online at http://www.waset.org/ijecse/v2/v2-3-23.pdf
24. H.B.Kekre, Sudeep D. Thepade, Archana Athawale, Anant Shah, Prathmesh Verlekar,
Suraj Shirke, “Performance Evaluation of Image Retrieval using Energy Compaction and
Image Tiling over DCT Row Mean and DCT Column Mean”, Springer-International
Conference on Contours of Computing Technology (Thinkquest-2010),
BabasahebGawde Institute of Technology, Mumbai, 13-14 March 2010, The paper will
be uploaded on online Springerlink.
25. H.B.Kekre, Tanuja K. Sarode, Sudeep D. Thepade, VaishaliSuryavanshi,“Improved
Texture Feature Based Image Retrieval using Kekre’s Fast Codebook Generation
Algorithm”, Springer-International Conference on Contours of Computing Technology
(Thinkquest-2010), BabasahebGawde Institute of Technology, Mumbai, 13-14 March
2010, The paper will be uploaded on online Springerlink.
26. H.B.Kekre, Tanuja K. Sarode, Sudeep D. Thepade, “Image Retrieval by Kekre’s
Transform Applied on Each Row of Walsh Transformed VQ Codebook”, (Invited), ACM-
International Conference and Workshop on Emerging Trends in Technology (ICWET
2010),Thakur College of Engg. And Tech., Mumbai, 26-27 Feb 2010, The paper is
invited at ICWET 2010. Also will be uploaded on online ACM Portal.
27. H.B.Kekre, Sudeep D. Thepade, AkshayMaloo, “Image Retrieval using Fractional
Coefficients of Transformed Image using DCT and Walsh Transform”, IJEST.
28. Haar, Alfred, “ZurTheorie der orthogonalenFunktionensysteme”. (German),
MathematischeAnnalen, volume 69, No. 3, 1910, pp. 331–371.
29. Charles K. Chui, “An Introduction to Wavelets”, Academic Press, 1992, San Diego, ISBN
0585470901.
30. H. B. Kekre, Tanuja K. Sarode, V. A. Bharadi, A. Agrawal. R. Arora,, M. Nair,
“Performance Comparison of Full 2-D DCT, 2-D Walsh and 1-D Transform over Row
Mean and Column Mean for Iris Recognition” International Conference and Workshop on
Emerging Trends in Technology (ICWET 2010) – 26-27 February 2010, TCET, Mumbai,
India.
31. M.C. Padma,P. A. Vijaya, “Wavelet Packet Based Features for Automatic Script
Identification”, International Journal Of Image Processing (IJIP), CSC Journals, 2009,
Volume 4, Issue 1, Pg.53-65.