This document presents a scalable method for image classification using sparse coding and dictionary learning. It proposes parallelizing the computation of image similarity for faster recognition. Specifically, it distributes the task of measuring similarity between images among multiple cores in a cluster. Experimental results on a face recognition dataset show nearly linear speedup when balancing the dataset size and number of nodes. Reconstruction errors are used as a similarity measure, with dictionaries learned using K-SVD for each image. The proposed parallel method distributes this similarity computation process to achieve faster image classification.
Content Based Image Retrieval (CBIR) is one of the
most active in the current research field of multimedia retrieval.
It retrieves the images from the large databases based on images
feature like color, texture and shape. In this paper, Image
retrieval based on multi feature fusion is achieved by color and
texture features as well as the similarity measures are
investigated. The work of color feature extraction is obtained by
using Quadratic Distance and texture features by using Pyramid
Structure Wavelet Transforms and Gray level co-occurrence
matrix. We are comparing all these methods for best image
retrieval
Image Super-Resolution Reconstruction Based On Multi-Dictionary LearningIJRESJOURNAL
ABSTRACT: In order to overcome the problems that the single dictionary cannot be adapted to variety types of images and the reconstruction quality couldn’t meet the application, we propose a novel Multi-Dictionary Learning algorithm for feature classification. The algorithm uses the orientation information of the low resolution image to guide the image patches in the database to classify, and designs the classification dictionary which can effectively express the reconstructed image patches. Considering the nonlocal similarity of the image, we construct the combined nonlocal mean value(C-NLM) regularizer, and take the steering kernel regression(SKR) to formulate a local regularization ,and establish a unified reconstruction framework. Extensive experiments on single image validate that the proposed method, compared with several other state-of-the-art learning based algorithms, achieves improvement in image quality and provides more details.
Behavior study of entropy in a digital image through an iterative algorithmijscmcj
Image segmentation is a critical step in computer vision tasks constituting an essential issue for pattern recognition and visual interpretation. In this paper, we study the behavior of entropy in digital images through an iterative algorithm of mean shift filtering. The order of a digital image in gray levels is defined. The behavior of Shannon entropy is analyzed and then compared, taking into account the number of iterations of our algorithm, with the maximum entropy that could be achieved under the same order. The use of equivalence classes it induced, which allow us to interpret entropy as a hyper-surface in real m dimensional space. The difference of the maximum entropy of order n and the entropy of the image is used to group the the iterations, in order to caractrizes the performance of the algorithm.
Content Based Image Retrieval (CBIR) is one of the
most active in the current research field of multimedia retrieval.
It retrieves the images from the large databases based on images
feature like color, texture and shape. In this paper, Image
retrieval based on multi feature fusion is achieved by color and
texture features as well as the similarity measures are
investigated. The work of color feature extraction is obtained by
using Quadratic Distance and texture features by using Pyramid
Structure Wavelet Transforms and Gray level co-occurrence
matrix. We are comparing all these methods for best image
retrieval
Image Super-Resolution Reconstruction Based On Multi-Dictionary LearningIJRESJOURNAL
ABSTRACT: In order to overcome the problems that the single dictionary cannot be adapted to variety types of images and the reconstruction quality couldn’t meet the application, we propose a novel Multi-Dictionary Learning algorithm for feature classification. The algorithm uses the orientation information of the low resolution image to guide the image patches in the database to classify, and designs the classification dictionary which can effectively express the reconstructed image patches. Considering the nonlocal similarity of the image, we construct the combined nonlocal mean value(C-NLM) regularizer, and take the steering kernel regression(SKR) to formulate a local regularization ,and establish a unified reconstruction framework. Extensive experiments on single image validate that the proposed method, compared with several other state-of-the-art learning based algorithms, achieves improvement in image quality and provides more details.
Behavior study of entropy in a digital image through an iterative algorithmijscmcj
Image segmentation is a critical step in computer vision tasks constituting an essential issue for pattern recognition and visual interpretation. In this paper, we study the behavior of entropy in digital images through an iterative algorithm of mean shift filtering. The order of a digital image in gray levels is defined. The behavior of Shannon entropy is analyzed and then compared, taking into account the number of iterations of our algorithm, with the maximum entropy that could be achieved under the same order. The use of equivalence classes it induced, which allow us to interpret entropy as a hyper-surface in real m dimensional space. The difference of the maximum entropy of order n and the entropy of the image is used to group the the iterations, in order to caractrizes the performance of the algorithm.
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
A Hybrid Steganography Technique Based On LSBMR And OPAPIOSRJVSP
This paper presents a steganography technique based on two existing methods of data hiding i.e. LSBMR and OPAP. The proposed method uses the non-overlapping blocks, having three consecutive pixels. The center pixel of this block embeds k- bits of secret data using OPAP and the remaining pixels of the block embed data using LSBMR. The experimental results show that the proposed method provides better embedding capacity while maintaining the good image quality. The improved performance is shown in comparison to other data hiding methods that are investigated in this study.
A Scheme for Joint Signal Reconstruction in Wireless Multimedia Sensor Networksijma
In context aware wireless multimedia sensor networks, scenarios are usually such that
signals of multiple distributed sensors contain a common sparse component and each individual
signal owns an innovation sparse component. So distributed compressive sensing based on joint
sparsity of a signal ensemble concept exploits both these intra- and inter- signal correlation structures
and compress signals to the extent possible. This paper proposes an optimized reconstruction
scheme based on joint sparsity model which is derived from the distributed compressive sensing. In
this regard, based on distributed compressive sensing, a joint reconstruction scheme is proposed to
compress and reconstruct ensemble of signals even in large scale data transmission. Furthermore,
simulation results show the effectiveness of the proposed method in diverse compression ratios and
processing times in comparison with the joint sparsity model and individual compressive sensing
reconstruction methods.
Computer apparition plays the most important role in human perception, which is limited to only the visual band of the electromagnetic spectrum. The need for Radar imaging systems, to recover some sources that
are not within human visual band, is raised. This paper present new algorithm for Synthetic Aperture Radar (SAR) images segmentation based on thresholding technique. Entropy based image thresholding has
received sustainable interest in recent years. It is an important concept in the area of image processing.
Pal (1996) proposed a cross entropy thresholding method based on Gaussian distribution for bi-modal images. Our method is derived from Pal method that segment images using cross entropy thresholding based on Gamma distribution and can handle bi-modal and multimodal images. Our method is tested using
Synthetic Aperture Radar (SAR) images and it gave good results for bi-modal and multimodal images. The
results obtained are encouraging.
Comparative Analysis of Lossless Image Compression Based On Row By Row Classi...IJERA Editor
Lossless image compression is needed in many fields like medical imaging, telemetry, geophysics, remote
sensing and other applications, which require exact replica of original image and loss of information is not
tolerable. In this paper, a near lossless image compression algorithm based on row by row classifier with
encoding schemes like Lempel Ziv Welch (LZW), Huffman and Run Length Encoding (RLE) on color images
is proposed. The algorithm divides the image into three parts R, G and B, apply row by row classification on
each part and result of this classification is records in the mask image. After classification the image data is
decomposed into two sequences each for R, G and B and mask image is hidden in them. These sequences are
encoded using different encoding schemes like LZW, Huffman and RLE. An exhaustive comparative analysis is
performed to evaluate these techniques, which reveals that the pro
Steganalysis of LSB Embedded Images Using Gray Level Co-Occurrence MatrixCSCJournals
This paper proposes a steganalysis technique for both grayscale and color images. It uses the feature vectors derived from gray level co-occurrence matrix (GLCM) in spatial domain, which is sensitive to data embedding process. This GLCM matrix is derived from an image. Several combinations of diagonal elements of GLCM are considered as features. There is difference between the features of stego and non-stego images and this characteristic is used for steganalysis. Distance measures like Absolute distance, Euclidean distance and Normalized Euclidean distance are used for classification. Experimental results demonstrate that the proposed scheme outperforms the existing steganalysis techniques in attacking LSB steganographic schemes applied to spatial domain.
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
A Hybrid Steganography Technique Based On LSBMR And OPAPIOSRJVSP
This paper presents a steganography technique based on two existing methods of data hiding i.e. LSBMR and OPAP. The proposed method uses the non-overlapping blocks, having three consecutive pixels. The center pixel of this block embeds k- bits of secret data using OPAP and the remaining pixels of the block embed data using LSBMR. The experimental results show that the proposed method provides better embedding capacity while maintaining the good image quality. The improved performance is shown in comparison to other data hiding methods that are investigated in this study.
A Scheme for Joint Signal Reconstruction in Wireless Multimedia Sensor Networksijma
In context aware wireless multimedia sensor networks, scenarios are usually such that
signals of multiple distributed sensors contain a common sparse component and each individual
signal owns an innovation sparse component. So distributed compressive sensing based on joint
sparsity of a signal ensemble concept exploits both these intra- and inter- signal correlation structures
and compress signals to the extent possible. This paper proposes an optimized reconstruction
scheme based on joint sparsity model which is derived from the distributed compressive sensing. In
this regard, based on distributed compressive sensing, a joint reconstruction scheme is proposed to
compress and reconstruct ensemble of signals even in large scale data transmission. Furthermore,
simulation results show the effectiveness of the proposed method in diverse compression ratios and
processing times in comparison with the joint sparsity model and individual compressive sensing
reconstruction methods.
Computer apparition plays the most important role in human perception, which is limited to only the visual band of the electromagnetic spectrum. The need for Radar imaging systems, to recover some sources that
are not within human visual band, is raised. This paper present new algorithm for Synthetic Aperture Radar (SAR) images segmentation based on thresholding technique. Entropy based image thresholding has
received sustainable interest in recent years. It is an important concept in the area of image processing.
Pal (1996) proposed a cross entropy thresholding method based on Gaussian distribution for bi-modal images. Our method is derived from Pal method that segment images using cross entropy thresholding based on Gamma distribution and can handle bi-modal and multimodal images. Our method is tested using
Synthetic Aperture Radar (SAR) images and it gave good results for bi-modal and multimodal images. The
results obtained are encouraging.
Comparative Analysis of Lossless Image Compression Based On Row By Row Classi...IJERA Editor
Lossless image compression is needed in many fields like medical imaging, telemetry, geophysics, remote
sensing and other applications, which require exact replica of original image and loss of information is not
tolerable. In this paper, a near lossless image compression algorithm based on row by row classifier with
encoding schemes like Lempel Ziv Welch (LZW), Huffman and Run Length Encoding (RLE) on color images
is proposed. The algorithm divides the image into three parts R, G and B, apply row by row classification on
each part and result of this classification is records in the mask image. After classification the image data is
decomposed into two sequences each for R, G and B and mask image is hidden in them. These sequences are
encoded using different encoding schemes like LZW, Huffman and RLE. An exhaustive comparative analysis is
performed to evaluate these techniques, which reveals that the pro
Steganalysis of LSB Embedded Images Using Gray Level Co-Occurrence MatrixCSCJournals
This paper proposes a steganalysis technique for both grayscale and color images. It uses the feature vectors derived from gray level co-occurrence matrix (GLCM) in spatial domain, which is sensitive to data embedding process. This GLCM matrix is derived from an image. Several combinations of diagonal elements of GLCM are considered as features. There is difference between the features of stego and non-stego images and this characteristic is used for steganalysis. Distance measures like Absolute distance, Euclidean distance and Normalized Euclidean distance are used for classification. Experimental results demonstrate that the proposed scheme outperforms the existing steganalysis techniques in attacking LSB steganographic schemes applied to spatial domain.
Analysis of Multimedia Traffic Performance in a Multi-Class traffic environme...IOSR Journals
Abstract: The computer networks have evolved themselves into an altogether new generation with Mobile Ad-Hoc networks. The Mobile Ad-Hoc Networks are increasingly becoming more sophisticated and complex in terms of topology, routing and security[1][2][3]The new age MANETs incorporate routing of heavier traffic classes like audio, video and multimedia. It has become important to study the performance characteristics of the multimedia traffic class in MANETS [10] that includes packet loss rate and throughput. In this paper we will discuss these performance parameter throughput under different scenarios like varying Bandwidth, channel error rate, delay and fragment size. Keywords: MANET, Packet loss rate, Throughput, Bandwidth, Fragment size
A Comparative Study on Balance and Flexibility between Dancer and Non-Dancer ...IOSR Journals
Abstract: Dance is a form of art that normally involves rhythmic movement of the body and accompanied with
music. Movement of human body while performing dance can become a significant medium for communication,
feelings and emotions. It embraces movement, creation and performance. Dance helps to extend the limits of
human physical ability, expressiveness and spirit. When it comes to health dance can be a very effective way of
establishing a lasting healthy living. Anecdotally it can be said that dance potentially motivate and excite young
people. Dance is a non-competitive form of exercise which has positive effects on physical and mental health.
Young girls can be engaged in physical activity through dance. The author being a dancer in fervor and passion
as well as an aspirant of the profession Physical Education strived to conduct the study bearing the title “A
Comparative Study on Balance and Flexibility between Dancer and Non-Dancer Girls”. The researcher
selected 30 girls who are regularly involved in Dance and 30 girls who are non-dancer or rather sedentary on
the basis of purposive stratified random sampling from Bidhan Govt. Girl’s School, Dist. Nadia West Bengal as
the subjects of her study. She incorporated Sit and Reach test and Stork Stand Balance tests for assessment of
Flexibility and Balance respectively. With respect to data analysis initially descriptive statics like mean SD and
range and further paired sample T test was conducted to ascertain the degree of difference between the means
with the help of SPSS soft ware. Data analysis proved significant difference between the Dancer and NonDancer
girls both with respect to flexibility and Balance. In both the cases the Dancer girls proved to be better
though the differences were not statistically significant. Thus the author arrived at the conclusion that dance
involving passion, strength, stamina, enthusiasm, rhythm, amusement and many more could be a wonderful
fitness activity similar to other fitness activities like jogging, running, cycling, swimming etc.
Key words: Dance, Flexibility, Balance, Dancer, Non-Dancer.
Performance Evaluation of IEEE STD 802.16d TransceiverIOSR Journals
WiMAX ("Worldwide Interoperability for Microwave Access") technology is developed to meet the
growing demand of increased data rate and accessing the internet at high speeds. 802.16 family of standards is
officially called Wireless MAN in IEEE. Orthogonal frequency division multiplexing (OFDM) is multicarrier
modulation technique used in IEEE 802.16d (fixed WiMAX) communication standard. OFDM is used to
increase data rate of wireless medium with higher spectral efficiency. The proposed work is to evaluate
performance of IEEE Std 802.16d transceiver in MATLAB R2009b simulink environment. System performance
evaluated using BER vs SNR for different modulation technique such as 4 QAM, 16 QAM, 64 QAM under
different channel condition
BEHAVIOR STUDY OF ENTROPY IN A DIGITAL IMAGE THROUGH AN ITERATIVE ALGORITHM O...ijscmcj
Image segmentation is a critical step in computer vision tasks constituting an essential issue for pattern
recognition and visual interpretation. In this paper, we study the behavior of entropy in digital images
through an iterative algorithm of mean shift filtering. The order of a digital image in gray levels is defined.
The behavior of Shannon entropy is analyzed and then compared, taking into account the number of
iterations of our algorithm, with the maximum entropy that could be achieved under the same order. The
use of equivalence classes it induced, which allow us to interpret entropy as a hyper-surface in real m-
dimensional space. The difference of the maximum entropy of order n and the entropy of the image is used
to group the the iterations, in order to caractrizes the performance of the algorithm
Dictionary based Image Compression via Sparse Representation IJECEIAES
Nowadays image compression has become a necessity due to a large volume of images. For efficient use of storage space and data transmission, it becomes essential to compress the image. In this paper, we propose a dictionary based image compression framework via sparse representation, with the construction of a trained over-complete dictionary. The overcomplete dictionary is trained using the intra-prediction residuals obtained from different images and is applied for sparse representation. In this method, the current image block is first predicted from its spatially neighboring blocks, and then the prediction residuals are encoded via sparse representation. Sparse approximation algorithm and the trained overcomplete dictionary are applied for sparse representation of prediction residuals. The detail coefficients obtained from sparse representation are used for encoding. Experimental result shows that the proposed method yields both improved coding efficiency and image quality as compared to some state-of-the-art image compression methods.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
Super resolution image reconstruction via dual dictionary learning in sparse...IJECEIAES
Patch-based super resolution is a method in which spatial features from a low-resolution (LR) patch are used as references for the reconstruction of high-resolution (HR) image patches. Sparse representation for each patch is extracted. These coefficients obtained are used to recover HR patch. One dictionary is trained for LR image patches, and another dictionary is trained for HR image patches and both dictionaries are jointly trained. In the proposed method, high frequency (HF) details required are treated as combination of main high frequency (MHF) and residual high frequency (RHF). Hence, dual-dictionary learning is proposed for main dictionary learning and residual dictionary learning. This is required to recover MHF and RHF respectively for recovering finer image details. Experiments are carried out to test the proposed technique on different test images. The results illustrate the efficacy of the proposed algorithm.
Web image annotation by diffusion maps manifold learning algorithmijfcstjournal
Automatic image annotation is one of the most challenging problems in machine vision areas. The goal of this task is to predict number of keywords automatically for images captured in real data. Many methods are based on visual features in order to calculate similarities between image samples. But the computation cost of these approaches is very high. These methods require many training samples to be stored in memory. To lessen thisburden, a number of techniques have been developed to reduce the number
of features in a dataset. Manifold learning is a popular approach to nonlinear dimensionality reduction. In
this paper, we investigate Diffusion maps manifold learning method for webimage auto-annotation task.Diffusion maps
manifold learning method isused to reduce the dimension of some visual features. Extensive experiments and analysis onNUS-WIDE-LITE web image dataset with
different visual featuresshow how this manifold learning dimensionality reduction method can be applied effectively to image annotation.
Clustered Compressive Sensingbased Image Denoising Using Bayesian Frameworkcsandit
This paper provides a compressive sensing (CS) method of denoising images using Bayesian
framework. Some images, for example like magnetic resonance images (MRI) are usually very
weak due to the presence of noise and due to the weak nature of the signal itself. So denoising
boosts the true signal strength. Under Bayesian framework, we have used two different priors:
sparsity and clusterdness in an image data as prior information to remove noise. Therefore, it is
named as clustered compressive sensing based denoising (CCSD). After developing the
Bayesian framework, we applied our method on synthetic data, Shepp-logan phantom and
sequences of fMRI images. The results show that applying the CCSD give better results than
using only the conventional compressive sensing (CS) methods in terms of Peak Signal to Noise
Ratio (PSNR) and Mean Square Error (MSE). In addition, we showed that this algorithm could
have some advantages over the state-of-the-art methods like Block-Matching and 3D
Filtering (BM3D).
This paper provides a compressive sensing (CS) method of denoising images using Bayesian framework. Some images, for example like magnetic resonance images (MRI) are usually very
weak due to the presence of noise and due to the weak nature of the signal itself. So denoising
boosts the true signal strength. Under Bayesian framework, we have used two different priors:
sparsity and clusterdness in an image data as prior information to remove noise. Therefore, it is
named as clustered compressive sensing based denoising (CCSD). After developing the
Bayesian framework, we applied our method on synthetic data, Shepp-logan phantom and
sequences of fMRI images. The results show that applying the CCSD give better results than
using only the conventional compressive sensing (CS) methods in terms of Peak Signal to Noise
Ratio (PSNR) and Mean Square Error (MSE). In addition, we showed that this algorithm could
have some advantages over the state-of-the-art methods like Block-Matching and 3D
Filtering (BM3D).
Template matching is a basic method in image analysis to extract useful information from images. In this
paper, we suggest a new method for pattern matching. Our method transform the template image from two
dimensional image into one dimensional vector. Also all sub-windows (same size of template) in the
reference image will transform into one dimensional vectors. The three similarity measures SAD, SSD, and
Euclidean are used to compute the likeness between template and all sub-windows in the reference image
to find the best match. The experimental results show the superior performance of the proposed method
over the conventional methods on various template of different sizes.
Sinusoidal Function for Population Size in Quantum Evolutionary Algorithm and...sipij
Fractal Image Compression is a well-known problem which is in the class of NP-Hard problems. Quantum Evolutionary Algorithm is a novel optimization algorithm which uses a probabilistic representation for solutions and is highly suitable for combinatorial problems like Knapsack problem. Genetic algorithms are widely used for fractal image compression problems, but QEA is not used for this kind of problems yet. This paper improves QEA whit change population size and used it in fractal image compression. Utilizing the self-similarity property of a natural image, the partitioned iterated function system (PIFS) will be found to encode an image through Quantum Evolutionary Algorithm (QEA) method Experimental results show that our method has a better performance than GA and conventional fractal image compression algorithms.
Joint Image Registration And Example-Based Super-Resolution Algorithmaciijournal
Supper-resolution (SR) methods are classified into two different methods: image registration (IR)-based
methods and example-based methods. The proposed joint SR method is focused on estimating highresolution (HR) video sequences from low-resolution (LR) ones by combining the two different methods.
The IR-based SR method collects information from adjacent frames to reconstruct HR images in the video
sequence. Example-based SR methods give good textures and strong edges in the result HR video. In this
paper, IR-based and example-based SR methods are fused based on the gradient features. The proposed
joint SR method gives smaller peak signal to noise ratio than the example-based method, however it shows
better reconstruction results on high-level features such as characters in images. Experimental result of the
proposed joint SR method shows less noise and higher contrast than the example-based method.
Band Clustering for the Lossless Compression of AVIRIS Hyperspectral ImagesIDES Editor
Hyperspectral images can be efficiently compressed
through a linear predictive model, as for example the one
used in the SLSQ algorithm. In this paper we exploit this
predictive model on the AVIRIS images by individuating,
through an off-line approach, a common subset of bands, which
are not spectrally related with any other bands. These bands
are not useful as prediction reference for the SLSQ 3-D
predictive model and we need to encode them via other
prediction strategies which consider only spatial correlation.
We have obtained this subset by clustering the AVIRIS bands
via the clustering by compression approach. The main result
of this paper is the list of the bands, not related with the
others, for AVIRIS images. The clustering trees obtained for
AVIRIS and the relationship among bands they depict is also
an interesting starting point for future research.
A systematic image compression in the combination of linear vector quantisati...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Blind Image Seperation Using Forward Difference Method (FDM)sipij
In this paper, blind image separation is performed, exploiting the property of sparseness to represent images. A new sparse representation called forward difference method is proposed. It is known that most of the independent component analysis (ICA) basis functions, extracted from images are sparse and gives unreliable sparseness measure. In the proposed method, the image mixture is first transformed to sparse images. These images are divided into blocks and for each block the sparseness measure ε0 norm is applied. The block having the most sparseness is considered to determine the separation matrix. The efficiency of the proposed method is compared with other sparse representation functions.
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...CSCJournals
The Internet paved way for information sharing all over the world decades ago and its popularity for distribution of data has spread like a wildfire ever since. Data in the form of images, sounds, animations and videos is gaining users’ preference in comparison to plain text all across the globe. Despite unprecedented progress in the fields of data storage, computing speed and data transmission speed, the demands of available data and its size (due to the increase in both, quality and quantity) continue to overpower the supply of resources. One of the reasons for this may be how the uncompressed data is compressed in order to send it across the network. This paper compares the two most widely used training algorithms for multilayer perceptron (MLP) image compression – the Levenberg-Marquardt algorithm and the Scaled Conjugate Gradient algorithm. We test the performance of the two training algorithms by compressing the standard test image (Lena or Lenna) in terms of accuracy and speed. Based on our results, we conclude that both algorithms were comparable in terms of speed and accuracy. However, the Levenberg- Marquardt algorithm has shown slightly better performance in terms of accuracy (as found in the average training accuracy and mean squared error), whereas the Scaled Conjugate Gradient algorithm faired better in terms of speed (as found in the average training iteration) on a simple MLP structure (2 hidden layers).
SVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONSijscmcj
The purpose of this article is to determine the usefulness of the Graphics Processing Unit (GPU) calculations used to implement the Latent Semantic Indexing (LSI) reduction of the TERM-BY DOCUMENT matrix. Considered reduction of the matrix is based on the use of the SVD (Singular Value Decomposition) decomposition. A high computational complexity of the SVD decomposition - O(n3), causes that a reduction of a large indexing structure is a difficult task. In this article there is a comparison of the time complexity and accuracy of the algorithms implemented for two different environments. The first environment is associated with the CPU and MATLAB R2011a. The second environment is related to graphics processors and the CULA library. The calculations were carried out on generally available benchmark matrices, which were combined to achieve the resulting matrix of high size. For both considered environments computations were performed for double and single precision data.
An Image representation using Compressive Sensing and Arithmetic Coding IJCERT
The demand for graphics and multimedia communication over intenet is growing day by day. Generally the coding efficiency achieved by CS measurements is below the widely used wavelet coding schemes (e.g., JPEG 2000). In the existing wavelet-based CS schemes, DWT is mainly applied for sparse representation and the correlation of DWT coefficients has not been fully exploited yet. To improve the coding efficiency, the statistics of DWT coefficients has been investigated. A novel CS-based image representation scheme has been proposed by considering the intra- and inter-similarity among DWT coefficients. Multi-scale DWT is first applied. The low- and high-frequency subbands of Multi-scale DWT are coded separately due to the fact that scaling coefficients capture most of the image energy. At the decoder side, two different recovery algorithms have been presented to exploit the correlation of scaling and wavelet coefficients well. In essence, the proposed CS-based coding method can be viewed as a hybrid compressed sensing schemes which gives better coding efficiency compared to other CS based coding methods.
SINGLE IMAGE SUPER RESOLUTION: A COMPARATIVE STUDYcsandit
The majority of applications requiring high resolution images to derive and analyze data
accurately and easily. Image super resolution is playing an effective role in those applications.
Image super resolution is the process of producing high resolution image from low resolution
image. In this paper, we study various image super resolution techniques with respect to the
quality of results and processing time. This comparative study introduces a comparison between
four algorithms of single image super-resolution. For fair comparison, the compared algorithms
are tested on the same dataset and same platform to show the major advantages of one over the
others.
A COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVALcscpconf
Basic group of visual techniques such as color, shape, texture are used in Content Based Image Retrievals (CBIR) to retrieve query image or sub region of image to find similar images in image database. To improve query result, relevance feedback is used many times in CBIR to help user to express their preference and improve query results. In this paper, a new approach for image retrieval is proposed which is based on the features such as Color Histogram, Eigen Values and Match Point. Images from various types of database are first identified by using edge detection techniques .Once the image is identified, then the image is searched in the particular database, then all related images are displayed. This will save the retrieval time. Further to retrieve the precise query image, any of the three techniques are used and comparison is done w.r.t. average retrieval time. Eigen value technique found to be the best as compared with other two techniques.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
A1804010105
1. IOSR Journal of Computer Engineering (IOSR-JCE)
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 18, Issue 4, Ver. I (Jul.-Aug. 2016), PP 01-05
www.iosrjournals.org
DOI: 10.9790/0661-1804010105 www.iosrjournals.org 1 | Page
Scalable Image Classification Using Compression
Akshata K. Naik1
, Dr. Dinesh Acharya U2
1
(Computer Science and Engineering, Manipal Institute of Technology, India)
2
(Computer Science and Engineering, Manipal Institute of Technology, India)
Abstract: The increasing rate of data growth has led to finding techniques for faster processing of data. Big
Data analytics has recently emerged as a promising field for examining huge volume of datasets containing
different data types. It is a known fact that image processing and retrieval involves high computation especially
with a large dataset. We present a scalable method for face recognition based on sparse coding and dictionary
learning. Sparse representation has closer resemblance with a cortex like image representation and thus more
closer to human perception. The proposed method parallelizes the computation of image similarity for faster
recognition.
Keywords: Dictionary learning, Image classification, Parallel computation, Sparse representation.
I. Introduction
Data mining is a field of computer science, dealing with finding of hidden patterns from data. Recently,
a line of research has developed towards parameter free data mining [1] [2]. The field exploits the Kolmogorov
complexity which is a theoretical base for measuring randomness of data. The complexity however is not a
computable quantity. It is thus closely measured by the size of data. The interesting aspect of this field is that,
the methods employed are generic and application independent. One of the most widely used distance metric is
the Normalized Compression Distance (NCD). It has been proved that the distance metric is successful in
clustering and classification of one-dimensional data [1] [2]. The successful implementation of NCD in the field
of multi-dimensional data like images is limited, as the compressors used for images are not a normal
compressor [3].
Sparse representation is one of the effective methods for representation of signals and image processing
[4] [5]. It has been widely used for various applications in image processing viz. image denoising, image
compression and image classification [6] [7] [8]. Sparse coding is a class of unsupervised learning to find a set
of over-complete basis referred to as the dictionary. A dictionary of dimension m x n is called an
over-complete dictionary if m < n. The aim is to learn such that an input vector b can be represented as a
linear combination of these basis vectors. An over-complete dictionary produces a system of linear equations
with infinite number of solutions [8]. The solution with few number of non-zero coefficients is more desirable.
Such a solution is formulated as :
The vector is the sparse representation of input vector and is the permissible error. An
over-complete dictionary can either be pre-specified or learnt from training data. Learning a dictionary has an
added advantage of adapting itself to fit a given set of signals. K-SVD algorithm, which is a generalization of
the K-means clustering, is a popular dictionary learning algorithm.
Image similarity based on sparse reconstruction error uses the above mentioned dictionary learning
technique [9]. A similarity measure viz. sparse Signal-to-noise ratio (SSNR) has been proposed by the authors.
The measure involves the sparse coding which is computationally intensive. This paper proposes a parallel
computation method for face recognition using SSNR. The proposed method distributes the task of measuring
similarity between images, among multiple cores of the same as well as different computers in a cluster. The
method has shown an almost linear speedup when there is a balance between the size of dataset and the number
of nodes.
1.1. Dictionary Learning
Dictionary learning is based on a given set of examples. Given such a set B = { } such that, 1 ≤ i ≤
N, a dictionary is learnt by solving the equation (1) for each of the example . K-SVD is one of the
successfully used dictionary learning algorithm. The algorithm iterates between a sparse coding stage and
dictionary update stage. Dictionary for an image is learnt by extracting patches of dimension . The
patches are then converted to vector ϵ Rm
. Each of these patches act as a column vector for the matrix B.
2. Scalable Image Classification Using Compression
DOI: 10.9790/0661-1804010105 www.iosrjournals.org 2 | Page
We then learn an over-complete dictionary that has n atoms (basis vector) using the patches from
B as input. The dictionary , should be able to closely approximate each patch as a linear combination of a
small number of atoms [8].
Fig. 1 (a) Block diagram of K-SVD algorithm; (b) Initialized patches for dictionary learning
K-SVD algorithm finds an approximation to equation (1) using two steps. The 2 steps are 1) sparse
coding and 2) dictionary update. The first step employs any pursuit algorithm to compute a sparse vector ,
when the dictionary is fixed. We use a greedy algorithm called Orthogonal Matching Pursuit (OMP) due to
its simplicity and faster implementation. The second step updates the dictionary columns sequentially along with
the corresponding coefficients in . The detail implementation of K-SVD algorithm can be referred from the
paper [10]. The initialized patches for dictionary learning and the block diagram for K-SVD is shown in Fig.
1(a) and 1(b) respectively.
1.2. Similarity Measure
The similarity measure SSNR [9] computes the similarity between two images X and Y. Computation
of similarity between X and Y requires learning of dictionaries Dx and Dy for the images respectively. Each of
the images are then reconstructed using both the dictionaries. The SSNR uses the Signal-to-noise ratio (SNR)
between the original image and reconstructed image. SSNR function is given as:
In equation (2) and are set of random patches belonging to images X and Y respectively. The
term and are closest approximation of and using the dictionaries Dy and Dx respectively. Similarly
and are approximation of and using the dictionaries Dx and Dy respectively. SNR is the Signal-
to-noise ratio. Higher value of numerator indicates that the images are similar. The denominator is used for
normalization. The SSNR has the following properties [9]:
a) Non-negativity: value of SSNR(X, Y) lies between 0 and 1. SSNR = 1 when X = Y.
b) Symmetry: SSNR(X, Y) = SSNR(Y, X)
It should be noted that the reconstruction of patches has sparsity (number of non-zero coefficients) constraint on
the sparse vector. The sparsity is set to a value β. It is formulated as below:
In equation (3) is a vector ϵ and is a sparse vector with non-zero coefficients. Similarly for vector
ϵ and the sparse vector , reconstruction is given by equation (4)
The rest of the paper is organized as follows. Section 2 gives a brief description of the background
work. The section 3 describes the methodology employed. Section 4 presents the experimental results. The
paper is concluded in section 5 with possible directions for improvisation of the current method.
3. Scalable Image Classification Using Compression
DOI: 10.9790/0661-1804010105 www.iosrjournals.org 3 | Page
II. Background work
Data mining paradigm based on compression allows parameter free or parameter light solutions to data
mining tasks. It further helps in exploratory data mining rather than imposing presumptions on data [1].
Normalized Information Distance (NID) is based on the concepts of Kolmogorov‟s complexity, which is a
measure of randomness of strings based on their information content [11]. In algorithmic theory, the
Kolmogorov complexity K(x) of a string x, is defined as “the length of the shortest program capable of
producing x on a universal computer”.
The quantity K(x) is incomputable and therefore can be approximated by the compression algorithms.
The Normalized Compression Distance (NCD) is based on this result and is given by:
where, C(x) is the length of the (lossless) compressed file of x, and C(x, y) the length of compressed
file obtained by concatenation of x and y. The concept is that, if x and y share a large amount of mutual
information, then the compressor will reuse the repeating structures found in one, to compress the other. NCD
metric has been successfully applied in clustering languages and music [2]. Researchers have also tried to apply
the concept for classification of images. The successful application of NCD metric in the field of images is
limited, as most of the image compressors are not a normal compressor. The definition of normal compressor
can be found in the paper [3].
Another research founded on Kolmogorov principle has given a distance measure called CK-1 for
classification of texture [12]. The researchers use the MPEG compressor to measure the similarity between two
images by constructing a video consisting of both the images. A different research work has proposed similarity
measure based on sparse representation [8].The concept of sparseness of natural images is used to measure the
degree of compression. Using the K-SVD algorithm, one dictionary per image is learnt. Sparse Complexity and
Relative Sparse Complexity [8] is then used to measure compressibility of an image. This, in turn, is used to
quantify similarity. Dictionary learning techniques have also been employed for image classification in other
works [13] [14]. Another approach uses sparse reconstruction errors for measuring the similarity between two
images [9]. The authors have proposed a similarity measure SSNR. Our work is based on the similarity measure
described in the paper [9]. The accuracy for face recognition [9] is remarkable in comparison to state-of-the-art
methods.
Fig. 2. Examples of few images from AT&T dataset used for face recognition.
Fig. 3. Sanity check: Image reconstruction using the dictionaries learnt.
4. Scalable Image Classification Using Compression
DOI: 10.9790/0661-1804010105 www.iosrjournals.org 4 | Page
III. Methodology
Our main objective is to parallelize the process of computing similarity between images using the
similarity measure given in [9]. It is worth noting that the image reconstruction process using the Orthogonal
Matching Pursuit algorithm involves matrix computation of higher dimensions. The size of matrices and number
of computations involved thus make the parallel approach a suitable candidate for solving the problem faster. To
test the efficiency of the parallel approach, task of face recognition is implemented on multiple nodes. The entire
process of finding similar images from a given dataset involves the following two phases:
Training Phase : To construct dictionary using K-SVD algorithm for the training set
Testing Phase: To construct the dictionary for query image and compute similarity with all images in the
training set.
Once the similarities are computed, k-NN with k =3 is used to classify the query image. The value of k
is chosen based upon the previous research work in face recognition [9]. Considering the speed of the testing
phase in real time system, the main focus is on the testing aspect of the problem. The training phase can also be
parallelized in the similar manner, but it is out of the scope of this paper. We exploit the concept of data
parallelism or single instruction multiple data (SIMD) computation. The calculation of similarity for each image
involves four reconstruction operations (as mentioned in equation 2). Following are the four reconstructions:
1) Reconstruction of dataset image patches using dataset dictionary
2) Reconstruction of dataset image patches using query dictionary
3) Reconstruction of query image patches using dataset dictionary
4) Reconstruction of query image patches using query dictionary
Equations (3) and (4) are the basis for the reconstruction task. The paper proposes to implement the
function of similarity measure (which includes the above 4 sub tasks) in parallel on a cluster of four nodes. Each
of the node in turn uses its own multiple cores (in this case 4 cores) to implement the method. Thus the task of
similarity computation of n images is divided among the 4 nodes on the cluster in a load balancing fashion.
IV. Experimental validation
In this section we measure the efficiency of our method in terms of speedup and scaleup. In order to
measure the speedup, the data set size is kept constant and number of computers on the cluster are changed.
Scaleup measures the scalability of the system and is calculated by increasing both the size of data set and the
number of computers on the cluster proportionally. The training data has to be communicated to all the nodes in
cluster. This communication process occurs only once and is considered as a part of system startup. Hence, the
communication cost for training data has been neglected in computation of speedup and scaleup. At the runtime,
only the query data needs to be sent to the nodes. It can be noted that, this runtime communication cost is almost
same for the various number of nodes we considered. Therefore, the speedup and scaleup for varying number of
nodes in the cluster is determined largely by the processing time. The experiments are performed using R 3.2.2
on a cluster of 4 computers, each having the following specification: 8 GB RAM and 3.0 GHz Intel core i5
processor. Additional R packages viz. parallel, EBImage [15] and pixmap are used for parallelization and image
processing. The „pixmap‟ package is used for reading bitmap images and converting it to grayscale images for
color invariance. Similarly the „EBImage‟ is used for processing jpeg images. The „parallel‟ package is
integrated as a part of R from R 2.14.0 onwards. This package is based on „multicore‟ and „snow‟ packages. In
this paper, a set of worker processes are created that listens to the master node commands in the cluster via
sockets. As a sanity check, the images are reconstructed using its own dictionary to validate the reconstruction
process. Fig. 3 shows an example of the sanity check performed. The number of K-SVD iterations and structural
similarity (SSIM) values are also as indicated in the figure.
Table 1. Processing time in minutes
Nodes per Cluster Time for 70 images Time for 140 images Time for 210 images Time for 280 images
1 2.13 4.22 6.23 8.37
2 1.05 2.06 3.05 4.06
3 0.71 1.35 2.02 2.67
4 0.56 1.02 1.51 2.01
The dataset used for validation is AT&T face dataset [16]. The dataset consists of 400 grayscale
images. The images belong to 40 individuals with different facial expressions. Each individual category has 10
images each. Fig 2 shows few examples from the dataset. To learn a dictionary, a total of 3000 patches of
dimensions 8 X 8 are extracted from each image. Dictionary of dimensions 64 X 128 is learnt using β = 8. A
total number of 10 K-SVD iterations are used. The results are tabulated in table 1.The speedup for 140, 210 and
5. Scalable Image Classification Using Compression
DOI: 10.9790/0661-1804010105 www.iosrjournals.org 5 | Page
280 images are almost identical and hence they overlap in the graph. The graph of speedup and scaleup is given
in fig. 4(a) and 4(b).
Fig. 4 (a) Speedup graph for the method; (b) Scaleup graph for the method
V. Conclusion
Image similarity using dictionary learning techniques has attracted significant attention from
researchers lately. The dictionary and sparse coding are not yet fully optimized in terms of memory and speed.
Our experimental results are based on data parallelism and does not involve any sophisticated parallelism
approach for dictionary learning. The experimental results show that the speedup and scaleup is almost linear
when the system initialization communication cost is not taken into account. A considerable amount of speedup
is shown with the increase in the size of dataset. A perfect balance between the number of nodes and dataset size
is the key to an almost linear speedup and scaleup, without which a communication overhead occurs. Further
work on using other techniques for parallelization using map-reduce paradigm can be taken up. The
communication overhead, which occurs during the system startup can also be improvised using a different
memory architecture.
References
[1]. M.Li, X. Chen,X.Li,B.Ma, P.Vintayi, The similarity metric, IEEE Trans. Inf. Theory, 2004, 50:3250-3264
[2]. R.Calibrasi,P.Vintayi, Clustering by compression, IEEE Trans.Inf.Theory, 2005, 51:1523-1545
[3]. Cebrian Manuel,Alofonseca Manuel, Ortega Alfonso, Common pitfalls using the normalized compression distance : what to watch
out for in a compressor?, Communications in Information & Systems, 2005, 5:367-384
[4]. Elad Michael, on the role of sparse and redundant representations in image processing, IEEE proceedings, 2010, 15: 3736-3745
[5]. K.P. Soman, R. Ramanathan, Digital signal and image processing – The sparse way, Isa publication, 2012.
[6]. Ori Bryt, Elad Michael, Compression of facial images using the K-SVD algorithm, Elsevier, 2008, 19:270-282
[7]. Qiang Qiu, Patel Vishal,Chellapa Rama, Information-theoretic dictionary learning for image classification, IEEE Transactions on
pattern analysis and machine learning, 2014, 36:2173-2184
[8]. Guha Tanaya, Ward Rabab K., Image similarity using sparse representation and compression distance, IEEE Transactions on
multimedia, 2014, 16: 980-987
[9]. Guha Tanaya, Ward Rabab K., Aboulnasr Tyseer, Image similarity measurement from sparse reconstruction errors, ICASSP, 2013,
1937-1941
[10]. Aharon Michal, Elad Michael,Bruckstein Alfred, K-SVD : An algorithm for designing overcomplete dictionaries for sparse
representations, IEEE Transactions on signal processing, 2006, 54:4311-4322
[11]. Vitányi Paul, M Balbach, F. J., Cilibrasi, R. L. & Li, M., Normalized information distance In Information theory and statistical
learning, Springer, 2009, 45-82.
[12]. Campana, Bilson JL, and Eamonn J. Keogh, A compression‐based distance measure for texture, Statistical Analysis and Data
Mining, 2010, 6: 381-398.
[13]. Zhang Qiang, Li Baoxin, Discriminative K-SVD for dictionary learning in face recognition, CVPR, 2010, 2691-2698
[14]. Yang Meng, Zhang Lei, Yang Jian, Zhang David, Robust sparse coding for face recognition, CVPR, 2011, 625-632
[15]. Pau G, Fuchs F, Sklyar O, Boutros M and Huber W, EBImage-an R package for image processing with applications to cellular
phenotypes, Bioinformatics, 2010 ,27:979-981
[16]. AT&T, AT&Tarchive. [Online] Available at: http://www.cl.cam.ac.uk/Research/DTG/attarchive:pub/data/att_faces.zip [Accessed
October 2015].