This document summarizes a research paper that proposes a parallel k-nearest neighbors (kNN) algorithm using OpenCL on a GPU architecture. The key points are:
1) kNN classification is computationally intensive, especially for large datasets, creating a need for parallelization.
2) The authors designed and implemented a parallel kNN algorithm using OpenCL to distribute distance computations across GPU cores.
3) Experimental results on UCI datasets showed the parallel kNN approach improved performance over a serial kNN implementation.
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and ArchitecturesMLAI2
Regularization and transfer learning are two popular techniques to enhance generalization on unseen data, which is a fundamental problem of machine learning. Regularization techniques are versatile, as they are task- and architecture-agnostic, but they do not exploit a large amount of data available. Transfer learning methods learn to transfer knowledge from one domain to another, but may not generalize across tasks and architectures, and may introduce new training cost for adapting to the target task. To bridge the gap between the two, we propose a transferable perturbation, MetaPerturb, which is meta-learned to improve generalization performance on unseen data. MetaPerturb is implemented as a set-based lightweight network that is agnostic to the size and the order of the input, which is shared across the layers. Then, we propose a meta-learning framework, to jointly train the perturbation function over heterogeneous tasks in parallel. As MetaPerturb is a set-function trained over diverse distributions across layers and tasks, it can generalize to heterogeneous tasks and architectures. We validate the efficacy and generality of MetaPerturb trained on a specific source domain and architecture, by applying it to the training of diverse neural architectures on heterogeneous target datasets against various regularizers and fine-tuning. The results show that the networks trained with MetaPerturb significantly outperform the baselines on most of the tasks and architectures, with a negligible increase in the parameter size and no hyperparameters to tune.
Robust foreground modelling to segment and detect multiple moving objects in ...IJECEIAES
Last decade has witnessed an ever increasing number of video surveillance installa- tions due to the rise of security concerns worldwide. With this comes the need for video analysis for fraud detection, crime investigation, traffic monitoring to name a few. For any kind of video analysis application, detection of moving objects in videos is a fundamental step. In this paper, an efficient foreground modelling method to segment multiple moving objects is implemented. Proposed method significantly reduces noise thereby accurately segmenting region of interest under dynamic conditions while handling occlusion to a large extent. Extensive performance analysis shows that the proposed method was found to give far better results when compared to the de facto standard as well as relatively new approaches used for moving object detection.
In recent machine learning community, there is a trend of constructing a linear logarithm version of
nonlinear version through the ‘kernel method’ for example kernel principal component analysis, kernel
fisher discriminant analysis, support Vector Machines (SVMs), and the current kernel clustering
algorithms. Typically, in unsupervised methods of clustering algorithms utilizing kernel method, a
nonlinear mapping is operated initially in order to map the data into a much higher space feature, and then
clustering is executed. A hitch of these kernel clustering algorithms is that the clustering prototype resides
in increased features specs of dimensions and therefore lack intuitive and clear descriptions without
utilizing added approximation of projection from the specs to the data as executed in the literature
presented. This paper aims to utilize the ‘kernel method’, a novel clustering algorithm, founded on the
conventional fuzzy clustering algorithm (FCM) is anticipated and known as kernel fuzzy c-means algorithm
(KFCM). This method embraces a novel kernel-induced metric in the space of data in order to interchange
the novel Euclidean matric norm in cluster prototype and fuzzy clustering algorithm still reside in the space
of data so that the results of clustering could be interpreted and reformulated in the spaces which are
original. This property is used for clustering incomplete data. Execution on supposed data illustrate that
KFCM has improved performance of clustering and stout as compare to other transformations of FCM for
clustering incomplete data.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
The objective of this paper is to present the hybrid approach for edge detection. Under this technique, edge
detection is performed in two phase. In first phase, Canny Algorithm is applied for image smoothing and in
second phase neural network is to detecting actual edges. Neural network is a wonderful tool for edge
detection. As it is a non-linear network with built-in thresholding capability. Neural Network can be trained
with back propagation technique using few training patterns but the most important and difficult part is to
identify the correct and proper training set.
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and ArchitecturesMLAI2
Regularization and transfer learning are two popular techniques to enhance generalization on unseen data, which is a fundamental problem of machine learning. Regularization techniques are versatile, as they are task- and architecture-agnostic, but they do not exploit a large amount of data available. Transfer learning methods learn to transfer knowledge from one domain to another, but may not generalize across tasks and architectures, and may introduce new training cost for adapting to the target task. To bridge the gap between the two, we propose a transferable perturbation, MetaPerturb, which is meta-learned to improve generalization performance on unseen data. MetaPerturb is implemented as a set-based lightweight network that is agnostic to the size and the order of the input, which is shared across the layers. Then, we propose a meta-learning framework, to jointly train the perturbation function over heterogeneous tasks in parallel. As MetaPerturb is a set-function trained over diverse distributions across layers and tasks, it can generalize to heterogeneous tasks and architectures. We validate the efficacy and generality of MetaPerturb trained on a specific source domain and architecture, by applying it to the training of diverse neural architectures on heterogeneous target datasets against various regularizers and fine-tuning. The results show that the networks trained with MetaPerturb significantly outperform the baselines on most of the tasks and architectures, with a negligible increase in the parameter size and no hyperparameters to tune.
Robust foreground modelling to segment and detect multiple moving objects in ...IJECEIAES
Last decade has witnessed an ever increasing number of video surveillance installa- tions due to the rise of security concerns worldwide. With this comes the need for video analysis for fraud detection, crime investigation, traffic monitoring to name a few. For any kind of video analysis application, detection of moving objects in videos is a fundamental step. In this paper, an efficient foreground modelling method to segment multiple moving objects is implemented. Proposed method significantly reduces noise thereby accurately segmenting region of interest under dynamic conditions while handling occlusion to a large extent. Extensive performance analysis shows that the proposed method was found to give far better results when compared to the de facto standard as well as relatively new approaches used for moving object detection.
In recent machine learning community, there is a trend of constructing a linear logarithm version of
nonlinear version through the ‘kernel method’ for example kernel principal component analysis, kernel
fisher discriminant analysis, support Vector Machines (SVMs), and the current kernel clustering
algorithms. Typically, in unsupervised methods of clustering algorithms utilizing kernel method, a
nonlinear mapping is operated initially in order to map the data into a much higher space feature, and then
clustering is executed. A hitch of these kernel clustering algorithms is that the clustering prototype resides
in increased features specs of dimensions and therefore lack intuitive and clear descriptions without
utilizing added approximation of projection from the specs to the data as executed in the literature
presented. This paper aims to utilize the ‘kernel method’, a novel clustering algorithm, founded on the
conventional fuzzy clustering algorithm (FCM) is anticipated and known as kernel fuzzy c-means algorithm
(KFCM). This method embraces a novel kernel-induced metric in the space of data in order to interchange
the novel Euclidean matric norm in cluster prototype and fuzzy clustering algorithm still reside in the space
of data so that the results of clustering could be interpreted and reformulated in the spaces which are
original. This property is used for clustering incomplete data. Execution on supposed data illustrate that
KFCM has improved performance of clustering and stout as compare to other transformations of FCM for
clustering incomplete data.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
The objective of this paper is to present the hybrid approach for edge detection. Under this technique, edge
detection is performed in two phase. In first phase, Canny Algorithm is applied for image smoothing and in
second phase neural network is to detecting actual edges. Neural network is a wonderful tool for edge
detection. As it is a non-linear network with built-in thresholding capability. Neural Network can be trained
with back propagation technique using few training patterns but the most important and difficult part is to
identify the correct and proper training set.
Clustering using kernel entropy principal component analysis and variable ker...IJECEIAES
Clustering as unsupervised learning method is the mission of dividing data objects into clusters with common characteristics. In the present paper, we introduce an enhanced technique of the existing EPCA data transformation method. Incorporating the kernel function into the EPCA, the input space can be mapped implicitly into a high-dimensional of feature space. Then, the Shannon’s entropy estimated via the inertia provided by the contribution of every mapped object in data is the key measure to determine the optimal extracted features space. Our proposed method performs very well the clustering algorithm of the fast search of clusters’ centers based on the local densities’ computing. Experimental results disclose that the approach is feasible and efficient on the performance query.
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETScsandit
The ability to mine and extract useful information automatically, from large datasets, is a
common concern for organizations (having large datasets), over the last few decades. Over the
internet, data is vastly increasing gradually and consequently the capacity to collect and store
very large data is significantly increasing.
Existing clustering algorithms are not always efficient and accurate in solving clustering
problems for large datasets.
However, the development of accurate and fast data classification algorithms for very large
scale datasets is still a challenge. In this paper, various algorithms and techniques especially,
approach using non-smooth optimization formulation of the clustering problem, are proposed
for solving the minimum sum-of-squares clustering problems in very large datasets. This
research also develops accurate and real time L2-DC algorithm based with the incremental
approach to solve the minimum
Welcome to International Journal of Engineering Research and Development (IJERD)IJERD Editor
call for paper 2012, hard copy of journal, research paper publishing, where to publish research paper,
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGijaia
Imbalanced data is a significant challenge in classification with machine learning algorithms. This is particularly important with log message data as negative logs are sparse so this data is typically imbalanced. In this paper, a model to generate text log messages is proposed which employs a SeqGAN network. An Autoencoder is used for feature extraction and anomaly detection is done using a GRU network. The proposed model is evaluated with three imbalanced log data sets, namely BGL, OpenStack, and Thunderbird. Results are presented which show that appropriate oversampling and data balancing
improves anomaly detection accuracy.
Design a New Image Encryption using Fuzzy Integral Permutation with Coupled C...IJORCS
This article introduces a novel image encryption algorithm based on DNA addition combining and coupled two-dimensional piecewise nonlinear chaotic map. This algorithm consists of two parts. In the first part of the algorithm, a DNA sequence matrix is obtained by encoding each color component, and is divided into some equal blocks and then the generated sequence of Sugeno integral fuzzy and the DNA sequence addition operation is used to add these blocks. Next, the DNA sequence matrix from the previous step is decoded and the complement operation to the result of the added matrix is performed by using Sugeno fuzzy integral. In the second part of the algorithm, the three modified color components are encrypted in a coupling fashion in such a way to strengthen the cryptosystem security. It is observed that the histogram, the correlation and avalanche criterion, can satisfy security and performance requirements (Avalanche criterion > 0.49916283). The experimental results obtained for the CVG-UGR image databases reveal the fact that the proposed algorithm is suitable for practical use to protect the security of digital image information over the Internet.
An Efficient Frame Embedding Using Haar Wavelet Coefficients And Orthogonal C...IJERA Editor
Digital media, applications, copyright defense, and multimedia security have become a vital aspect of our daily life. Digital watermarking is a technology used for the copyright security of digital applications. In this work we have dealt with a process able to mark digital pictures with a visible and semi invisible hided information, called watermark. This process may be the basis of a complete copyright protection system. Watermarking is implemented here using Haar Wavelet Coefficients and Principal Component analysis. Experimental results show high imperceptibility where there is no noticeable difference between the watermarked video frames and the original frame in case of invisible watermarking, vice-versa for semi visible implementation. The watermark is embedded in lower frequency band of Wavelet Transformed cover image. The combination of the two transform algorithm has been found to improve performance of the watermark algorithm. The robustness of the watermarking scheme is analyzed by means of two distinct performance measures viz. Peak Signal to Noise Ratio (PSNR) and Normalized Coefficient (NC).
The paper proposes a new image encryption scheme based on chaotic encryption. It provides a fast
encryption algorithm based on coupled chaotic map. The image is encrypted using a pseudorandom key
stream generator. The image is partially encrypted by selecting most important components of image. To
obtain most important components of an image, discrete wavelet transform is applied.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Performance Optimization of Clustering On GPUijsrd.com
In today's digital world, Data sets are increasing exponentially. Statistical analysis using clustering in various scientific and engineering applications become very challenging issue for such large data set. Clustering on huge data set and its performance are two major factors demand for optimization. Parallelization is well-known approach to optimize performance. It has been observed from recent research work that GPU based parallelization help to achieve high degree of performance. Hence, this thesis focuses on optimizing hierarchical clustering algorithms using parallelization. It covers implementation of optimized algorithm on various parallel environments using Open MP on multi-core architecture and using CUDA on may-core architecture.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Appraisal of multilevel car parking facility at kg road cbd area, bengaluru cityeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Clustering using kernel entropy principal component analysis and variable ker...IJECEIAES
Clustering as unsupervised learning method is the mission of dividing data objects into clusters with common characteristics. In the present paper, we introduce an enhanced technique of the existing EPCA data transformation method. Incorporating the kernel function into the EPCA, the input space can be mapped implicitly into a high-dimensional of feature space. Then, the Shannon’s entropy estimated via the inertia provided by the contribution of every mapped object in data is the key measure to determine the optimal extracted features space. Our proposed method performs very well the clustering algorithm of the fast search of clusters’ centers based on the local densities’ computing. Experimental results disclose that the approach is feasible and efficient on the performance query.
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETScsandit
The ability to mine and extract useful information automatically, from large datasets, is a
common concern for organizations (having large datasets), over the last few decades. Over the
internet, data is vastly increasing gradually and consequently the capacity to collect and store
very large data is significantly increasing.
Existing clustering algorithms are not always efficient and accurate in solving clustering
problems for large datasets.
However, the development of accurate and fast data classification algorithms for very large
scale datasets is still a challenge. In this paper, various algorithms and techniques especially,
approach using non-smooth optimization formulation of the clustering problem, are proposed
for solving the minimum sum-of-squares clustering problems in very large datasets. This
research also develops accurate and real time L2-DC algorithm based with the incremental
approach to solve the minimum
Welcome to International Journal of Engineering Research and Development (IJERD)IJERD Editor
call for paper 2012, hard copy of journal, research paper publishing, where to publish research paper,
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGijaia
Imbalanced data is a significant challenge in classification with machine learning algorithms. This is particularly important with log message data as negative logs are sparse so this data is typically imbalanced. In this paper, a model to generate text log messages is proposed which employs a SeqGAN network. An Autoencoder is used for feature extraction and anomaly detection is done using a GRU network. The proposed model is evaluated with three imbalanced log data sets, namely BGL, OpenStack, and Thunderbird. Results are presented which show that appropriate oversampling and data balancing
improves anomaly detection accuracy.
Design a New Image Encryption using Fuzzy Integral Permutation with Coupled C...IJORCS
This article introduces a novel image encryption algorithm based on DNA addition combining and coupled two-dimensional piecewise nonlinear chaotic map. This algorithm consists of two parts. In the first part of the algorithm, a DNA sequence matrix is obtained by encoding each color component, and is divided into some equal blocks and then the generated sequence of Sugeno integral fuzzy and the DNA sequence addition operation is used to add these blocks. Next, the DNA sequence matrix from the previous step is decoded and the complement operation to the result of the added matrix is performed by using Sugeno fuzzy integral. In the second part of the algorithm, the three modified color components are encrypted in a coupling fashion in such a way to strengthen the cryptosystem security. It is observed that the histogram, the correlation and avalanche criterion, can satisfy security and performance requirements (Avalanche criterion > 0.49916283). The experimental results obtained for the CVG-UGR image databases reveal the fact that the proposed algorithm is suitable for practical use to protect the security of digital image information over the Internet.
An Efficient Frame Embedding Using Haar Wavelet Coefficients And Orthogonal C...IJERA Editor
Digital media, applications, copyright defense, and multimedia security have become a vital aspect of our daily life. Digital watermarking is a technology used for the copyright security of digital applications. In this work we have dealt with a process able to mark digital pictures with a visible and semi invisible hided information, called watermark. This process may be the basis of a complete copyright protection system. Watermarking is implemented here using Haar Wavelet Coefficients and Principal Component analysis. Experimental results show high imperceptibility where there is no noticeable difference between the watermarked video frames and the original frame in case of invisible watermarking, vice-versa for semi visible implementation. The watermark is embedded in lower frequency band of Wavelet Transformed cover image. The combination of the two transform algorithm has been found to improve performance of the watermark algorithm. The robustness of the watermarking scheme is analyzed by means of two distinct performance measures viz. Peak Signal to Noise Ratio (PSNR) and Normalized Coefficient (NC).
The paper proposes a new image encryption scheme based on chaotic encryption. It provides a fast
encryption algorithm based on coupled chaotic map. The image is encrypted using a pseudorandom key
stream generator. The image is partially encrypted by selecting most important components of image. To
obtain most important components of an image, discrete wavelet transform is applied.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Performance Optimization of Clustering On GPUijsrd.com
In today's digital world, Data sets are increasing exponentially. Statistical analysis using clustering in various scientific and engineering applications become very challenging issue for such large data set. Clustering on huge data set and its performance are two major factors demand for optimization. Parallelization is well-known approach to optimize performance. It has been observed from recent research work that GPU based parallelization help to achieve high degree of performance. Hence, this thesis focuses on optimizing hierarchical clustering algorithms using parallelization. It covers implementation of optimized algorithm on various parallel environments using Open MP on multi-core architecture and using CUDA on may-core architecture.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Appraisal of multilevel car parking facility at kg road cbd area, bengaluru cityeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Investigation on effective thermal conductivity of foams using transient plan...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Study on transmission energy losses and finding the hazards using what if ana...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Comparative assessment of noise levels in various laboratories and constructi...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A novel predicate for active region merging in automatic image segmentationeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Directly coupled microstrip array antennas for wideband applicationeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Performance improvement of bottleneck link in red vegas over heterogeneous ne...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Experimental study of Data clustering using k- Means and modified algorithmsIJDKP
The k- Means clustering algorithm is an old algorithm that has been intensely researched owing to its ease
and simplicity of implementation. Clustering algorithm has a broad attraction and usefulness in
exploratory data analysis. This paper presents results of the experimental study of different approaches to
k- Means clustering, thereby comparing results on different datasets using Original k-Means and other
modified algorithms implemented using MATLAB R2009b. The results are calculated on some performance
measures such as no. of iterations, no. of points misclassified, accuracy, Silhouette validity index and
execution time
Scalable Rough C-Means clustering using Firefly algorithm..................................................................1
Abhilash Namdev and B.K. Tripathy
Significance of Embedded Systems to IoT................................................................................................. 15
P. R. S. M. Lakshmi, P. Lakshmi Narayanamma and K. Santhi Sri
Cognitive Abilities, Information Literacy Knowledge and Retrieval Skills of Undergraduates: A
Comparison of Public and Private Universities in Nigeria ........................................................................ 24
Janet O. Adekannbi and Testimony Morenike Oluwayinka
Risk Assessment in Constructing Horseshoe Vault Tunnels using Fuzzy Technique................................ 48
Erfan Shafaghat and Mostafa Yousefi Rad
Evaluating the Adoption of Deductive Database Technology in Augmenting Criminal Intelligence in
Zimbabwe: Case of Zimbabwe Republic Police......................................................................................... 68
Mahlangu Gilbert, Furusa Samuel Simbarashe, Chikonye Musafare and Mugoniwa Beauty
Analysis of Petrol Pumps Reachability in Anand District of Gujarat ....................................................... 77
Nidhi Arora
Available transfer capability computations in the indian southern e.h.v power...eSAT Journals
Abstract This paper presents three methods for computing the available transfer capability (ATC). One method is the conventional method known as continuation repeated power flow (CRPF) and other two are the intelligent techniques known as radial basis function neural network (RBFNN), basic adaptive neuro fuzzy inference system (ANFIS). In these two intelligent techniques, the basic ANFIS works with a multiple input single output (MISO) and it is modified as multiple input multiple output (MIMO) ANFIS using the proposed MIMO ANFIS. The main intension of this proposed method is to utilise the significant features of ANFIS with respect to the multiple outputs, as the basic ANFIS has been proved as the best intelligent techniques in the modelling of any application, but it has a disadvantage of single output and this drawback will be overcome using the proposed MIMO ANFIS. In this paper, the latest Indian southern region extra high voltage (SREHV) 72-bus system considered as test system to obtain the ATC computations using three methods. The ATC computations at desired buses are obtained with and without contingencies and compared the conventional ATC computations with the intelligent techniques. The obtained results are scrupulously verified with different test patterns and observed that the accuracy of proposed method is proved as the best as compared to the other methods for computing ATC. In this way, this paper shows a better way to compute ATC for the different open power market system Keywords— Available Transfer Capability, Total Transfer Capability, Open Power Markets, Repeated Power Flow, Radial Basis Function Neural Networks, Adaptive Neuro Fuzzy Inference System etc.
Stochastic Computing Correlation Utilization in Convolutional Neural Network ...TELKOMNIKA JOURNAL
In recent years, many applications have been implemented in embedded systems and mobile Internet of Things (IoT) devices that typically have constrained resources, smaller power budget, and exhibit "smartness" or intelligence. To implement computation-intensive and resource-hungry Convolutional Neural Network (CNN) in this class of devices, many research groups have developed specialized parallel accelerators using Graphical Processing Units (GPU), Field-Programmable Gate Arrays (FPGA), or Application-Specific Integrated Circuits (ASIC). An alternative computing paradigm called Stochastic Computing (SC) can implement CNN with low hardware footprint and power consumption. To enable building more efficient SC CNN, this work incorporates the CNN basic functions in SC that exploit correlation, share Random Number Generators (RNG), and is more robust to rounding error. Experimental results show our proposed solution provides significant savings in hardware footprint and increased accuracy for the SC CNN basic functions circuits compared to previous work.
K-Means clustering uses an iterative procedure which is very much sensitive and dependent upon the initial centroids. The initial centroids in the k-means clustering are chosen randomly, and hence the clustering also changes with respect to the initial centroids. This paper tries to overcome this problem of random selection of centroids and hence change of clusters with a premeditated selection of initial centroids. We have used the iris, abalone and wine data sets to demonstrate that the proposed method of finding the initial centroids and using the centroids in k-means algorithm improves the clustering performance. The clustering also remains the same in every run as the initial centroids are not randomly selected but through premeditated method.
A PSO-Based Subtractive Data Clustering AlgorithmIJORCS
There is a tremendous proliferation in the amount of information available on the largest shared information source, the World Wide Web. Fast and high-quality clustering algorithms play an important role in helping users to effectively navigate, summarize, and organize the information. Recent studies have shown that partitional clustering algorithms such as the k-means algorithm are the most popular algorithms for clustering large datasets. The major problem with partitional clustering algorithms is that they are sensitive to the selection of the initial partitions and are prone to premature converge to local optima. Subtractive clustering is a fast, one-pass algorithm for estimating the number of clusters and cluster centers for any given set of data. The cluster estimates can be used to initialize iterative optimization-based clustering methods and model identification methods. In this paper, we present a hybrid Particle Swarm Optimization, Subtractive + (PSO) clustering algorithm that performs fast clustering. For comparison purpose, we applied the Subtractive + (PSO) clustering algorithm, PSO, and the Subtractive clustering algorithms on three different datasets. The results illustrate that the Subtractive + (PSO) clustering algorithm can generate the most compact clustering results as compared to other algorithms.
Improving K-NN Internet Traffic Classification Using Clustering and Principle...journalBEEI
K-Nearest Neighbour (K-NN) is one of the popular classification algorithm, in this research K-NN use to classify internet traffic, the K-NN is appropriate for huge amounts of data and have more accurate classification, K-NN algorithm has a disadvantages in computation process because K-NN algorithm calculate the distance of all existing data in dataset. Clustering is one of the solution to conquer the K-NN weaknesses, clustering process should be done before the K-NN classification process, the clustering process does not need high computing time to conqest the data which have same characteristic, Fuzzy C-Mean is the clustering algorithm used in this research. The Fuzzy C-Mean algorithm no need to determine the first number of clusters to be formed, clusters that form on this algorithm will be formed naturally based datasets be entered. The Fuzzy C-Mean has weakness in clustering results obtained are frequently not same even though the input of dataset was same because the initial dataset that of the Fuzzy C-Mean is less optimal, to optimize the initial datasets needs feature selection algorithm. Feature selection is a method to produce an optimum initial dataset Fuzzy C-Means. Feature selection algorithm in this research is Principal Component Analysis (PCA). PCA can reduce non significant attribute or feature to create optimal dataset and can improve performance for clustering and classification algorithm. The resultsof this research is the combination method of classification, clustering and feature selection of internet traffic dataset was successfully modeled internet traffic classification method that higher accuracy and faster performance.
A New Method for Figuring the Number of Hidden Layer Nodes in BP Algorithmrahulmonikasharma
In the field of artificial neural network, BP neural network is a multi-layer feed-forward neural network. Because it is difficult to figure the number of hidden layer nodes in a BP neural network, the theoretical basis and the existing methods for BP network hidden layer nodes are studied. Then based on traditional empirical formulas, we propose a new approach to rapidly figure the quantity of hidden layer nodes in two-layer network. That is, with the assistance of experience formulas, the horizon of unit number in hidden layer can be confirmed and its optimal value will be found in this horizon. Finally, a new formula for figuring the quantity of hidden layer codes is obtained through fitting input dimension, output dimension and the optimal value of hidden layer codes. Under some given input dimension and output dimension, efficiency and precision of BP algorithm may be improved by applying the proposed formula.
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...IAEME Publication
This paper presents an approach based on applying an aggregated predictor formed by multiple versions of a multilayer neural network with a back-propagation optimization algorithm for helping the engineer to get a list of the most appropriate well-test interpretation models for a given set of pressure/ production data. The proposed method consists of three stages: (1) data decorrelation through principal component analysis to reduce the covariance between the variables and the dimension of the input layer in the artificial neural network, (2) bootstrap replicates of the learning set where the data is repeatedly sampled with a random split of the data into train sets and using these as new learning sets, and (3) automatic reservoir model identification through aggregated predictor formed by a plurality vote when predicting a new class. This method is described in detail to ensure successful replication of results. The required training and test dataset were generated by using analytical solution models. In our case, there were used 600 samples: 300 for training, 100 for cross-validation, and 200 for testing. Different network structures were tested during this study to arrive at optimum network design. We notice that the single net methodology always brings about confusion in selecting the correct model even though the training results for the constructed networks are close to 1. We notice also that the principal component analysis is an effective strategy in reducing the number of input features, simplifying the network structure, and lowering the training time of the ANN. The results obtained show that the proposed model provides better performance when predicting new data with a coefficient of correlation approximately equal to 95% Compared to a previous approach 80%, the combination of the PCA and ANN is more stable and determine the more accurate results with lesser computational complexity than was feasible previously. Clearly, the aggregated predictor is more stable and shows less bad classes compared to the previous approach.
k-Means is a rather simple but well known algorithms for grouping objects, clustering. Again all objects need to be represented as a set of numerical features. In addition the user has to specify the number of groups (referred to as k) he wishes to identify. Each object can be thought of as being represented by some feature vector in an n dimensional space, n being the number of all features used to describe the objects to cluster. The algorithm then randomly chooses k points in that vector space, these point serve as the initial centers of the clusters. Afterwards all objects are each assigned to center they are closest to. Usually the distance measure is chosen by the user and determined by the learning task. After that, for each cluster a new center is computed by averaging the feature vectors of all objects assigned to it. The process of assigning objects and recomputing centers is repeated until the process converges. The algorithm can be proven to converge after a finite number of iterations. Several tweaks concerning distance measure, initial center choice and computation of new average centers have been explored, as well as the estimation of the number of clusters k. Yet the main principle always remains the same. In this project we will discuss about K-means clustering algorithm, implementation and its application to the problem of unsupervised learning
Hierarchical Digital Twin of a Naval Power SystemKerry Sado
A hierarchical digital twin of a Naval DC power system has been developed and experimentally verified. Similar to other state-of-the-art digital twins, this technology creates a digital replica of the physical system executed in real-time or faster, which can modify hardware controls. However, its advantage stems from distributing computational efforts by utilizing a hierarchical structure composed of lower-level digital twin blocks and a higher-level system digital twin. Each digital twin block is associated with a physical subsystem of the hardware and communicates with a singular system digital twin, which creates a system-level response. By extracting information from each level of the hierarchy, power system controls of the hardware were reconfigured autonomously. This hierarchical digital twin development offers several advantages over other digital twins, particularly in the field of naval power systems. The hierarchical structure allows for greater computational efficiency and scalability while the ability to autonomously reconfigure hardware controls offers increased flexibility and responsiveness. The hierarchical decomposition and models utilized were well aligned with the physical twin, as indicated by the maximum deviations between the developed digital twin hierarchy and the hardware.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Water scarcity is the lack of fresh water resources to meet the standard water demand. There are two type of water scarcity. One is physical. The other is economic water scarcity.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptx
Parallel k nn on gpu architecture using opencl
1. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 10 | Oct-2014, Available @ http://www.ijret.org 367
PARALLEL kNN ON GPU ARCHITECTURE USING OpenCL
V.B.Nikam1
, B.B.Meshram2
1
Associate Professor, Department of Computer Engineering and Information Technology, Jijabai Technological
Institute, Matunga, Mumbai, Maharashtra, India
2
Professor, Department of Computer Engineering and Information Technology, Veermata Jijabai Technological
Institute, Matunga, Mumbai, Maharashtra, India
Abstract
In data mining applications, one of the useful algorithms for classification is the kNN algorithm. The kNN search has a wide
usage in many research and industrial domains like 3-dimensional object rendering, content-based image retrieval, statistics,
biology (gene classification), etc. In spite of some improvements in the last decades, the computation time required by the kNN
search remains the bottleneck for kNN classification, especially in high dimensional spaces. This bottleneck has created the
necessity of the parallel kNN on commodity hardware. GPU and OpenCL architecture are the low cost high performance
solutions for parallelising the kNN classifier. In regard to this, we have designed, implemented our proposed parallel kNN model
to improve upon performance bottleneck issue of kNN algorithm.
In this paper, we have proposed parallel kNN algorithm on GPU and OpenCL framework. In our approach, we distributed the
distance computations of the data points among all GPU cores. Multiple threads invoked for each GPU core. We have
implemented and tested our parallel kNN implementation on UCI datasets. The experimental results show that the speedup of the
KNN algorithm is improved over the serial performance.
Keywords: kNN, GPU, CPU, Parallel Computing, Data Mining, Clustering Algorithm.
--------------------------------------------------------------------***----------------------------------------------------------------------
1. INTRODUCTION
In data mining, classification is the problem of identifying
set of categories/sub-populations from new observations;
based on a training dataset containing observations/instances
whose category membership known in advance. The data
mining algorithm performs classification is called classifier.
Classifier is a supervised learning category. In supervised
learning, input set of observations has an effect on output set
of observations; the model includes the mediating variables
between the input and output variables. kNN uses a
continuous data for building classification model. Being the
capability of handling large datasets for solving the big
problems, kNN has been widely accepted in research and
industry applications. Though, kNN has been accepted
widely; due to the large datasets, performance became the
performance bottleneck for the algorithm. kNN classifier is
an instance based learning (IBL) which follows case based
reasoning (CBR) approach to solve new problems. This
learning approach is a lazy learning. kNN classifier is the
most widely used classifier in the data mining domain, and
has been accepted for wide range and volume of datasets. A
new instance is classified as the most frequent class of its
‘K’ nearest neighbours.
OpenCL is a generic many core computing programming
framework used for parallelizing applications on the GPU’s
parallel architecture. OpenCL and GPU is most preferred for
parallelizing kNN. It is observed that, kNN is
embarrassingly parallel in nature and a good candidate for
parallelism. In this paper, we have explored the design and
experimental results of our parallel kNN on GPU and
OpenCL architecture. We have proposed, implemented and
tested OpenCL based parallel kNN algorithm on GPU
platform. We observed that, our proposed parallel kNN on
OpenCL & GPU platform performs well over the sequential
performance of kNN. In addition to the performance, the
accuracy of our proposed kNN algorithm is achieved to the
level of satisfaction.
This paper is organised as below: Section2 discusses the
literature survey on kNN classifier, related work in parallel
kNN algorithm, OpenCL and GPU technology. Section3
describes our methodology on OpenCL framework for
parallel kNN algorithm. Section4 discusses on result
analysis. We have observed the results for scaleup on
performance. Section5 is the conclusion of our work and
the paper.
2. LITERATURE SURVEY
The basic process of sequential kNN algorithm is as follows:
First, the data pre-processing phase is to initialize the
labelled dimensional training dataset as well as the test
dataset to be classified. Second, select one test point in the
test dataset and calculate the distances between it and each
point in the train dataset. The next phase is to sort the results
of distances computed, and find out ‘k’ smallest results. The
fourth step is to determine the class label of the test point by
the election result of ‘k’ points. Thus, selecting another
point in the test dataset and repeat the process to classify the
test dataset[5].
2. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 10 | Oct-2014, Available @ http://www.ijret.org 368
The pseudo-code for the sequential kNN algorithm is as
follows,
Algorithm: kNN
Input: Dataset, D = {(x1,c1), . . . , (xN,cN)},
Input query t = (x1, . . . ,xn), k- number of
neighbour.
Output: class ‘c’ to be identified for new instance of
dataset ‘t’
Begin
For each labelled instance (xi, ci)
calculate d(xi, x)
Order d(xi,x) from lowest to highest, (i = 1, . . . ,N)
Select ‘K’ nearest instances to x:
Assign to x the most frequent class in
End
If we have 4 test cases and 12 training points, then in
traditional algorithm we visit each training point 4 times and
will be perform total 48 serial operations. Thus, for ‘m’ test
cases, and ‘n’ training points there would be O(mn) time
complexity. The time complexity is very high if test cases
and training points become very large in volume. This has
created the need to parallelize operations for kNN algorithm.
The compute intensive operational part of kNN algorithm
taken up for the parallel and simultaneous operations, can
lead to reasonable reduction in compute time. The optimal
value for ‘k’ identified in the range of 3 to 10. In general,
large ‘k’ value is more precise as it reduces the overall
noise.
kNN has been used in statistical methodologies and pattern
recognition since beginning. ‘k’ nearest neighbours
measured by a distance function. Here we have used
Euclidian distance. The Euclidian distance measured
between two points, ‘a’ and ‘b’, is shown below in equation
(1).
d(a,b) = .........................................(1)
In addition to Euclidian distance, there are Manhattan
distance, Minkowski distance are also used for measuring
distances. These distance measures are only used for
continuous data variables, however, in case of categorical
variables the Hamming distance must be used. If there is
continuous and categorical variables exist in the dataset,
standardization of the numerical variables between 0 and 1
is done to achieve same type data values for the entire
dataset. This is also called as normalization of dataset.
2.1 OpenCL Framework
Due to tremendous computing power, General Purpose
computing on GPU (GPGPU) has been widely adopted by
the developers and researchers [7]. Khronos group has
proposed OpenCL (Open Computing Language) framework
for programming many core architecture [8]. OpenCL is an
open industry standard for general-purpose parallel
programming. OpenCL device ‘kernels’ contains
computation to be carried out in parallel. In order to achieve
parallel computation, many work-items, work groups, etc
the components of OpenCL data parallel execution model,
as shown in the 0, spawns on multiple data items [9].
Fig 1: Data parallel execution using OpenCL
The data parallelisation task happens through the work items
that executes in work-groups as shown in 0 For parallel
computing and performance, OpenCL memory hierarchy
has four types of device memories viz private, global,
constant and local memories. Efficient use of each of them
need to be done for high performance achievement [7][8].
2.2 Related Work on Parallel kNN
kNN is widely used classification algorithm and very
parallel friendly because of number of independent
operations. It is required to scan all the objects for any new
entry. It is known for lazy learning as all the time dataset
needs to be scanned. When training and test dataset size is
large, the speed will be quite slow which makes it the best
candidate to be processed in parallel. S. Liang , C. Wang, Y.
Liu and L. Jian (2009) have focused on parallelizing sorting
and distance calculation process due to high compute
intensiveness. Distances between the current object and all
the data set objects are calculated in parallel. They made two
kernels for the same. Sorting Kernel will find the rank of
particular object in data set.[1]. Q. Kuang and L. Zhao
(2009) proposed parallel scheme for processing testing data
set and training data set by dividing computation of result
data set. During distance calculation, result data-set is
divided into tiles and each tiles of width T is processed by
an independent block in parallel. Each block contains TxT
thread each of which calculates the single distance[2].
Garcia V. Debreuve E. Nielsen F. and Barlaud M. (2010)
suggested CUDA based and CUBLAS based approaches for
the parallelism of kNN. The approach except sorting is
almost similar to S. Liang et al[1] approach. CUBLAS
implementation proved high speed up [3] for performance.
G. Nolan proposed improved bubble sort on CUDA for kNN
process. they stated iterative version of bitonic sort used for
parallelisation[4]. A. Arefin, C. Riveros, R. Berretta and P.
Moscato (2011) proposed fast and scalable kNN algorithm
on GPU for very large data-sets. In their strategy, they
divided reluctant distance matrix into the chunks of size,
which can fit into the GPU’s onboard memory. This
3. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 10 | Oct-2014, Available @ http://www.ijret.org 369
approach is independent of the size of the distance matrix
and overall data size[5]. A. Arefin, et. al. (2012) proposed
graph based agglomerative clustering method based on kNN
graph and Boruvka’s algorithm on GPU[6] .
3. PROPOSED METHOLOLOGY
In our proposed parallel kNN algorithm, we have proposed
the methodology on OpenCL framework. Our model will
performs best on high-end GPU platform, also, the model
responds to the multi core commodity architecture. In our
parallel kNN model, we compute distance between each
training case and all test cases concurrently. Hence,
computes all distances in parallel in a step. We extended this
concept to calculate distance between all training cases and
all testing cases in parallel using many cores. These
computations we have taken up on many core GPU
platform, and developed kernels in OpenCL to compute the
task in parallel.
Finding distance ‘d’ for finding nearest neighbour is the
crucial and compute intensive task in kNN classifier. We
have ported the ‘distance’ computation on GPU to perform
the parallel operation, which has resulted to considerable
improvement in kNN performance. The major two kernels
work for parallelising the computation on multiple threads
on GPU, 1.Distance computation to find the distance
between the individual attributes points and 2.Sum up the
partial computed distance to find the total sum to get the
distance between the points for finding the nearest
neighbouring points among the available.
Fig 2: Proposed Parallel kNN flowchart
The compute intensive step in kNN is distance
computations, which is done on GPU where as others are
performed on CPU.
The pseudo code for our parallel kNN algorithm is as
follows,
Algorithm: parkNN
Input: Dataset, D = {(a1,c1), . . . , (xn,cn)},
Input query, t = (a1, . . . ,xn),
k- number of neighbour.
Output: Class ‘c’ to be identified for new instance of
dataset‘t’
Begin
On CPU:
1. Load data set to global memory
2. Decide ‘k’ for the cluster definition/generation
3. Transfer dataset to GPU memory
On GPU:
4. For all labelled instance (ai, ci), compute d(ai, a)
On CPU:
5. Sort d(ai,a) from lowest to highest, ( i = 1, . . . ,n)
6. Select ‘k’ nearest instances to x:
7. Assign to x the most frequent class in
End
Distributing difference computations on many core GPU
performs computations faster due to parallelising the
difference computations. The following OpenCL kernels
demonstrate parallel computation on GPU cores
Kernel for Parallel difference computation
__kernel void diff(__global const float * query, __global
const float *data,
__global const int *nVal, __global float * diff)
{ int id = get_global_id(0);
int indexCent = id % ( nVal[0] * nVal[1] );
int p = id % nVal[0];
int q = (( id / (nVal[0] * nVal[1]) ) % nVal[2] ) * nVal[0];
diff[ id ] = (query[indexCent] - data[ p + q ]) *
(query[indexCent] - data[ p + q ]);
}
nVal contains following parameters,
0 -> #Attributes, 1 -> #Query , 2 -> #Data Points,
# Workers = (#Attributes * #Query * #Data Points)
Kernel for Parallel Sum Computations
__kernel void Sum(__global const float * data, __global
float *sum, __global const int *nVal)
{ int id = get_global_id(0);
int index = id + (( id / nVal[2] ) * nVal[1] );
if( (id % nVal[2]) == (nVal[0] - nVal[2]))
sum[id] = data[index];
else
sum[id] = data[index] + data[(index + nVal[2])];
}
nVal contains following parameters,
0 -> Data Point Length (DPL), 1 -> DPL / 2, 2 -> Offset ,
Offset = DPL/2 + DPL%2,
# workers = offset * # Data Points
The parallel computation of the difference computations in
two points, A & B, is illustrated in 0, as shown below. The
OpenCL kernels written for parallel performance on GPU
4. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 10 | Oct-2014, Available @ http://www.ijret.org 370
device, and are achieving computation methodology as
shown in 0.
Fig 3: Distance calculation between two points in Parallel
Parallel sum kernel as described in kernel ‘sum’. Multi-step
execution (partial data) for parallel sum calculation
generated partial results on each level which further used as
inputs to next level of parallel computations.
Illustrative Example: We illustrate working of our
proposed parallel kNN algorithm with the UCI Balloon
dataset, as described in Table1. The attribute values of
dataset are transformed into bit-mapped represented
datasets. The sample dataset split into training dataset and
test dataset with the ratio of 70:30 respectively. For UCI
balloon dataset 12 records are taken as training dataset, and
4 instances are taken as test datasets. The training dataset is
used to build model, however test dataset used to test the
built model for records to classify the query instance. The
test data set is shown in Table1, and the training dataset is
shown in Table2, as below. The balloon dataset has 4
attributes, 16 records, and is classified in two classes.
Table: 1 Balloons Dataset. (TEST)
Sr. No. Color Size Act Age Class
1 1 2 1 1 1
2 1 2 1 2 2
3 2 1 1 1 1
4 2 1 1 2 2
Table: 2 Balloons Dataset (TRAINING): Bitmap
representation
Sr. No. Color Size Act Age Class
1 1 1 1 1 1
2 1 1 1 2 2
3 1 1 2 1 2
4 1 1 2 2 2
5 1 2 2 1 2
6 1 2 2 2 2
7 2 1 2 1 2
8 2 1 2 2 2
9 2 2 1 1 1
10 2 2 1 2 2
11 2 2 2 1 2
12 2 2 2 2 2
Our proposed parallel kNN algorithm takes into account the
same inputs and produces the same output as sequential
algorithm, but the kNN operations performed are parallel for
performance achievement.
The following steps illustrate the parallel operations of kNN
algorithm.
CPU:
1. Load data set to global memory
2. Decide ‘k’ for the cluster definition/generation
3. Transfer dataset to GPU memory
GPU:
4. For all labelled instance (ai,ci), compute d(ai,a)
CPU:
5. Apply input query on build model for class
prediction,
a. Sort d(ai,a) from lowest to highest,(i=1,...,n
)
b. Select ‘k’ nearest instances to x:
c. Assign to ‘x’ the most frequent class in
The algorithmic steps are illustrated a bit detail as below
Step1: Load Data
Load data from CPU memory to GPU global memory.
The data is transformed from actual dataset, to bit
represented dataset at CPU. The ‘bit’ represented
dataset is then transferred GPU memory.
Step2 : Define ‘k’, Let say k=3,
Step 3: Transform dataset to GPU memory.
// for distance computations, we used squared
Euclidean distance method.
Step 4: Make Data vector ‘D’ and Query vector ‘Q’
Vector D (Training Data) contents:
Vector Q (Query Vector) contents:
Computing an Euclidian distance in parallel using openCL
kernel __diff( )
5. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 10 | Oct-2014, Available @ http://www.ijret.org 371
Here, the parallel model is ready which can be used to
predict the input query class. The model is tested in step5.
This we have explained in the following four test cases.
Step 1: Find Class for each Test case
Test Case 1
Distance list is <1, 2, 2, 3, 1, 2, 3, 4, 1, 2, 2, 3>, and their
corresponding
Class vector is <1, 2, 2, 2, 2, 2, 2, 2, 1, 2, 2, 2 >
Pairing the <Distance vector, Class vector> we get,
<(1,1), (2,2), (2,2), (3,2), (1,2), (2,2), (3,2), (4,2),
(1,1),(2,2),(2,2),(3,2) >
After sorting paring vector in ascending order distance wise,
we get
<(1,1), (1,2), (1,1), (2,2), (2,2), (2,2), (2,2), (2,2), (3,2),
(3,2), (3,2), (4,2)>;
Selecting first ‘k’ samples we get training samples with
corresponding classes as <1, 2, 1>, for k=3.
Thus using majority,we can classify this test case in class
1”.
Test Case 2
For distance list <2, 1, 3, 2, 2, 1, 4, 3, 2, 1, 3, 2>,
After sorting and selecting first ‘k’ samples, we get training
samples as <2, 5, 10> with corresponding classes as
<2,2,2>.
Thus using majority, we can classify this test case in class
“2”.
Test Case 3
Distance list is <1, 2, 2, 3, 3, 4, 1, 2, 1, 2, 2, 3>
After sorting and selecting first k samples we get training
samples as <1, 7, 9> with corresponding classes as <1, 2, 1>
Thus using majority we can classify this test case in
class“1”.
Test Case 4
Distance list is <2, 1, 3, 2, 4, 3, 2, 1, 2, 1, 3, 2>
After sorting and selecting first k samples we get training
samples as <2, 8, 10> with corresponding classes as <2,2,2>
Thus using majority we can classify this test case in
class“2”.
We have tested our proposed parallel kNN algorithm on
different UCI datasets, as listed in Table-3.
Table 3: UCI Classification Dataset
Sr.
No.
Data Set
Name
Attributes Training
Records
Testing
Records
1. Balloons 4 12 4
2.
Space
Shuttle
6 12 4
3. Seeds 7 180 30
4.
Met
Data
6 432 60
5. ILPD 8 499 80
6.
CMC
Data
7 1400 74
The experimental results shown that, the model created to
predict for target query variable, is working fine and
produces almost 85 to 95% accuracy for prediction. As
represented in Table4, the accuracy of the prediction
depends on the size of the training dataset and the value of
the ‘k’. It is also observed that, large the volume of the
training dataset, the more mature the model is. Also, large
size of ‘k’ refines clustering output for the input query, but it
does not work always. The best values of the ‘k’ for the
given dataset has to identify by multiple observation and
outputs. We have tested our parallel kNN model on
following UCI datasets, categorised for classification only.
We have proved our model for performance and
classification accuracy, as shown in Table4 and Table5
respectively.
Table 4: Parallel kNN Algorithm (Classification Accuracy)
Data Set
Name
‘k’ for
kNN
Parallel
Accuracy
Balloons 3 100 %
Space Shuttle 5 75 %
Seeds 7 86 %
Met Data 9 95 %
ILPD 7 93 %
CMC Data 11 96 %
Table:5 Parallel kNN Performance on UCI Dataset
Sr
.N
o
DataSe
t Name
‘k’
Value
Traini
ng
Record
s
Testing
Rector
s
CPU
Time GPU
Time
1
Balloon
s
3 12 4
0.102sec
9 ms
2
Space
Shuttle
5 12 4
0.157sec
13 ms
3 Seeds 7 180 30 0.477sec 21 ms
4
Met
Data
9 432 60
0.621sec
56 ms
5 ILPD 7 499 80 1.253sec 76 ms
6
CMC
Data
11 1400 74
26.72sec 1384m
s
4. CONCLUSION
kNN, the instance-based classifier operate on the premises
that classification of unknown instances can be done by
relating the unknown to the known using the distance
6. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 10 | Oct-2014, Available @ http://www.ijret.org 372
function. Closer the distance, proximity for the similarity of
the class is more. In order to minimize the kNN compute
time, we have designed OpenCL based parallel algorithm
for many core highly parallel architectures. The results for
our OpenCL based parkNN proves that the performance
scales up sub-linearly towards the improved performance of
parallel kNN. The classification accuracy is not identified
with the proportional value of ‘k’, the better ‘k’ is to be
identified by multiple observations only. However,
concerned to performance our parallel kNN, parkNN, we
observed satisfactory scale up.
REFERENCES
[1] Shenshen Liang; Cheng Wang; Ying Liu; Liheng
Jian, "CUKNN: A parallel implementation of K-
nearest neighbor on CUDA-enabled GPU,"
Information, Computing and Telecommunication,
2009. YC-ICT '09. IEEE Youth Conference on ,
vol., no., pp.415,418, 20-21 Sept. 2009
[2] Quansheng Kuang, and Lei Zhao, “A Practical
GPU based KNN algorithm”, Proceedings of the
Second Symposium International Computer
Science and Computational Technology(ISCSCT
’09)
[3] Garcia, V.; Debreuve, E.; Nielsen, F.; Barlaud, M.,
"K-nearest neighbor search: Fast GPU-based
implementations and application to high-
dimensional feature matching," Image Processing
(ICIP), 2010 17th IEEE International Conference
on , vol., no., pp.3757,3760, 26-29 Sept. 2010
[4] G. Nolan, Improving the k-Nearest Neighbour
Algorithm with CUDA
[5] A. S. Arefin, C. Riveros, R. Berretta, and P.
Moscato, “Gpu-fs-knn: A fast and scalable knn
computation technique using gpu,” Faculty
Postgrad Research Poster, The University of
Newcastle, Australia, Sept 2011.
[6] Ahmed Shamsul Arefin, Carlos Riveros, Regina
Berretta, Pablo Moscato, “kNN-MST-
Agglomerative: A Fast and Scalable Graph-based
Data Clustering Approach” on GPU The 7th
International Conference on Computer Science &
Education (ICCSE 2012), July 14-17, 2012.
Melbourne, Australia
[7] Franco-Arcega, A.; Suarez-Cansino, J.; Flores-
Flores, L.G., "A parallel algorithm to induce
decision trees for large datasets," Information,
Communication and Automation Technologies
(ICAT), 2013 XXIV International Symposium on ,
pp.1,6, Oct. 30th - Nov. 1st , 2013.
[8] J.Zhu, G.Chen, B. Wu, "GPGPU Memory
Estimation and Optimization Targeting OpenCL
Architecture", IEEE International Conference on
Cluster Computing (CLUSTER), pp. 449-458,
Sept. 2012.
[9] Khronos OpenCL working group, “The OpenCL
Specification Version2.0”, Edited by A. Munshi,
Nov. 2013.
https://www.khronos.org/registry/cl/specs/opencl-
2.0.pdf.
BIOGRAPHIES
Valmik B Nikam is Bachelor of
Engineering (Computer Science and
Engineering) from Government College of
Engineering Aurangabad, Master of
Engineering (Computer Engineering) from
VJTI, Matunga, Mumbai, Maharashtra
state, and pursuing PhD in Computer
Department of VJTI. He was faculty at Dr. Babasaheb
Ambedkar Technological University, Lonere. He has 12
years of academic experience and 5 years of administrative
experience as a Head of Department. He has one year of
industry experience. He has attended many short-term
training programs and has been invited for expert lectures in
the workshops. Presently he is Associate Professor at
department of Computer Engineering & Information
Technology of VJTI, Matunga, Mumbai. His research
interests include Scalability of Data Mining Algorithms,
Data Warehousing, Big Data, Parallel Computing, GPU
Computing, Cloud Computing. He is member of CSI, ACM,
IEEE research organizations, and a life member of ISTE. He
has been felicitated with IBM-DRONA award in 2011.
B.B. Meshram is a Professor and Head of
Department of Computer Engineering and
Information Technology, Veermata Jijabai
Technological Institute, Matunga,
Mumbai. He is Ph.D. in Computer
Engineering. He has been in the
academics & research since 20 years. His current research
includes database technologies, data mining, securities,
forensic analysis, video processing, distributed computing.
He has authored over 203 research publications, out of
which over 38 publications at National, 91 publications at
international conferences, and more than 71 in international
journals, also he has filed eight patents. He has given
numerous invited talks at various conferences, workshops,
training programs and also served as chair/co-chair for many
conferences/workshops in the area of computer science and
engineering. The industry demanded M.Tech program on
Network Infrastructure Management System, and the
International conference “Interface” are his brain childs to
interface the industry, academia & researchers. Beyond the
researcher, he also runs the Jeman Educational Society to
uplift the needy and deprived students of the society, as a
responsibility towards the society and hence the Nation.