In this paper, we introduce a set of new kernel functions derived from the generalized Legendre polynomials to obtain more robust and higher support vector machine (SVM) classification accuracy. The generalized Legendre kernel functions are suggested to provide a value of how two given vectors are like each other by changing the inner product of these two vectors into a greater dimensional space. The proposed kernel functions satisfy the Mercer’s condition and orthogonality properties for reaching the optimal result with low number support vector (SV). For that, the new set of Legendre kernel functions could be utilized in classification applications as effective substitutes to those generally used like Gaussian, Polynomial and Wavelet kernel functions. The suggested kernel functions are calculated in compared to the current kernels such as Gaussian, Polynomial, Wavelets and Chebyshev kernels by application to various non-separable data sets with some attributes. It is seen that the suggested kernel functions could give competitive classification outcomes in comparison with other kernel functions. Thus, on the basis test outcomes, we show that the suggested kernel functions are more robust about the kernel parameter change and reach the minimal SV number for classification generally.
The International Journal of Engineering and Science (The IJES)theijes
This document summarizes a research paper that proposes a novel approach to improving the k-means clustering algorithm. The standard k-means algorithm is computationally expensive and produces results that depend heavily on the initial centroid selection. The proposed approach determines initial centroids systematically and uses a heuristic to efficiently assign data points to clusters. It improves both the accuracy and efficiency of k-means clustering by ensuring the entire process takes O(n2) time without sacrificing cluster quality.
This document provides an overview of clustering techniques. It defines clustering as grouping a set of similar objects into classes, with objects within a cluster being similar to each other and dissimilar to objects in other clusters. The document then discusses partitioning, hierarchical, and density-based clustering methods. It also covers mathematical elements of clustering like partitions, distances, and data types. The goal of clustering is to minimize a similarity function to create high similarity within clusters and low similarity between clusters.
Clustering: Large Databases in data miningZHAO Sam
The document discusses different approaches for clustering large databases, including divide-and-conquer, incremental, and parallel clustering. It describes three major scalable clustering algorithms: BIRCH, which incrementally clusters incoming records and organizes clusters in a tree structure; CURE, which uses a divide-and-conquer approach to partition data and cluster subsets independently; and DBSCAN, a density-based algorithm that groups together densely populated areas of points.
This document describes a study that used fuzzy kohonen clustering network (FKCN) and inverse distance weighting (IDW) to create a building damage hazard zonation for Banda Aceh city, Indonesia. FKCN was used to cluster physical parameter data (lithology, topography, peak ground acceleration) into 3 classes. IDW was then used to interpolate the data and divide the city into 3 zones - low, medium, and high building damage areas. The results showed most of the city is in the low to medium hazard zones, with only a small high hazard zone in the northwest near the coast. This differs from a previous study that found relatively more high hazard areas along the coast.
An image can be seen as a matrix I, where I(x, y) is the brightness of the pixel located at coordinates (x, y). In the Convolutional neural network, the kernel is nothing but a filter
that is used to extract the features from the images.
Content Based Image Retrieval Using 2-D Discrete Wavelet TransformIOSR Journals
This document proposes a content-based image retrieval system using 2D discrete wavelet transform and texture features. The system decomposes images using 2D DWT, extracts texture features from low frequency coefficients using GLCM, and retrieves similar images by calculating Euclidean distances between feature vectors. Experimental results on Wang's database show the proposed approach achieves 89.8% average retrieval accuracy.
This document compares the Efficient Fuzzy Kohonen Clustering Network (EFKCN) algorithm to the original Fuzzy Kohonen Clustering Network (FKCN) algorithm to determine if EFKCN improves clustering results. Through empirical testing and simulations using expanded Fisher's Iris Data, the results showed that EFKCN did not achieve better accuracy than FKCN and is not yet a better clustering algorithm. While EFKCN had smaller errors in cluster center positions, FKCN consistently achieved higher correct clustering rates, even when the algorithms were run for the same number of iterations. Therefore, the document concludes that EFKCN does not actually improve upon the clustering performance of the original FKCN algorithm.
The International Journal of Engineering and Science (The IJES)theijes
This document summarizes a research paper that proposes a novel approach to improving the k-means clustering algorithm. The standard k-means algorithm is computationally expensive and produces results that depend heavily on the initial centroid selection. The proposed approach determines initial centroids systematically and uses a heuristic to efficiently assign data points to clusters. It improves both the accuracy and efficiency of k-means clustering by ensuring the entire process takes O(n2) time without sacrificing cluster quality.
This document provides an overview of clustering techniques. It defines clustering as grouping a set of similar objects into classes, with objects within a cluster being similar to each other and dissimilar to objects in other clusters. The document then discusses partitioning, hierarchical, and density-based clustering methods. It also covers mathematical elements of clustering like partitions, distances, and data types. The goal of clustering is to minimize a similarity function to create high similarity within clusters and low similarity between clusters.
Clustering: Large Databases in data miningZHAO Sam
The document discusses different approaches for clustering large databases, including divide-and-conquer, incremental, and parallel clustering. It describes three major scalable clustering algorithms: BIRCH, which incrementally clusters incoming records and organizes clusters in a tree structure; CURE, which uses a divide-and-conquer approach to partition data and cluster subsets independently; and DBSCAN, a density-based algorithm that groups together densely populated areas of points.
This document describes a study that used fuzzy kohonen clustering network (FKCN) and inverse distance weighting (IDW) to create a building damage hazard zonation for Banda Aceh city, Indonesia. FKCN was used to cluster physical parameter data (lithology, topography, peak ground acceleration) into 3 classes. IDW was then used to interpolate the data and divide the city into 3 zones - low, medium, and high building damage areas. The results showed most of the city is in the low to medium hazard zones, with only a small high hazard zone in the northwest near the coast. This differs from a previous study that found relatively more high hazard areas along the coast.
An image can be seen as a matrix I, where I(x, y) is the brightness of the pixel located at coordinates (x, y). In the Convolutional neural network, the kernel is nothing but a filter
that is used to extract the features from the images.
Content Based Image Retrieval Using 2-D Discrete Wavelet TransformIOSR Journals
This document proposes a content-based image retrieval system using 2D discrete wavelet transform and texture features. The system decomposes images using 2D DWT, extracts texture features from low frequency coefficients using GLCM, and retrieves similar images by calculating Euclidean distances between feature vectors. Experimental results on Wang's database show the proposed approach achieves 89.8% average retrieval accuracy.
This document compares the Efficient Fuzzy Kohonen Clustering Network (EFKCN) algorithm to the original Fuzzy Kohonen Clustering Network (FKCN) algorithm to determine if EFKCN improves clustering results. Through empirical testing and simulations using expanded Fisher's Iris Data, the results showed that EFKCN did not achieve better accuracy than FKCN and is not yet a better clustering algorithm. While EFKCN had smaller errors in cluster center positions, FKCN consistently achieved higher correct clustering rates, even when the algorithms were run for the same number of iterations. Therefore, the document concludes that EFKCN does not actually improve upon the clustering performance of the original FKCN algorithm.
An empirical assessment of different kernel functions on the performance of s...riyaniaes
Artificial intelligence (AI) and machine learning (ML) have influenced every part of our day-to-day activities in this era of technological advancement, making a living more comfortable on the earth. Among the several AI and ML algorithms, the support vector machine (SVM) has become one of the most generally used algorithms for data mining, prediction and other (AI and ML) activities in several domains. The SVM’s performance is significantly centred on the kernel function (KF); nonetheless, there is no universal accepted ground for selecting an optimal KF for a specific domain. In this paper, we investigate empirically different KFs on the SVM performance in various fields. We illustrated the performance of the SVM based on different KF through extensive experimental results. Our empirical results show that no single KF is always suitable for achieving high accuracy and generalisation in all domains. However, the gaussian radial basis function (RBF) kernel is often the default choice. Also, if the KF parameters of the RBF and exponential RBF are optimised, they outperform the linear and sigmoid KF based SVM method in terms of accuracy. Besides, the linear KF is more suitable for the linearly separable dataset.
Support Vector Machine Optimal Kernel SelectionIRJET Journal
This document discusses selecting the optimal kernel for support vector machines (SVMs) based on different datasets. It provides background on SVMs and how their performance depends on the kernel function used. The document evaluates 4 kernel types (linear, polynomial, radial basis function (RBF), sigmoid) on 3 datasets: heart disease data, digit recognition data, and social network ads data. For each dataset and kernel combination, it reports accuracy, sensitivity, specificity, and kappa statistic metrics from implementing SVMs in R. The linear and RBF kernels generally performed best, with RBF working best for datasets with larger numbers of features like digit recognition data.
Relevance Vector Machines for Earthquake Response Spectra drboon
This study uses Relevance Vector Machine (RVM) regression to develop a probabilistic model for the average horizontal component of 5%-damped earthquake response spectra. Unlike conventional models, the proposed approach does not require a functional form, and constructs the model based on a set predictive variables and a set of representative ground motion records. The RVM uses Bayesian inference to determine the confidence intervals, instead of estimating them from the mean squared errors on the training set. An example application using three predictive variables (magnitude, distance and fault mechanism) is presented for sites with shear wave velocities ranging from 450 m/s to 900 m/s. The predictions from the proposed model are compared to an existing parametric model. The results demonstrate the validity of the proposed model, and suggest that it can be used as an alternative to the conventional ground motion models. Future studies will investigate the effect of additional predictive variables on the predictive performance of the model.
Relevance Vector Machines for Earthquake Response Spectra drboon
This study uses Relevance Vector Machine (RVM) regression to develop a probabilistic model for the average horizontal component of 5%-damped earthquake response spectra. Unlike conventional models, the proposed approach does not require a functional form, and constructs the model based on a set predictive variables and a set of representative ground motion records. The RVM uses Bayesian inference to determine the confidence intervals, instead of estimating them from the mean squared errors on the training set. An example application using three predictive variables (magnitude, distance and fault mechanism) is presented for sites with shear wave velocities ranging from 450 m/s to 900 m/s. The predictions from the proposed model are compared to an existing parametric model. The results demonstrate the validity of the proposed model, and suggest that it can be used as an alternative to the conventional ground motion models. Future studies will investigate the effect of additional predictive variables on the predictive performance of the model.
A general frame for building optimal multiple SVM kernelsinfopapers
Dana Simian, Florin Stoica, A General Frame for Building Optimal Multiple SVM Kernels, Large-Scale Scientific Computing, Lecture Notes in Computer Science, 2012, Volume 7116/2012, 256-263, DOI: 10.1007/978-3-642-29843-1_29
APPLICATION OF PARTICLE SWARM OPTIMIZATION TO MICROWAVE TAPERED MICROSTRIP LINEScseij
This document discusses using Particle Swarm Optimization (PSO) to design a tapered microstrip transmission line to match an arbitrary load to a 50Ω line. PSO was used to optimize the impedances of a three section tapered line to minimize reflections. Simulations found impedances that gave good matching at 5GHz. PSO converged to solutions in under 1000 iterations. This demonstrates PSO's effectiveness in solving multi-objective microwave engineering optimization problems.
Application of particle swarm optimization to microwave tapered microstrip linescseij
Application of metaheuristic algorithms has been of continued interest in the field of electrical engineering
because of their powerful features. In this work special design is done for a tapered transmission line used
for matching an arbitrary real load to a 50Ω line. The problem at hand is to match this arbitray load to 50
Ω line using three section tapered transmission line with impedances in decreasing order from the load. So
the problem becomes optimizing an equation with three unknowns with various conditions. The optimized
values are obtained using Particle Swarm Optimization. It can easily be shown that PSO is very strong in
solving this kind of multiobjective optimization problems.
Evaluation of a hybrid method for constructing multiple SVM kernelsinfopapers
Dana Simian, Florin Stoica, Evaluation of a hybrid method for constructing multiple SVM kernels, Recent Advances in Computers, Proceedings of the 13th WSEAS International Conference on Computers, Recent Advances in Computer Engineering Series, WSEAS Press, Rodos, Greece, July 23-25, 2009, ISSN: 1790-5109, ISBN: 978-960-474-099-4, pp. 619-623
The document discusses a Bayesian approach called localized multi-kernel relevance vector machine (LMK-RVM) that uses multiple kernel functions to perform classification. LMK-RVM allows different kernel functions or parameters to be used in different areas of feature space, providing more flexibility than single-kernel models. It combines multi-kernel learning with the sparsity of the relevance vector machine (RVM) model. The document outlines LMK-RVM and provides examples showing it can improve classification accuracy and potentially provide sparser models compared to single-kernel approaches.
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...IRJET Journal
This document discusses machine learning algorithms for image classification using five different classification schemes. It summarizes the mathematical models behind each classification algorithm, including Nearest Class Centroid classifier, Nearest Sub-Class Centroid classifier, k-Nearest Neighbor classifier, Perceptron trained using Backpropagation, and Perceptron trained using Mean Squared Error. It also describes two datasets used in the experiments - the MNIST dataset of handwritten digits and the ORL face recognition dataset. The performance of the five classification schemes are compared on these datasets.
Fitness function X-means for prolonging wireless sensor networks lifetimeIJECEIAES
X-means and k-means are clustering algorithms proposed as a solution for prolonging wireless sensor networks (WSN) lifetime. In general, X-means overcomes k-means limitations such as predetermined number of clusters. The main concept of X-means is to create a network with basic clusters called parents and then generate (j) number of children clusters by parents splitting. X-means did not provide any criteria for splitting parent’s clusters, nor does it provide a method to determine the acceptable number of children. This article proposes fitness function X-means (FFX-means) as an enhancement of X-means; FFX-means has a new method that determines if the parent clusters are worth splitting or not based on predefined network criteria, and later on it determines the number of children. Furthermore, FFX-means proposes a new cluster-heads selection method, where the cluster-head is selected based on the remaining energy of the node and the intra-cluster distance. The simulation results show that FFX-means extend network lifetime by 11.5% over X-means and 75.34% over k-means. Furthermore, the results show that FFX-means balance the node’s energy consumption, and nearly all nodes depleted their energy within an acceptable range of simulation rounds.
This summary provides an overview of the key points from the document:
1) The document presents the use of General Regression Neural Networks (GRNN) to predict propagation path loss in an urban environment based on measurements taken in Kavala, Greece.
2) Two neural network models are studied - one for path loss prediction and another using error control. Their performance is compared to measured path loss values based on error metrics.
3) For line-of-sight predictions, the GRNN model achieves better performance than empirical models due to using multiple input parameters and generalization. For non-line-of-sight, a third GRNN model including street orientation has the lowest error rates.
Expert system design for elastic scattering neutrons optical model using bpnnijcsa
In present paper, a proposed expert system is designed to obtain a trained formulae for the optical model
parameters used in elastic scattering neutrons of light nuclei for (7Li), at energy range between [(1) to
(20)] MeV. A simple algorithm has used to design this expert system, while a multi-layer backwardpropagation
neural network (BPNN) is applied for training and testing the data used in this model. This
group of formulae may get a simple expert system occurring from governing formulae model, and predicts
the critical parameters usually resulted from the complicated computer coding methods. This expert system
may use in nuclear reactions yields in both fission and fusion nature who gives more closely results to the
real model.
This document discusses using a particle swarm optimization algorithm to solve the K-node set reliability optimization problem for distributed computing systems. The K-node set reliability optimization problem aims to maximize the reliability of a subset of k nodes in a distributed computing system while satisfying a specified capacity constraint. It presents the problem formulation and describes particle swarm optimization, a metaheuristic optimization technique inspired by swarm intelligence. The proposed algorithm applies a discrete particle swarm optimization approach to solve the K-node set reliability optimization problem, which is demonstrated on an example distributed computing system topology.
The document proposes a new method called Kernel-based Dynamic Subspace Method (KDSM) for classifying high-dimensional data. KDSM combines an ensemble technique of support vector machines with an optimal kernel method. It uses a dynamic subspace approach to select informative feature subsets and an optimal algorithm to select parameters for the radial basis function kernel. The method is tested on hyperspectral image data and achieves higher classification accuracy compared to other methods while also reducing computation time, especially for datasets with small training sizes.
A Fuzzy Interactive BI-objective Model for SVM to Identify the Best Compromis...ijfls
This document summarizes a research paper that proposes a fuzzy bi-objective support vector machine (SVM) model to identify infected COVID-19 patients. The model uses SVM classification with two objectives - maximizing margin between classes and minimizing misclassification errors. An α-cut transforms the fuzzy model into a classical bi-objective problem solved using weighting methods. This generates multiple efficient solutions. An interactive process then identifies the best compromise based on minimizing the number of support vectors in each class. The model constructs a utility function to measure COVID-19 infection levels based on the SVM classification.
A FUZZY INTERACTIVE BI-OBJECTIVE MODEL FOR SVM TO IDENTIFY THE BEST COMPROMIS...ijfls
A support vector machine (SVM) learns the decision surface from two different classes of the input points. In several applications, some of the input points are misclassified and each is not fully allocated to either of these two groups. In this paper a bi-objective quadratic programming model with fuzzy parameters is utilized and different feature quality measures are optimized simultaneously. An α-cut is defined to transform the fuzzy model to a family of classical bi-objective quadratic programming problems. The weighting method is used to optimize each of these problems. For the proposed fuzzy bi-objective quadratic programming model, a major contribution will be added by obtaining different effective support vectors due to changes in weighting values. The experimental results, show the effectiveness of the α-cut with the weighting parameters on reducing the misclassification between two classes of the input points. An interactive procedure will be added to identify the best compromise solution from the generated efficient solutions. The main contribution of this paper includes constructing a utility function for measuring the degree of infection with coronavirus disease (COVID-19).
Application of support vector machines for prediction of anti hiv activity of...Alexander Decker
This document describes a study that used support vector machines (SVM) to develop a quantitative structure-activity relationship (QSAR) model to predict the anti-HIV activity of TIBO derivatives. The SVM model achieved high correlation (q2=0.96) and low error (RMSE=0.212), outperforming artificial neural networks and multiple linear regression models developed on the same data set. The results indicate that SVM is a valuable tool for QSAR modeling and predicting anti-HIV activity of chemical compounds.
Performance Analysis of Various Activation Functions in Generalized MLP Archi...Waqas Tariq
This document compares the performance of various activation functions in multilayer perceptron (MLP) neural networks. It analyzes MLP architectures using different activation functions, including bi-polar sigmoid, uni-polar sigmoid, hyperbolic tangent, conic section, and radial basis functions. Based on experiments, hyperbolic tangent performed the best in terms of accuracy, requiring fewer iterations than other functions to solve nonlinear problems. While conic section had the lowest training error, hyperbolic tangent produced the most accurate results during testing. In general, the hyperbolic tangent function achieved high accuracy and is a good choice for most MLP applications.
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...Scientific Review SR
This document summarizes a study that evaluated the performance of a kernel radial basis probabilistic neural network (Kernel RBPNN) model for classifying iris data, compared to backpropagation, radial basis function, and radial basis probabilistic neural network models. The Kernel RBPNN model achieved the highest classification accuracy of 89.12% on test data from the iris dataset, performing better than the other models. It also had the fastest training time, being over 80 times faster than the radial basis function model. Analysis of the receiver operating characteristic curves showed that the Kernel RBPNN model had the largest area under the curve, indicating it had the best classification prediction capability out of the four models evaluated.
An empirical assessment of different kernel functions on the performance of s...riyaniaes
Artificial intelligence (AI) and machine learning (ML) have influenced every part of our day-to-day activities in this era of technological advancement, making a living more comfortable on the earth. Among the several AI and ML algorithms, the support vector machine (SVM) has become one of the most generally used algorithms for data mining, prediction and other (AI and ML) activities in several domains. The SVM’s performance is significantly centred on the kernel function (KF); nonetheless, there is no universal accepted ground for selecting an optimal KF for a specific domain. In this paper, we investigate empirically different KFs on the SVM performance in various fields. We illustrated the performance of the SVM based on different KF through extensive experimental results. Our empirical results show that no single KF is always suitable for achieving high accuracy and generalisation in all domains. However, the gaussian radial basis function (RBF) kernel is often the default choice. Also, if the KF parameters of the RBF and exponential RBF are optimised, they outperform the linear and sigmoid KF based SVM method in terms of accuracy. Besides, the linear KF is more suitable for the linearly separable dataset.
Support Vector Machine Optimal Kernel SelectionIRJET Journal
This document discusses selecting the optimal kernel for support vector machines (SVMs) based on different datasets. It provides background on SVMs and how their performance depends on the kernel function used. The document evaluates 4 kernel types (linear, polynomial, radial basis function (RBF), sigmoid) on 3 datasets: heart disease data, digit recognition data, and social network ads data. For each dataset and kernel combination, it reports accuracy, sensitivity, specificity, and kappa statistic metrics from implementing SVMs in R. The linear and RBF kernels generally performed best, with RBF working best for datasets with larger numbers of features like digit recognition data.
Relevance Vector Machines for Earthquake Response Spectra drboon
This study uses Relevance Vector Machine (RVM) regression to develop a probabilistic model for the average horizontal component of 5%-damped earthquake response spectra. Unlike conventional models, the proposed approach does not require a functional form, and constructs the model based on a set predictive variables and a set of representative ground motion records. The RVM uses Bayesian inference to determine the confidence intervals, instead of estimating them from the mean squared errors on the training set. An example application using three predictive variables (magnitude, distance and fault mechanism) is presented for sites with shear wave velocities ranging from 450 m/s to 900 m/s. The predictions from the proposed model are compared to an existing parametric model. The results demonstrate the validity of the proposed model, and suggest that it can be used as an alternative to the conventional ground motion models. Future studies will investigate the effect of additional predictive variables on the predictive performance of the model.
Relevance Vector Machines for Earthquake Response Spectra drboon
This study uses Relevance Vector Machine (RVM) regression to develop a probabilistic model for the average horizontal component of 5%-damped earthquake response spectra. Unlike conventional models, the proposed approach does not require a functional form, and constructs the model based on a set predictive variables and a set of representative ground motion records. The RVM uses Bayesian inference to determine the confidence intervals, instead of estimating them from the mean squared errors on the training set. An example application using three predictive variables (magnitude, distance and fault mechanism) is presented for sites with shear wave velocities ranging from 450 m/s to 900 m/s. The predictions from the proposed model are compared to an existing parametric model. The results demonstrate the validity of the proposed model, and suggest that it can be used as an alternative to the conventional ground motion models. Future studies will investigate the effect of additional predictive variables on the predictive performance of the model.
A general frame for building optimal multiple SVM kernelsinfopapers
Dana Simian, Florin Stoica, A General Frame for Building Optimal Multiple SVM Kernels, Large-Scale Scientific Computing, Lecture Notes in Computer Science, 2012, Volume 7116/2012, 256-263, DOI: 10.1007/978-3-642-29843-1_29
APPLICATION OF PARTICLE SWARM OPTIMIZATION TO MICROWAVE TAPERED MICROSTRIP LINEScseij
This document discusses using Particle Swarm Optimization (PSO) to design a tapered microstrip transmission line to match an arbitrary load to a 50Ω line. PSO was used to optimize the impedances of a three section tapered line to minimize reflections. Simulations found impedances that gave good matching at 5GHz. PSO converged to solutions in under 1000 iterations. This demonstrates PSO's effectiveness in solving multi-objective microwave engineering optimization problems.
Application of particle swarm optimization to microwave tapered microstrip linescseij
Application of metaheuristic algorithms has been of continued interest in the field of electrical engineering
because of their powerful features. In this work special design is done for a tapered transmission line used
for matching an arbitrary real load to a 50Ω line. The problem at hand is to match this arbitray load to 50
Ω line using three section tapered transmission line with impedances in decreasing order from the load. So
the problem becomes optimizing an equation with three unknowns with various conditions. The optimized
values are obtained using Particle Swarm Optimization. It can easily be shown that PSO is very strong in
solving this kind of multiobjective optimization problems.
Evaluation of a hybrid method for constructing multiple SVM kernelsinfopapers
Dana Simian, Florin Stoica, Evaluation of a hybrid method for constructing multiple SVM kernels, Recent Advances in Computers, Proceedings of the 13th WSEAS International Conference on Computers, Recent Advances in Computer Engineering Series, WSEAS Press, Rodos, Greece, July 23-25, 2009, ISSN: 1790-5109, ISBN: 978-960-474-099-4, pp. 619-623
The document discusses a Bayesian approach called localized multi-kernel relevance vector machine (LMK-RVM) that uses multiple kernel functions to perform classification. LMK-RVM allows different kernel functions or parameters to be used in different areas of feature space, providing more flexibility than single-kernel models. It combines multi-kernel learning with the sparsity of the relevance vector machine (RVM) model. The document outlines LMK-RVM and provides examples showing it can improve classification accuracy and potentially provide sparser models compared to single-kernel approaches.
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...IRJET Journal
This document discusses machine learning algorithms for image classification using five different classification schemes. It summarizes the mathematical models behind each classification algorithm, including Nearest Class Centroid classifier, Nearest Sub-Class Centroid classifier, k-Nearest Neighbor classifier, Perceptron trained using Backpropagation, and Perceptron trained using Mean Squared Error. It also describes two datasets used in the experiments - the MNIST dataset of handwritten digits and the ORL face recognition dataset. The performance of the five classification schemes are compared on these datasets.
Fitness function X-means for prolonging wireless sensor networks lifetimeIJECEIAES
X-means and k-means are clustering algorithms proposed as a solution for prolonging wireless sensor networks (WSN) lifetime. In general, X-means overcomes k-means limitations such as predetermined number of clusters. The main concept of X-means is to create a network with basic clusters called parents and then generate (j) number of children clusters by parents splitting. X-means did not provide any criteria for splitting parent’s clusters, nor does it provide a method to determine the acceptable number of children. This article proposes fitness function X-means (FFX-means) as an enhancement of X-means; FFX-means has a new method that determines if the parent clusters are worth splitting or not based on predefined network criteria, and later on it determines the number of children. Furthermore, FFX-means proposes a new cluster-heads selection method, where the cluster-head is selected based on the remaining energy of the node and the intra-cluster distance. The simulation results show that FFX-means extend network lifetime by 11.5% over X-means and 75.34% over k-means. Furthermore, the results show that FFX-means balance the node’s energy consumption, and nearly all nodes depleted their energy within an acceptable range of simulation rounds.
This summary provides an overview of the key points from the document:
1) The document presents the use of General Regression Neural Networks (GRNN) to predict propagation path loss in an urban environment based on measurements taken in Kavala, Greece.
2) Two neural network models are studied - one for path loss prediction and another using error control. Their performance is compared to measured path loss values based on error metrics.
3) For line-of-sight predictions, the GRNN model achieves better performance than empirical models due to using multiple input parameters and generalization. For non-line-of-sight, a third GRNN model including street orientation has the lowest error rates.
Expert system design for elastic scattering neutrons optical model using bpnnijcsa
In present paper, a proposed expert system is designed to obtain a trained formulae for the optical model
parameters used in elastic scattering neutrons of light nuclei for (7Li), at energy range between [(1) to
(20)] MeV. A simple algorithm has used to design this expert system, while a multi-layer backwardpropagation
neural network (BPNN) is applied for training and testing the data used in this model. This
group of formulae may get a simple expert system occurring from governing formulae model, and predicts
the critical parameters usually resulted from the complicated computer coding methods. This expert system
may use in nuclear reactions yields in both fission and fusion nature who gives more closely results to the
real model.
This document discusses using a particle swarm optimization algorithm to solve the K-node set reliability optimization problem for distributed computing systems. The K-node set reliability optimization problem aims to maximize the reliability of a subset of k nodes in a distributed computing system while satisfying a specified capacity constraint. It presents the problem formulation and describes particle swarm optimization, a metaheuristic optimization technique inspired by swarm intelligence. The proposed algorithm applies a discrete particle swarm optimization approach to solve the K-node set reliability optimization problem, which is demonstrated on an example distributed computing system topology.
The document proposes a new method called Kernel-based Dynamic Subspace Method (KDSM) for classifying high-dimensional data. KDSM combines an ensemble technique of support vector machines with an optimal kernel method. It uses a dynamic subspace approach to select informative feature subsets and an optimal algorithm to select parameters for the radial basis function kernel. The method is tested on hyperspectral image data and achieves higher classification accuracy compared to other methods while also reducing computation time, especially for datasets with small training sizes.
A Fuzzy Interactive BI-objective Model for SVM to Identify the Best Compromis...ijfls
This document summarizes a research paper that proposes a fuzzy bi-objective support vector machine (SVM) model to identify infected COVID-19 patients. The model uses SVM classification with two objectives - maximizing margin between classes and minimizing misclassification errors. An α-cut transforms the fuzzy model into a classical bi-objective problem solved using weighting methods. This generates multiple efficient solutions. An interactive process then identifies the best compromise based on minimizing the number of support vectors in each class. The model constructs a utility function to measure COVID-19 infection levels based on the SVM classification.
A FUZZY INTERACTIVE BI-OBJECTIVE MODEL FOR SVM TO IDENTIFY THE BEST COMPROMIS...ijfls
A support vector machine (SVM) learns the decision surface from two different classes of the input points. In several applications, some of the input points are misclassified and each is not fully allocated to either of these two groups. In this paper a bi-objective quadratic programming model with fuzzy parameters is utilized and different feature quality measures are optimized simultaneously. An α-cut is defined to transform the fuzzy model to a family of classical bi-objective quadratic programming problems. The weighting method is used to optimize each of these problems. For the proposed fuzzy bi-objective quadratic programming model, a major contribution will be added by obtaining different effective support vectors due to changes in weighting values. The experimental results, show the effectiveness of the α-cut with the weighting parameters on reducing the misclassification between two classes of the input points. An interactive procedure will be added to identify the best compromise solution from the generated efficient solutions. The main contribution of this paper includes constructing a utility function for measuring the degree of infection with coronavirus disease (COVID-19).
Application of support vector machines for prediction of anti hiv activity of...Alexander Decker
This document describes a study that used support vector machines (SVM) to develop a quantitative structure-activity relationship (QSAR) model to predict the anti-HIV activity of TIBO derivatives. The SVM model achieved high correlation (q2=0.96) and low error (RMSE=0.212), outperforming artificial neural networks and multiple linear regression models developed on the same data set. The results indicate that SVM is a valuable tool for QSAR modeling and predicting anti-HIV activity of chemical compounds.
Performance Analysis of Various Activation Functions in Generalized MLP Archi...Waqas Tariq
This document compares the performance of various activation functions in multilayer perceptron (MLP) neural networks. It analyzes MLP architectures using different activation functions, including bi-polar sigmoid, uni-polar sigmoid, hyperbolic tangent, conic section, and radial basis functions. Based on experiments, hyperbolic tangent performed the best in terms of accuracy, requiring fewer iterations than other functions to solve nonlinear problems. While conic section had the lowest training error, hyperbolic tangent produced the most accurate results during testing. In general, the hyperbolic tangent function achieved high accuracy and is a good choice for most MLP applications.
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...Scientific Review SR
This document summarizes a study that evaluated the performance of a kernel radial basis probabilistic neural network (Kernel RBPNN) model for classifying iris data, compared to backpropagation, radial basis function, and radial basis probabilistic neural network models. The Kernel RBPNN model achieved the highest classification accuracy of 89.12% on test data from the iris dataset, performing better than the other models. It also had the fastest training time, being over 80 times faster than the radial basis function model. Analysis of the receiver operating characteristic curves showed that the Kernel RBPNN model had the largest area under the curve, indicating it had the best classification prediction capability out of the four models evaluated.
Similar to GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFICATION (20)
Null Bangalore | Pentesters Approach to AWS IAMDivyanshu
#Abstract:
- Learn more about the real-world methods for auditing AWS IAM (Identity and Access Management) as a pentester. So let us proceed with a brief discussion of IAM as well as some typical misconfigurations and their potential exploits in order to reinforce the understanding of IAM security best practices.
- Gain actionable insights into AWS IAM policies and roles, using hands on approach.
#Prerequisites:
- Basic understanding of AWS services and architecture
- Familiarity with cloud security concepts
- Experience using the AWS Management Console or AWS CLI.
- For hands on lab create account on [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
# Scenario Covered:
- Basics of IAM in AWS
- Implementing IAM Policies with Least Privilege to Manage S3 Bucket
- Objective: Create an S3 bucket with least privilege IAM policy and validate access.
- Steps:
- Create S3 bucket.
- Attach least privilege policy to IAM user.
- Validate access.
- Exploiting IAM PassRole Misconfiguration
-Allows a user to pass a specific IAM role to an AWS service (ec2), typically used for service access delegation. Then exploit PassRole Misconfiguration granting unauthorized access to sensitive resources.
- Objective: Demonstrate how a PassRole misconfiguration can grant unauthorized access.
- Steps:
- Allow user to pass IAM role to EC2.
- Exploit misconfiguration for unauthorized access.
- Access sensitive resources.
- Exploiting IAM AssumeRole Misconfiguration with Overly Permissive Role
- An overly permissive IAM role configuration can lead to privilege escalation by creating a role with administrative privileges and allow a user to assume this role.
- Objective: Show how overly permissive IAM roles can lead to privilege escalation.
- Steps:
- Create role with administrative privileges.
- Allow user to assume the role.
- Perform administrative actions.
- Differentiation between PassRole vs AssumeRole
Try at [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
Accident detection system project report.pdfKamal Acharya
The Rapid growth of technology and infrastructure has made our lives easier. The
advent of technology has also increased the traffic hazards and the road accidents take place
frequently which causes huge loss of life and property because of the poor emergency facilities.
Many lives could have been saved if emergency service could get accident information and
reach in time. Our project will provide an optimum solution to this draw back. A piezo electric
sensor can be used as a crash or rollover detector of the vehicle during and after a crash. With
signals from a piezo electric sensor, a severe accident can be recognized. According to this
project when a vehicle meets with an accident immediately piezo electric sensor will detect the
signal or if a car rolls over. Then with the help of GSM module and GPS module, the location
will be sent to the emergency contact. Then after conforming the location necessary action will
be taken. If the person meets with a small accident or if there is no serious threat to anyone’s
life, then the alert message can be terminated by the driver by a switch provided in order to
avoid wasting the valuable time of the medical rescue team.
Digital Twins Computer Networking Paper Presentation.pptxaryanpankaj78
A Digital Twin in computer networking is a virtual representation of a physical network, used to simulate, analyze, and optimize network performance and reliability. It leverages real-time data to enhance network management, predict issues, and improve decision-making processes.
Build the Next Generation of Apps with the Einstein 1 Platform.
Rejoignez Philippe Ozil pour une session de workshops qui vous guidera à travers les détails de la plateforme Einstein 1, l'importance des données pour la création d'applications d'intelligence artificielle et les différents outils et technologies que Salesforce propose pour vous apporter tous les bénéfices de l'IA.
Determination of Equivalent Circuit parameters and performance characteristic...pvpriya2
Includes the testing of induction motor to draw the circle diagram of induction motor with step wise procedure and calculation for the same. Also explains the working and application of Induction generator
Determination of Equivalent Circuit parameters and performance characteristic...
GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFICATION
1. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
DOI: 10.5121/ijnsa.2019.11406 87
GENERALIZED LEGENDRE POLYNOMIALS FOR
SUPPORT VECTOR MACHINES (SVMS)
CLASSIFICATION
Ashraf Afifi1
and E.A.Zanaty2
1
Department of Computer Engineering, Computers and Information Technology College,
Taif University, Al-Hawiya 21974, Kingdom of Saudi Arabia
2
Computer Science Dept., Faculty of Science, Sohag University, Sohag, Egypt.
ABSTRACT
In this paper, we introduce a set of new kernel functions derived from the generalized Legendre
polynomials to obtain more robust and higher support vector machine (SVM) classification accuracy. The
generalized Legendre kernel functions are suggested to provide a value of how two given vectors are like
each other by changing the inner product of these two vectors into a greater dimensional space. The
proposed kernel functions satisfy the Mercer’s condition and orthogonality properties for reaching the
optimal result with low number support vector (SV). For that, the new set of Legendre kernel functions
could be utilized in classification applications as effective substitutes to those generally used like Gaussian,
Polynomial and Wavelet kernel functions. The suggested kernel functions are calculated in compared to the
current kernels such as Gaussian, Polynomial, Wavelets and Chebyshev kernels by application to various
non-separable data sets with some attributes. It is seen that the suggested kernel functions could give
competitive classification outcomes in comparison with other kernel functions. Thus, on the basis test
outcomes, we show that the suggested kernel functions are more robust about the kernel parameter change
and reach the minimal SV number for classification generally.
KEYWORDS
Legendre Polynomials, Kernel Functions, Functional Analysis, SVMS, Classification Problem.
1. INTRODUCTION
Support Vector Machines (SVMs) has become famous machines for data classification as a result
of use for the vast data set and practical for application [1-3]. The operation of SVMs is based
upon selecting kernel functions [4-6]. Picking various kernel functions will give out various
SVMs [7- 9] and may turn out to be in various performances [10-11]. Some effort has been
carried out on curbing kernels by handling prior knowledge; however, the optimal selection of a
kernel for a provided problem is yet a free research crisis [12]. Chapelle and Schölkopf [13]
suggested a kernel to use constant transformations. The disadvantage here is that they are most
probably just suitable for linear SVM classifiers. Hastie et al. [14] had given comparisons among
multi-class SVMs algorithms when implied to defy data set. Zanaty et al. [15-17] mixed GF and
RBF functions to attain new kernel functions that can make use of their corresponding power. In
[18-19], the Hermite kernel functions were defined for advancing the operation of SVMs in a
variety of applications. Meng and Wenjian [20] proposed orthogonal polynomials to advance
generalization performance in both classification and regression duties. The particular estimation
of crossing kernel SVMs which is logarithmic in time was shown in Maji et al. [21]. They proved
that the procedure is approximately in complex and the classification efficacy is passable, but the
2. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
88
runtimes are symbolically boosted in comparison with the implanted radial bases function (RBF)
and polynomial kernel (POLY) because of the great number of SVs for every classifier [14, 21].
Ozer et al. [22] presented kernel functions coming from the Chebyshev polynomials. They built
various kernel functions so that they can catch the highly non-linear boundaries in the Euclidian
space. In Jiang and Ching [23], the managed kernel learning with an SVM classifier was
outstandingly applied in biomedical diagnosis like segregating various types of tumour tissues for
noisy Raman Spectra, see [24-25] for further details.
The problems of data classification remain in picking the most convenient kernel of SVMs for a
specific application, specifically since various functions and parameters can have vastly different
operations [19-22]. A vital research field in SVMs is to establish an effective kernel function for
constructing SVMs in a particular application, specifically because of variable current application
which will demand various methods [22].
In this paper, Legendre kernel functions are constructed to advance the classification certainty of
SVMs for both linear and non-linear data groups. We sustain a group of Legendre kernel
functions based on advancing SVMs classification certainty. The class of Legendre kernel
functions fulfils mercer conditions and gives competitive operation in comparison with all other
typical kernel functions with the same standard of the simulation datasets. The suggested kernels
can be used for categorizing compound data which have numerous properties.
The remainder of the paper is arranged in this way: In section 2, SVM classifications are
elaborated. The kernel theory is deliberated in section 3. The generalized Legendre kernels are
introduced in section 4. Section 5 shows functional examination on the presented Legendre
kernels. Experimental and comparative outcomes are given in section 6. Lastly, section 7 presents
the conclusion.
2. SUPPORT VECTOR MACHINE (SVM)
SVMs became popular for data classification and regression. They used to segregate the data by
constructing two hyperplanes. If you have a set of N points NkRx n
k
,.....,1, to be
conjoined with a label }1,1{ k
y which can be categorized the data into one of two
groups. According to SVMs formulation, the classifier )(xy will be the design of a hyper-plane
wT
xk + b which shows optimum separation 2
2
w
between points k
x belonging to the two
classes. This introduces an optimization issue of the form:
,1][s.t.
2
1
min: , bxwywwQ k
T
k
T
bw
(1)
where the wwT
2
1
term stands for a cost function to be minimized to maximize segregation.
The restraints are formulated so that the nearest points k
x with labels [either +1 or -1] are (with
suitable input space scaling) at least 2
1
w
away from the separating hyper-plane. Nevertheless,
3. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
89
for the Least-Squares SVM classification, changes are done so that at the target value, an error
variable ek is enabled so that misclassifying can be gone along with in case of coinciding
distributions and the subsequent optimization problem is framed in the primal weight space for a
given training set
N
kkk
yx 1
},{
N
k
k
T
p
cbw
ewwewJQ
1
2
,, 2
1
2
1
),(: min
(2)
Along with the N restrains as given in Eq.(3). This formula included the tradeoff between a cost
function term and a sum of squared errors governed by the trade-off parameter γ.
Nkebxwy kk
T
k ,..,1,1])([ (3)
To solve this ‘primal minimization’ issue, we design the dual maximization of Eq.(2) using the
Lagrangian form:
),;,,(max:
cbwLD
(4)
Where
N
k
kk
T
kkp ebxwyewJL
1
}1])([{),(
(5)
and αk are Lagrange multipliers. The circumstances for optimality are given by
(6)
After removal of the variables w and e we get this solution:
v
T
c
b
Iy
y
1
0
/
0
(7)
where ]1;...;1[],,..;1 vN Iyyy and ];;...;[ 1 N
The kernel trick is applied here as follows:
)()( L
T
kLkkl xxyy NLkxxKyy Lkk L
,...,1,),( (8)
The outcoming LS-SVM model for classifier turns out:
Nk
ebxwy
L
Nke
e
L
y
b
L
xyw
w
L
kk
T
k
k
kk
k
N
k kk
N
k kkk
,......,1,where
,01])(0
,..,1.0
00
)(0
1
1
4. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
90
N
k
kkk bxxKysignxy
1
),()( (9)
where αk , b are the answer to the linear system presented by Eq.(7) and N stand for the
number of non-zero Lagrange multipliers αk, called SVs.
According to Eq.(9), the kernel functions have been applied on the pairs of elements
separately, for a given pair of two input vectors x and z, the outcoming kernel can be
formulated as:
)(),( jjj zx(x,z)K (10)
Where (.)K j is the kernel function that is evaluated on the jth
elements of the vector
pair zandx . More various kernel functions were found in literature as in Table (1).
Table 1: Expressions of kernels list
kernel kernel expression
Polynomials kernel [14]: n
zx
zxK
1
).(
Gaussian kernel [25]
2
2
).(
ji xx
ezxK
Wavelet kernel [22]:
2
2
1 2
exp75.1).(
a
zx
a
zx
CoszxK
jjjjm
j
, m is
order of polynomials.
Chebyshev kernel [7]:
jj
n
i jijim
j zx
zTxT
zxK
1
)()(
).(
0
1
, m is polynomials order and
T(x) is the Chebyshev polynomials.
3. PROPOSED KERNEL FUNCTIONS
In the suggested modified Henon map will be defined in terms of two basic processes
namely ciphering and deciphering. To advance the classification certainty of SVMs,
various kernel functions are required for various applications. We figured that Legendre
function will ensure to be effective kernels for numerous applications. From the solution
of Legendre’s differential equation, the formula of Legendre polynomials may be written
down using Rodrigues’ formula:
])1[(
!2
1
)( 2 n
n
n
nn x
dx
d
n
xQ (11)
By differentiating )1( n times both sides of the identity:
5. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
91
nn
xnxx
dx
d
x )1(2)1()1( 222
and applying the general Leibniz rule for repeated differentiation.
3.1. Legendre Recurrence
From Eq.(16), expanding for the first two terms gives:
,)(,1)( 10 xxQxQ
Eq.(11) is discriminated concerning t on both sides to acquire more terms with no use of direct
broadening of the Taylor series, and reorganized to attain:
1.
0
12
2
)()21(
21 n
n
n txnQtxt
txt
tx
Replacing the quotient of the square root with its description in (11), and equating the coefficients
of powers of t in the outcoming expansion gives Bonnet’s recursion formula:
0
2
)(
21
1
),(
n
n
n txQ
txt
xt
(12)
Theorem 1: Taylor series expansion of Legendre’s differential equation:
0
2
)(
21
1
),(
n
n
n txQ
txt
xt
can be represented as the following recurrence relation:
2.
).()()12()()1( 11 xnQxxQnxQn nnn
Proof: differentiate Eq. (12) with respect to t .
0 1
12
1
1
1
2
3
2
)()21()()1()(
)(
)21(2
)(2
j j
j
j
j
j
j
j
j
txjQtxttxQjtx
txjQ
txt
xt
t
g
0 1 0 1 2
111 )()1()(2)()1()()(
j j j j j
j
j
j
j
j
j
j
j
j
j txQjtxjQxtxQjtxQtxQx
Compare the coefficients of t j
:
6. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
92
(j=0)
,)(,1)( 10 xxQxQ
(j=1)
2/)13(221
22
2
2
2
2
2
1201
xQxQx
xQQQxQ
)2( j
11
111
)12()1(
)1(2)1(
jjj
jjjjj
jQxQjQj
QjxjQQjQxQ
3.2. Orthogonally of Legendre Function
Legendre polynomials ),(xQn 0,1,2,3,...,n N form a whole orthogonal group on the
interval 1 1.x
Theorem 2: Let Pn(x) denote the Legendre polynomial of degree n . Then:
1
1
0)()( nmifdxxQxQ nm
Proof: We know that the polynomials )(xQn and )(xQm satisfy:
andxQnnxQx nn 0)()1())'(')1(( 2
(13)
0)()1())'(')1(( 2
xQmmxQx mm (14)
Multiplying equation (1) by )(xQm and equation (2) by ),(xQn , we can obtain:
),())'(')1(()())'(')1(()()()1()1(( 22
xQxQxxQxQxxQxQmmnn mnnmmn
Therefore,
1
1
22
1
1
)())'(')1(()())'(')1(()()()1()1(( dxxQxQxxQxQxdxxQxQmmnn mnnmmn
=
1
1
2
1
1
2
)())'(')1()())'(')1(
x
xmnnm xQxQxdxxQxQx
.0)())'(')1()())'(')1(
1
1
2
1
1
2
x
xmnnm xQxQxdxxQxQx
Since
)1()1(, mmnnmn
Then we have:
7. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
93
1
1
0)()( nmifdxxQxQ nm
This completes the proof.
3.3. Generalized Legendre Kernels
Here, we are suggesting a general method of conveying the kernel function to resolve the
vagueness on how to apply Legendre kernels. As of what we know, there was a preceding work
illustrating the Legendre polynomials for vector inputs recursively. thus for vector inputs, we
illustrate the generalized Legendre polynomials as:
1)(0 xQ
xxQ )(1
11 )12()1( jjj jQxQjQj
0),())1/(()()1/()12(()( 11 jxQjjxxQjjxQ jjj
Therefore, the generalized Legendre, )(xQj , yields a row vector; if not, it gives a scalar value.
Therefore, with the use of generalized Legendre polynomials, we describe generalized nth
order
Legendre kernel as
n
j
T
jj zQxQzxK
0
)()(),(
(15)
Where x and z are m-dimensional vectors. In Eq. (13), the denominator should be greater than
zero. To fulfill (14), as every element in x and z vectors has a value in between [-1,1], the
maximum value for the inner product zx. is equal to
m
I
mI
1
thus minimum a
value will be equal to mwhich is the dimension of input vector x.
Consequently, the 5th
order generalized Legendre kernel could be presented as:
13500
)88.55370124131.598571243243)(88.553701241321.598571243243(
9720
)1756820568693555)(1756820568693555(
50625
)66.2325.3825.1181)(66.2325.3825.1181(
144
)2335)(2335(
3
)25)(25(
1),(
2323
22
22
bbbaaa
bbaa
c
bbaabacba
czxk
Where xxa . , zzb . , and zxc . . In addition, the first 4th
order
kernel functions are listed in Table 1.
8. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
94
Table 2: List of the generated Legendre kernel functions up to 4th
order.
Parameter : n Kernel function: ( , )k x z , Parameter : n
0 1
1 1+c
2 1+c+
3
)25)(25( ba
3 1+c+
144
)2335)(2335(
3
)25)(25(
bacba
4
2 2
(5 2)(5 2) (35 23)(35 23)
1
3 144
(1181.25 38.325 2.66)(1181.25 38.325 2.66)
50625
a b c a b
c
a a b b
4. FUNCTIONAL ANALYSIS
In order to reproduce proposed kernels, we use functional analysis as described in [26] to prove
Mercer’s theorem conditions. The mapping could be designed from the eigen function
decomposition of k . With respect to Mercer’s work [27-28], it is known that if k is the
symmetrical and continuous kernel of an integral operator 22
: LLOk , in such a way that:
dzzzxKxOkg )(),()(
is positive, i.e.,
,0)()(),( 2
LdxdzzxzxK
then k can be expanded into a uniformly convergent series
1
,)()(),(
i
iii zxzxK with .0i In this case, the mapping from input space to feature
space produced by the kernel is expressed as ),...)(),((: 2211 xxx such
that k acts as the given dot product, i.e.,
).,()()())(),(( zxKzxzx T
Theorem 3: A kernel is a valid SVM kernel; if it satisfies the Mercer Conditions [29].
Proof: If the kernel does not fulfill the Mercer Conditions, SVM might not derive the best
parameters, but instead it might bring up suboptimal parameters. Additionally, in case of the
Mercer conditions not being fulfilled, the Hessian matrix for the optimization portion might not
be positive straightforward. Thus we inspect if the generalized Legendre kernel fulfill the Mercer
conditions.
9. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
95
To be a credible SVM kernel, for any finite function )(x , the subsequent integration must
always be non-negative for the given kernel function ),( zxk [1]:
0(z)dxdz(x)z)K(x,
(16)
Where
n
j
T
nn
TTT
jj zQxQzQxQzQxQzQxQzxk
0
1100 )()(.......)()()()()()(),( (17)
Regard that )(x is a function where : ,RRm
then we can calculate and validate the
Mercer condition for ),( zxk as follows by speculating each element is independent of others:
Thus, the kernel ),( zxk is a verified kernel.
Theorem 4: A nonnegative linear combination of Mercer kernels is also a Mercer kernel.
Proof: Let ),...,1( MiKi be Mercer kernels and let
M
i
ii zxKazxK
1
),,(),(
where 0ia is a nonnegative constant. According to Mercer’s theorem, we have
.,...,1,,0)()(),( 2
MiLdxdzzxzxKi (19)
By taking the sum of the positive combination of (19) with coefficients ia overi , one obtains
.,0)()(),( 2
1
LgdxdzzxzxKa i
M
i
i (20)
Therefore, one reaches
.,0)()(),( 2
LgdxdzzxzxK
In particular, if
M
i ii aa1
),0(,1 then we consider ),( zxK in (20) as the convex mixing of
the positive definite kernels ),( zxKi . This kind of kernel can find many applications in practice.
0)()()()(
)()()()()()()()(
)()()()()()(),(
0
00
0
n
j
T
jj
n
j
T
jj
n
j
T
jj
n
j
T
jj
dxxxQdxxxQ
dzzzQdxxgxQdxdzzxzQxQ
dxdzzxzQxQdxdzzxzxk
(18)
10. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
96
Theorem 5: The product of Mercer kernels is also a Mercer kernel.
Proof: is like that of the previous theorem.
5. EXPERIMENTAL RESULTS AND DISCUSSIONS
The multi-class problem is described as the categorizing problem that has numerous classes. To
prolong these classifiers to take care of many classes, the target of this method is to map the
generalization capabilities of the binary classifiers to the multi-class domain. Multi-class SVMs
are ordinarily applied by mixing a couple of two class SVMs. In multi-class experimentations, we
have trained SVM for every class individually so that one is against all. In every experiment, we
utilized the SVM toolbox accessible at [32]. The multi-class problem is described as a
classification problem which has numerous classes or characters. Current SVMs [24] are binary
classifiers, i.e., they could categorize two classes. To be capable of dealing with various classes
(greater than 2), the current classifiers should be prolonged. The target is to depict the
generalization abilities of the binary classifiers to the multi-class domain. Multi-class SVMs are
usually applied by merging several two - class SVMs. The classifier is constructed to read two
input data files, the training data and the test data (for more details see [11, 18]). Every file is
arranged as records, each of which is made up of a vector of attributes x: ),...,,( 211 mi xxxx
followed by the target y: ),...,,( 212 ci yyyx where c is the number of classes m and is the
number of attributes. The SVM designs binary classifiers, and utilizes the training data to find the
maximum separating hyperplane.
The classification experimentations are carried out on number image segmentation data sets like
Brickface, Foliage, Sky, Cement, Window, Path and Grass data set [30-31]. The data has 7
diverse image classes. It has 210 data for training and a different 2100 data for evaluating. Every
vector has 18 elements having diverse maximum and minimum values. For the training, we got
30 data for the class (+1) and 180 data for the class (1) and likely for testing, we have 300 against
1800 data, correspondingly for every class. With the test step, the kernel functions presented
various performance values on various classes and there was winning kernel presenting the
optimal performance one very class as shown in Table (3). The suggested Legendre kernel
operated better than the standard. The optimal performance values having the least SV numbers
are presented in bold. Table (3) illustrates the test results for every class with various kernel
functions. We carry out some evaluations to compare the suggested kernel with its preceding
opposite as well as the Gaussian (GF) [25], polynomial (POLY) [14], Wavelet [7] and Chebyshev
[22] kernel functions. The operation of the suggested kernel with SVMs according to
classification accuracy (ACC) and kernel parameter against SV, is calculated by implementation
to data sets in Table (3). As shown in Table (3), the generalized Legendre kernel results present
better generalization capability than the current GF, POLY, Wavelet and Chebyshev kernels. For
instance, the optimal ACC is retrieved for Brickface, Foliage, Window, Path, Grass with the least
SV to be 17, 6, 4, 6 and 15 correspondingly. Even though the ACC of the generalized Legendre
kernel and GF kernel for Foliage data is similar, we emphasize that SV of the generalized
Legendre kernel has the least. In Figs.(1-7), we give the ACC against SV of the POLY, GF,
Wavelet, Chebyshev, and generalized Legendre kernels for Brickface, Foliage, Sky, Cement,
Window, Path and Grass data set. The generalized Legendre kernel functions present the minimal
SV 4 while maintaining the generalization ability right for the dataset. The relation between ACC
and SV showed that, as the SV increases, the ACC increases and asymptotically arrive at a high
performance value.
11. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
97
Figs. (8-12) describe the relation between kernel parameter vs. SV for GF, POLY, Wavelet and
Chebyshev kernel functions when these methods are implemented on Brickface, Foliage, Sky,
Cement, Window, Path and Grass data set correspondingly. These figures prove that as the kernel
parameter increases, the Chebyshev, Legendre and wavelet kernel increase their operation and
asymptotically reach allow performance value. While as the kernel parameter increases, the
Chebyshev and wavelet kernels need more SV than Legendre kernel. During the evaluations, we
witnessed that the generalized Legendre kernel function reaches the minimal SV number in
general.
Table 3: Data classification results with different kernel functions.
Data
sets
GF POLY Wavelet Chebyshev Legendre
SV
No.
ACC σ
SV
no.
ACC n
SV
no.
ACC a
SV
no.
ACC n
SV
no.
ACC n
Brickf
ace
61 0.960 0.4 11
2
0.969 2 44 0.997 1.4 23 0.997 0 17 0.999 1
5
Sky 7 1.0 4.3 14 1.0 2 5 1.0 1.5 5 0.99 0 6 1.0 3
Foliag
e
106 0.98 1.4 12
2
0.999 8 99 0.989 2.4 122 0.971 0 6 0.999 2
Cemen
t
88 0.989 0.5 78 0.971 1
2
77 0.997 1.4 65 0.999 4 16 0.998 1
Windo
w
49 0.957 0.4 80 0.969 7 34 0.99 1.4 34 0.985 4 4 0.99 2
Path 44 0.969 0.43 68 0.959 6 66 0.98 1.5 22 0.976 3 6 0.98 1
Grass 22 0.986 11 25 0.964 2 45 0.979 1.5 54 0.975 0 15 0.990 2
Figure 1: SV vs. ACC for Brickface data.
12. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
98
Figure 2: SV vs. ACC for Sky data.
Figure 3: SV vs. ACC for Foliage data.
Figure 4: SV vs. ACC for Cement data
13. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
99
Figure 5: SV vs. ACC for Window data.
Figure 6: SV vs. ACC for Path data.
Figure 7: SV vs. ACC for Grass data
14. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
100
Figure 8: Polynomial kernel parameter vs. SV number.
Figure 9: Gaussian kernel parameter vs. SV number
Figure 10 : Wavelet kernel parameter vs. SV.
15. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
101
Figure 11: Chebyshev kernel parameter vs. SV.
Figure 12: Legendre kernel parameter vs. SV.
6. CONCLUSION
Presenting the current paper, the classification certainty of SVMs has become advanced by
mapping the training data into a feature space by the help groups of Legendre functions. A class
of Legendre kernel functions based upon the properties of the common kernels is suggested,
being able to recognize many applications in training. Normalization takes a vital job for
generalized Legendre kernel, and thus the whole data should be normalized between [-1,1] before
utilizing the kernel function. Upon the simulation results, it can be said that picking order of
Legendre polynomials from an integer group is usually sufficient to acquire a good classification
consequence from the generalized Legendre kernel function.
We have made a comparison between the classification efficacy of the Legendre kernel function
and the current kernels like the current GF, POLY and Wavelet kernels. In accordance with the
test outcomes, the generalized Legendre kernel shows the lowest number of support vectors on
almost every evaluation. In strictly, we suggest this is derived from the orthogonally characteristic
of the Legendre polynomials. This character of the kernel function can be vital and helpful in
16. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
102
several applications where the support vector number is greatly vital as in feature selection.
Therefore generalized Legendre kernel functions can be regarded as a valid substitute to the GF,
POLY, and Wavelet kernel functions for a couple of particular datasets. The test outcomes imply
that their outcomes have been analogs to the kernel functions developed from the generalized
Legendre polynomials of the primer kind. Thus, we have not comprised the family of kernel
functions and their outcomes in this study. Also, since handling the properties of the generalized
Legendre polynomials is beyond this study, we have not studied these properties specifically even
though this study could be helpful to design new kernel functions developed from generalized
Legendre polynomials and that could be the goal of upcoming work.
REFERENCES
[1] Vapnik V. N., (1995) “The nature of statistical learning theory”, Springer-Verlag, New York, NY,
USA.
[2] Kim H., Pang S., Je H., Kim D., Bang S.Y., (2003) “Constructing support vector machine ensemble”,
Pattern Recognition”, vol.36, no.12, pp.2757–2767.
[3] Du P., Peng J., Terlaky T.. (2009) “Self-adaptive support vector machines”, modeling and
experiments Computational Management Science, vol. 6, no.1, pp. 41–51.
[4] Boser B. E., Guyon I. M., Vapnik V. N., (1995) “A training algorithm for optimal margin classifiers”,
Proc. Fifth Ann. Workshop Computing Learning Theory, pp. 144-152.
[5] Vapnik V.N. , (1999 “An overview of statistical learning theory)”, IEEE Trans. Neural networks, vol.
10, no. 5, pp. 988-999.
[6] Cortes C., Vapnik V.N., (1995) “Support-vector networks”, Machine Learning, vol. 20, pp. 273- 297.
[7] Scholkopf B., (1997) “Support vector learning”, PhD dissertation, Technische Universitat Berlin,
Germany.
[8] Vapnik V.N., Golowich S., and Smola A., (1997) “Support vector method for function approximation,
regression estimation and signal processing”, Advances in Neural Information processing Systems,
vol. 9, Cambridge, Mass.: MIT Press.
[9] Mangasarian O.L., Musicant D.R., (1999) “Successive over relaxation for support vector
machines”, IEEE Trans. Neural Networks, vol. 10, no. 5, pp. 1032-1037.
[10] Aronszajn N. , (1950) “Theory of Reproducing Kernels”, Trans. Am. Math. Soc., vol. 68, pp. 337-
404.
[11] Shawe-Taylor J., Bartlett P.L., Willianmson R.C., Anthony M., (1998) “Structural risk
minimization over data-dependent hierarchies”, IEEE Trans. Information Theory, vol. 44, no. 5, pp.
1926-1940.
[12] Williamson R. C., Smola A., Schölkopf B., (1999) “Entropy numbers, operators and support vector
kernels”, MA: MIT Press, Cambridge, pp. 44-127.
[13] Chapelle O. and Schölkopf B., (2002) “Incorporating invariances in non-linear support vector
machines”, In T. G. Dietterich, S. Becker, and Z. Ghahramani (Eds.), Advances in Neural Information
Processing Systems, vol. 14, pp. 594-609, Cambridge, MA: MIT Press.
[14] Hastie T., Hsu C-W., Lin C-J, (2005) “A comparison of methods for multi-class support vector
Machines”, IEEE Transactions on Neural Networks, no. 13, pp.415–425.
17. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
103
[15] Zanaty , E.A. ,and Sultan Aljahdali, (2008) “Improving the accuracy of support vector machines”,
Proceedings of 23rd International conference on computers and their application April, pp. 75-83,
Cancun, Mexico.
[16] Zanaty E.A, Sultan Aljahdali, R.J. Cripps, (2009) “Accurate support vector machines for data
classification”, Int. J. Rapid Manufacturing, vol. 1, no. 2, pp. 114-127.
[17] Zanaty E.A, Ashraf Afifi, (2011) ”Support vector machines (SVMs) with universal kernels “, in
International journal of Artificial Intelligence, vol. 25 , pp.575-589.
[18] E. A. Zanaty &Ashraf Afifi, Jun (2018) “Generalized Hermite kernel function for support vector
machine classifications”, International Journal of Computer Applications, In press, Accepted 07 Jun
2018.
[19] Vahid H. M., Javad H.,., (2016) “New Hermite Orthogonal Polynomial Kernel and Combined
Kernels in Support Vector Machine Classifier”, Elsevier, Pattern Recognition, vol. 60, pp. 921–935.
[20] Meng T., Wenjian W., (2017) “Some Sets of Orthognal Polynomial Kernal Functions”, Elsevier,
Applied Soft Computing, vol. 61, pp. 741–756.
[21] Maji S., Berg A.C., Malik J. ,(2008) “Classification using intersection kernel support vector machines
is efficient”, IEEE Conference on Computer Vision and Pattern Recognition, June pp.1–8.
[22] Ozer Sedat, Chen ChiH., Cirpan HakanA., (2011) “A set of new Chebyshev kernel functions for
support vector machine pattern classification”, Pattern Recognition, no. 44, pp.1435–1447.
[23] Hao Jiang, Wai-Ki Ching, (2012) "Correlation kernels for support vector machines classification with
applications in cancer data", Computational and Mathematical Methods in Medicine,
doi:10.1155/2012/205025, pp.1-7.
[24] A. Kyriakides, E. Kastanos, K. Hadjigeorgiou, and C. Pitris, (2011) “Classification of Raman spectra
using the correlation kernel” , Journal of Raman Spectroscopy, vol. 42, no. 5, pp. 904–909.
[25] A. Kyriakides, E. Kastanos, K. Hadjigeorgiou, and C. Pitris, (January 2011) “Support vector machines
with the correlation kernel for the classification of Raman spectra”, Advanced Biomedical and
Clinical Diagnostic Systems IX, vol. 7890 of Proceedings of SPIE, pp. 78901B-1–78901B-7, San
Francisco, Calif, USA.
[26] Zhi-B., Hong C., Xin H., Luis C. P., Martin C., Alfonso, Rojas D., Hector P., Hector F., (2018) “A
novel Formulation of Orthognal Polynomial Kernal Functions for SVM Classifiers : The Gegenbauer
family”, Elsevier, Pattern Recognition, vol. 84, pp. 211–225.
[27] Mercer T. , ( 1909)“Functions of positive and negative type and their connection with the theory of
integral equations”, Philosophical Trans. of the Royal Soc. of London, Series A, pp. 415-446.
[28] Williamson R.C., Smola A.J., and Scholkopf B., (1998) “Generalization performance of
regularization networks and support vector machines via entropy numbers of compact operators”,
Technical Report 19, NeuroCOLT.
[29] Scholkopf B., A.J. Smola, (2001) “Learning with kernels: support vector machines, regularization,
optimization”, and Beyond, MIT Press.
[30] Web: http://www.liacc.up.pt/ML/old/statlog/datasets.html
[31] Web: http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/
[32] Web:http://www. http://svm.sourceforge.net/docs/3.00/api/
18. International Journal of Network Security & Its Applications (IJNSA) Vol. 11, No.4, July 2019
104
AUTHOR
Ashraf Afifi received the B.Sc.(Hons.), M.Sc., and Ph.D. degrees in Electronics and
Communication Engineering from Zagazig University, Egypt, in 1987, 1995, and
2002, respectively. He is currently Associate Professor with the Department of
Computer Engineering, Faculty of Computers and Information Technology, Taif
University, Saudi Arabia. He is a co-author of about 30 papers in international
journals and conference proceedings. His research interests cover communication
security, image processing, and image encryption. Email: a.afifi@tu.edu.sa, Mobile:
00966507275265.
Prof. E.A.Zanaty is a professor and of Computer Science at the Faculty of Science,
Sohag University, Sohag, Egypt. He received his Bachelor of Mathematics degree
from Sohag Faculty of Science, Assuit University, Egypt, 1992. He received his MSC
Degree in Computer Science in 1997 from South Valley University, Sohag, Egypt
and started his career as Assistant Lecturer in November 1997 at the same Faculty.
Prof. Zanaty completed his Ph.D Studies with Prof. Guido Bruneet supervision at
TU-Chemnitz, Germany, during the period 2000-2003. During this period, he worked
as Assistant researcher in Computer Graphics and Visualization Department, College
of Computer Science, Chemnitz University, Germany. After that, he backed to Egypt to work as senior of
Computer Science laboratory at Sohag University, Egypt. During 2007 till 2016, Prof. Zanaty contracted
with College of Computers and Information Technology, Taif University, Taif, Saudi Arabia. He
supervised them on several committees and he became the Head of Computer Science Department in 2009
while he earned a professor job in February 2015 from Sohag University, Egypt and also in May 2015 from
Taif University, Taif, Saudi Arabia.
Prof. Zanaty is an Editorial board of several journals and member of KACST and IAENG. He also is the
Editor in Chief of International Journal of Informatics and Medical Data Processing (IJIMDP). His research
activities are focused on artificial intelligence, data classification, reverse engineering, data reduction,
medical image segmentation, and reconstruction. In these areas, he has published several technical papers
in refereed international journals or conference proceedings.