This document proposes using machine learning techniques to predict COVID-19 infections based on chest x-ray images. Specifically, it involves using discrete wavelet transform to extract space-frequency features from chest x-rays, reducing the dimensionality of features using Shannon entropy, and then training standard machine learning classifiers like logistic regression, support vector machine, decision tree, and convolutional neural network on the extracted features to classify images as COVID-19 positive or negative. The document provides background on the proposed techniques of discrete wavelet transform, entropy, and various machine learning models.
The document reports on the results of three image processing projects. The first project implemented Lloyd-Max quantization to reduce image file sizes and Retinex theory to compensate for uneven illumination. The second project used principal component analysis to compute eigenfaces for face recognition. The third project performed linear discriminant analysis and tensor-based linear discriminant analysis for binary classification and visual object recognition. Illumination compensation subtracted an estimated illumination plane from image intensities to reduce shadows. Eigenfaces were the principal components of a training set of face images. Tensor-based linear discriminant analysis treated images as higher-order tensors to outperform conventional LDA.
Adaptive lifting based image compression scheme using interactive artificial ...csandit
This paper presents image compression method using Interactive Artificial Bee Colony (IABC) optimization algorithm. The proposed method reduces storage and facilitates data transmission by reducing transmission costs. To get the finest quality of compressed image, utilizing local search, IABC determines different update coefficient, and the best update coefficient is chosen
optimally. By using local search in the update step, we alter the center pixels with the coefficient in 8-different directions with a considerable window size, to produce the compressed image, expressed in terms of both PSNR and compression ratio. The IABC brings in the idea of
universal gravitation into the consideration of the affection between onlooker bees and the employed bees. By passing on different values of the control parameter, the universal gravitation involved in the IABC has various quantities of the single onlooker bee and employed bees. As a result when compared to existing methods, the proposed work gives better PSNR.
Improvement of Anomaly Detection Algorithms in Hyperspectral Images Using Dis...sipij
Recently anomaly detection (AD) has become an important application for target detection in hyperspectral remotely sensed images. In many applications, in addition to high accuracy of detection we need a fast and reliable algorithm as well. This paper presents a novel method to improve the performance of current AD algorithms. The proposed method first calculates Discrete Wavelet Transform (DWT) of every pixel vector of image using Daubechies4 wavelet. Then, AD algorithm performs on four bands of “Wavelet transform” matrix which are the approximation of main image. In this research some benchmark AD algorithms including Local RX, DWRX and DWEST have been implemented on Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) hyperspectral datasets. Experimental results demonstrate significant improvement of runtime in proposed method. In addition, this method improves the accuracy of AD algorithms because of DWT’s power in extracting approximation coefficients of signal, which contain the main behaviour of signal, and abandon the redundant information in hyperspectral image data.
Performance Evaluation of Object Tracking Technique Based on Position VectorsCSCJournals
This document presents a novel algorithm for object tracking in video frames based on position vectors. The algorithm first extracts position vectors for the object in the first frame. It then tracks the object across subsequent frames by cropping each new frame into blocks based on shifted position vectors and performing block matching between frames using feature vectors extracted by discrete wavelet transform (DWT) or dual tree complex wavelet transform (DTCWT). Experimental results on video sequences show the algorithm using DTCWT achieves higher tracking precision (95%) compared to DWT (92%). The algorithm is computationally efficient and can track multiple moving and still objects.
In this paper person identification is done based on sets of facial images. Each facial image is considered as the scattered point of logistic regression. The vertical distance of scattered point of facial image and the regression line is considered as the parameter to determine whether the image is of same person or not. The ratio of Euclidian distance (in terms of number of pixel of gray scale image based on ‘imtool’ of Matlab 13.0) between nasal and eye points are determined. The variance of the ration is considered another parameter to identify a facial image. The concept is combined with ghost image of Principal Component Analysis; where the mean square error and signal to noise ratio (SNR) in dB is considered as the parameters of detection. The combination of three methods, enhance the degree of accuracy compared to individual one.
For more info visit us at: http://www.siliconmentor.com/
Support vector machines are widely used binary classifiers known for its ability to handle high dimensional data that classifies data by separating classes with a hyper-plane that maximizes the margin between them. The data points that are closest to hyper-plane are known as support vectors. Thus the selected decision boundary will be the one that minimizes the generalization error (by maximizing the margin between classes).
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
Vector-Based Back Propagation Algorithm of.pdfNesrine Wagaa
This document presents a vector-based backpropagation algorithm for a supervised convolution neural network (CNN) model. The key points are:
- The CNN model consists of one convolution layer followed by three fully connected hidden layers for classification of handwritten digits using the MNIST dataset.
- The classical convolution operation is replaced by a matrix operation to avoid mathematical complexities. Convolution maps and filters are represented as vectors.
- Forward propagation involves applying the new convolution and pooling operations to extract features, then passing the output through the fully connected layers.
- Backpropagation is used to update the CNN parameters (filters, weights, biases) via gradient descent to minimize a cost function, with update equations derived for both the convolution
The document reports on the results of three image processing projects. The first project implemented Lloyd-Max quantization to reduce image file sizes and Retinex theory to compensate for uneven illumination. The second project used principal component analysis to compute eigenfaces for face recognition. The third project performed linear discriminant analysis and tensor-based linear discriminant analysis for binary classification and visual object recognition. Illumination compensation subtracted an estimated illumination plane from image intensities to reduce shadows. Eigenfaces were the principal components of a training set of face images. Tensor-based linear discriminant analysis treated images as higher-order tensors to outperform conventional LDA.
Adaptive lifting based image compression scheme using interactive artificial ...csandit
This paper presents image compression method using Interactive Artificial Bee Colony (IABC) optimization algorithm. The proposed method reduces storage and facilitates data transmission by reducing transmission costs. To get the finest quality of compressed image, utilizing local search, IABC determines different update coefficient, and the best update coefficient is chosen
optimally. By using local search in the update step, we alter the center pixels with the coefficient in 8-different directions with a considerable window size, to produce the compressed image, expressed in terms of both PSNR and compression ratio. The IABC brings in the idea of
universal gravitation into the consideration of the affection between onlooker bees and the employed bees. By passing on different values of the control parameter, the universal gravitation involved in the IABC has various quantities of the single onlooker bee and employed bees. As a result when compared to existing methods, the proposed work gives better PSNR.
Improvement of Anomaly Detection Algorithms in Hyperspectral Images Using Dis...sipij
Recently anomaly detection (AD) has become an important application for target detection in hyperspectral remotely sensed images. In many applications, in addition to high accuracy of detection we need a fast and reliable algorithm as well. This paper presents a novel method to improve the performance of current AD algorithms. The proposed method first calculates Discrete Wavelet Transform (DWT) of every pixel vector of image using Daubechies4 wavelet. Then, AD algorithm performs on four bands of “Wavelet transform” matrix which are the approximation of main image. In this research some benchmark AD algorithms including Local RX, DWRX and DWEST have been implemented on Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) hyperspectral datasets. Experimental results demonstrate significant improvement of runtime in proposed method. In addition, this method improves the accuracy of AD algorithms because of DWT’s power in extracting approximation coefficients of signal, which contain the main behaviour of signal, and abandon the redundant information in hyperspectral image data.
Performance Evaluation of Object Tracking Technique Based on Position VectorsCSCJournals
This document presents a novel algorithm for object tracking in video frames based on position vectors. The algorithm first extracts position vectors for the object in the first frame. It then tracks the object across subsequent frames by cropping each new frame into blocks based on shifted position vectors and performing block matching between frames using feature vectors extracted by discrete wavelet transform (DWT) or dual tree complex wavelet transform (DTCWT). Experimental results on video sequences show the algorithm using DTCWT achieves higher tracking precision (95%) compared to DWT (92%). The algorithm is computationally efficient and can track multiple moving and still objects.
In this paper person identification is done based on sets of facial images. Each facial image is considered as the scattered point of logistic regression. The vertical distance of scattered point of facial image and the regression line is considered as the parameter to determine whether the image is of same person or not. The ratio of Euclidian distance (in terms of number of pixel of gray scale image based on ‘imtool’ of Matlab 13.0) between nasal and eye points are determined. The variance of the ration is considered another parameter to identify a facial image. The concept is combined with ghost image of Principal Component Analysis; where the mean square error and signal to noise ratio (SNR) in dB is considered as the parameters of detection. The combination of three methods, enhance the degree of accuracy compared to individual one.
For more info visit us at: http://www.siliconmentor.com/
Support vector machines are widely used binary classifiers known for its ability to handle high dimensional data that classifies data by separating classes with a hyper-plane that maximizes the margin between them. The data points that are closest to hyper-plane are known as support vectors. Thus the selected decision boundary will be the one that minimizes the generalization error (by maximizing the margin between classes).
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
Vector-Based Back Propagation Algorithm of.pdfNesrine Wagaa
This document presents a vector-based backpropagation algorithm for a supervised convolution neural network (CNN) model. The key points are:
- The CNN model consists of one convolution layer followed by three fully connected hidden layers for classification of handwritten digits using the MNIST dataset.
- The classical convolution operation is replaced by a matrix operation to avoid mathematical complexities. Convolution maps and filters are represented as vectors.
- Forward propagation involves applying the new convolution and pooling operations to extract features, then passing the output through the fully connected layers.
- Backpropagation is used to update the CNN parameters (filters, weights, biases) via gradient descent to minimize a cost function, with update equations derived for both the convolution
This document describes a machine learning project that uses support vector machines (SVM) and k-nearest neighbors (k-NN) algorithms to segment gesture phases based on radial basis function (RBF) kernels and k-nearest neighbors. The project aims to classify frames of movement data into five gesture phases (rest, preparation, stroke, hold, retraction) using two classifiers. The SVM approach achieved 53.27% accuracy on test data while the k-NN approach achieved significantly higher accuracy of 92.53%. The document provides details on the dataset, feature extraction methods, model selection process and results of applying each classifier to the test data.
Linear regression [Theory and Application (In physics point of view) using py...ANIRBANMAJUMDAR18
Machine-learning models are behind many recent technological advances, including high-accuracy translations of the text and self-driving cars. They are also increasingly used by researchers to help in solving physics problems, like Finding new phases of matter, Detecting interesting outliers
in data from high-energy physics experiments, Founding astronomical objects are known as gravitational lenses in maps of the night sky etc. The rudimentary algorithm that every Machine Learning enthusiast starts with is a linear regression algorithm. In statistics, linear regression is a linear approach to modelling the relationship between a scalar response (or dependent variable) and one or more explanatory variables (or independent
variables). Linear regression analysis (least squares) is used in a physics lab to prepare the computer-aided report and to fit data. In this article, the application is made to experiment: 'DETERMINATION OF DIELECTRIC CONSTANT OF NON-CONDUCTING LIQUIDS'. The entire computation is made through Python 3.6 programming language in this article.
This document discusses principal component analysis (PCA) and its applications in image processing and facial recognition. PCA is a technique used to reduce the dimensionality of data while retaining as much information as possible. It works by transforming a set of correlated variables into a set of linearly uncorrelated variables called principal components. The first principal component accounts for as much of the variability in the data as possible, and each succeeding component accounts for as much of the remaining variability as possible. The document provides an example of applying PCA to a set of facial images to reduce them to their principal components for analysis and recognition.
NUMBER PLATE IMAGE DETECTION FOR FAST MOTION VEHICLES USING BLUR KERNEL ESTIM...paperpublications3
This document discusses a proposed method for detecting number plates on images of fast moving vehicles that have been blurred due to motion. It begins with an introduction to image processing and digital images. It then discusses estimating the blur kernel caused by vehicle motion in order to model it as a linear uniform blur with parameters for angle and length. Existing related works on image deblurring are reviewed. The proposed system estimates the blur kernel parameters using sparse representation and Radon transform methods, allows deblurring the image, and then uses artificial neural networks to identify numbers and characters in the deblurred image. The system is evaluated on real blurred images and shown to improve license plate recognition compared to previous methods.
1) The document proposes using homogeneous motion discovery to generate additional reference frames for 4K video coding. Motion is estimated between reference frames and the current frame to generate affine motion models and associated masks.
2) Experimental results on 3 video sequences show average bit rate savings of 3.78% over HEVC by using the additional reference frames generated from the affine motion models.
3) The approach provides a simpler computation method for high resolution video coding compared to motion hint estimation, which requires super-pixel segmentation that becomes impractical for resolutions like 4K.
Q UANTUM C LUSTERING -B ASED F EATURE SUBSET S ELECTION FOR MAMMOGRAPHIC I...ijcsit
In this paper, we present an algorithm for feature selection. This algorithm labeled QC-FS: Quantum
Clustering for Feature Selection performs the selection in two steps. Partitioning the original features
space in order to group similar features is performed using the Quantum Clustering algorithm. Then the
selection of a representative for each cluster is carried out. It uses similarity measures such as correlation
coefficient (CC) and the mutual information (MI). The feature which maximizes this information is chosen
by the algorithm
Density Driven Image Coding for Tumor Detection in mri ImageIOSRjournaljce
The significant of multi spectral band resolution is explored towards selection of feature coefficients based on its energy density. Toward the feature representiaon in transformed domain, multi wavelet transformations were used for finer spectral representation. However, due to a large feature count these features are not optimal under low resource computing system. In the recognition units, running with low resources a new coding approach of feature selection, considering the band spectral density is developed. The effective selection of feature element, based on its spectral density achieve two objective of pattern recognition, the feature coefficient representiaon is minimized, hence leading to lower resource requirement, and dominant feature representation, resulting in higher retrieval performance.
Propagation of Error Bounds due to Active Subspace ReductionMohammad
This document summarizes the propagation of error bounds due to active subspace reduction in computational models. It presents two algorithms for performing active subspace reduction: one that is gradient-free and reduces the response or state space, and one that is gradient-based and reduces the parameter space. It then develops a theorem for propagating error bounds across multiple reductions, both in the parameter and response spaces. Numerical experiments on an analytic function and a nuclear reactor pin cell model are used to validate the error bound approach.
A Novel Algorithm for Design Tree Classification with PCAEditor Jacotech
This document summarizes a research paper titled "A Novel Algorithm for Design Tree Classification with PCA". It discusses dimensionality reduction techniques like principal component analysis (PCA) that can improve the efficiency of classification algorithms on high-dimensional data. PCA transforms data to a new coordinate system such that the greatest variance by any projection of the data comes to lie on the first coordinate, called the first principal component. The paper proposes applying PCA and linear transformation on an original dataset before using a decision tree classification algorithm, in order to get better classification results.
This document summarizes a research paper titled "A Novel Algorithm for Design Tree Classification with PCA". It discusses dimensionality reduction techniques like principal component analysis (PCA) that can improve the efficiency of classification algorithms on high-dimensional data. PCA transforms data to a new coordinate system such that the greatest variance by any projection of the data comes to lie on the first coordinate, called the first principal component. The paper proposes applying PCA and linear transformation on an original dataset before using a decision tree classification algorithm, in order to get better classification results.
The document discusses an algorithm called Adaptive Multichannel Component Analysis (AMMCA) for separating image sources from mixtures using adaptively learned dictionaries. It begins by reviewing image denoising using learned dictionaries, then extends this to image separation from single mixtures. The key contribution is applying this approach to separating sources from multichannel mixtures by learning local dictionaries for each source during the separation process. The algorithm is described and simulated results are shown separating two images from a noisy mixture using the learned dictionaries. In conclusion, AMMCA is able to separate sources without prior knowledge of their sparsity domains by fusing dictionary learning into the separation process.
Abstract This paper presents a novel approach for the gesture recognition system using software. In this paper the real time image is taken and is compared with a training set of images and displays a matched training image. In this approach we have used skin detection techniques for detecting the skin threshold regions, Principle Component Analysis (PCA) algorithm and Linear Discriminant Analysis (LDA) for data compressing and analyzing and K-Nearest Neighbor (KNN), Support Vector Machine (SVM) classification for matching the appropriate training image to the real-time image. The software used is MATLAB. The hand gestures used are taken from the American Sign Language. Keywords— PCA algorithm, LDA algorithm, skin detection, KNN and SVM classification
This document provides an overview of knowledge representation techniques and object recognition. It discusses syntax and semantics in representation, as well as descriptions, features, grammars, languages, predicate logic, production rules, fuzzy logic, semantic nets, and frames. It then covers statistical and cluster-based pattern recognition methods, feedforward and backpropagation neural networks, unsupervised learning including Kohonen feature maps, and Hopfield neural networks. The goal is to represent knowledge in a way that enables object classification and decision-making.
This document provides an overview of knowledge representation techniques and object recognition. It discusses syntax and semantics in representation, as well as descriptions, features, grammars, languages, predicate logic, production rules, fuzzy logic, semantic nets, and frames. It then covers statistical and cluster-based pattern recognition methods, feedforward and backpropagation neural networks, unsupervised learning including Kohonen feature maps, and Hopfield neural networks. The goal is to represent knowledge in a way that enables object classification and decision-making.
This document discusses single object tracking and velocity determination. It begins with an introduction and objectives of the project which is to develop an algorithm for tracking a single object and determining its velocity in a sequence of video frames. It then provides details on preprocessing techniques like mean filtering, Gaussian smoothing and median filtering to reduce noise. It describes segmentation methods including histogram-based, single Gaussian background and frame difference approaches. Feature extraction methods like edges, bounding boxes and color are explained. Object detection using optical flow and block matching is covered. Finally, it discusses tracking and calculating velocity of the moving object. MATLAB is introduced as a technical computing language for solving these types of problems.
Gesture Recognition using Principle Component Analysis & Viola-Jones AlgorithmIJMER
Gesture recognition pertains to recognizing meaningful expressions of motion by a human,
involving the hands, arms, face, head, and/or body. It is of utmost importance in designing an intelligent
and efficient human–computer interface. The applications of gesture recognition are manifold, ranging
from sign language through medical rehabilitation to virtual reality. In this paper, we provide a survey on
gesture recognition with particular emphasis on hand gestures and facial expressions. Applications
involving wavelet transform and principal component analysis for face and hand gesture recognition on
digital images
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...IRJET Journal
This document discusses machine learning algorithms for image classification using five different classification schemes. It summarizes the mathematical models behind each classification algorithm, including Nearest Class Centroid classifier, Nearest Sub-Class Centroid classifier, k-Nearest Neighbor classifier, Perceptron trained using Backpropagation, and Perceptron trained using Mean Squared Error. It also describes two datasets used in the experiments - the MNIST dataset of handwritten digits and the ORL face recognition dataset. The performance of the five classification schemes are compared on these datasets.
A systematic image compression in the combination of linear vector quantisati...eSAT Publishing House
1) The document presents a method for image compression that combines linear vector quantization and discrete wavelet transform.
2) Linear vector quantization is used to generate codebooks and encode image blocks, achieving better PSNR and MSE than self-organizing maps.
3) The encoded blocks are then subjected to discrete wavelet transform. Low-low subbands are stored for reconstruction while other subbands are discarded.
4) Experimental results show the proposed method achieves higher PSNR and lower MSE than existing techniques, preserving both texture and edge information.
This document describes a novel statistical damage detection approach using unsupervised support vector machines (SVM). It aims to identify damage in structural components through vibration-based methods. The proposed approach builds a statistical model through unsupervised learning, avoiding the need for measurements from damaged structures. It is computationally efficient even with large numbers of features and does not suffer from local minima problems like artificial neural networks. Numerical simulations show the approach can accurately detect both the occurrence and location of damage.
This document describes a novel statistical damage detection approach using unsupervised support vector machines (SVM). It begins with an introduction to vibration-based damage detection methods and their limitations. It then provides an overview of SVM and describes how one-class SVM and support vector regression can be applied for unsupervised learning in damage detection, without requiring data from damaged structures. Feature selection is discussed to improve computational efficiency when dealing with large datasets. Numerical simulations are analyzed to examine the accuracy and scalability of the proposed approach.
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
This document describes a machine learning project that uses support vector machines (SVM) and k-nearest neighbors (k-NN) algorithms to segment gesture phases based on radial basis function (RBF) kernels and k-nearest neighbors. The project aims to classify frames of movement data into five gesture phases (rest, preparation, stroke, hold, retraction) using two classifiers. The SVM approach achieved 53.27% accuracy on test data while the k-NN approach achieved significantly higher accuracy of 92.53%. The document provides details on the dataset, feature extraction methods, model selection process and results of applying each classifier to the test data.
Linear regression [Theory and Application (In physics point of view) using py...ANIRBANMAJUMDAR18
Machine-learning models are behind many recent technological advances, including high-accuracy translations of the text and self-driving cars. They are also increasingly used by researchers to help in solving physics problems, like Finding new phases of matter, Detecting interesting outliers
in data from high-energy physics experiments, Founding astronomical objects are known as gravitational lenses in maps of the night sky etc. The rudimentary algorithm that every Machine Learning enthusiast starts with is a linear regression algorithm. In statistics, linear regression is a linear approach to modelling the relationship between a scalar response (or dependent variable) and one or more explanatory variables (or independent
variables). Linear regression analysis (least squares) is used in a physics lab to prepare the computer-aided report and to fit data. In this article, the application is made to experiment: 'DETERMINATION OF DIELECTRIC CONSTANT OF NON-CONDUCTING LIQUIDS'. The entire computation is made through Python 3.6 programming language in this article.
This document discusses principal component analysis (PCA) and its applications in image processing and facial recognition. PCA is a technique used to reduce the dimensionality of data while retaining as much information as possible. It works by transforming a set of correlated variables into a set of linearly uncorrelated variables called principal components. The first principal component accounts for as much of the variability in the data as possible, and each succeeding component accounts for as much of the remaining variability as possible. The document provides an example of applying PCA to a set of facial images to reduce them to their principal components for analysis and recognition.
NUMBER PLATE IMAGE DETECTION FOR FAST MOTION VEHICLES USING BLUR KERNEL ESTIM...paperpublications3
This document discusses a proposed method for detecting number plates on images of fast moving vehicles that have been blurred due to motion. It begins with an introduction to image processing and digital images. It then discusses estimating the blur kernel caused by vehicle motion in order to model it as a linear uniform blur with parameters for angle and length. Existing related works on image deblurring are reviewed. The proposed system estimates the blur kernel parameters using sparse representation and Radon transform methods, allows deblurring the image, and then uses artificial neural networks to identify numbers and characters in the deblurred image. The system is evaluated on real blurred images and shown to improve license plate recognition compared to previous methods.
1) The document proposes using homogeneous motion discovery to generate additional reference frames for 4K video coding. Motion is estimated between reference frames and the current frame to generate affine motion models and associated masks.
2) Experimental results on 3 video sequences show average bit rate savings of 3.78% over HEVC by using the additional reference frames generated from the affine motion models.
3) The approach provides a simpler computation method for high resolution video coding compared to motion hint estimation, which requires super-pixel segmentation that becomes impractical for resolutions like 4K.
Q UANTUM C LUSTERING -B ASED F EATURE SUBSET S ELECTION FOR MAMMOGRAPHIC I...ijcsit
In this paper, we present an algorithm for feature selection. This algorithm labeled QC-FS: Quantum
Clustering for Feature Selection performs the selection in two steps. Partitioning the original features
space in order to group similar features is performed using the Quantum Clustering algorithm. Then the
selection of a representative for each cluster is carried out. It uses similarity measures such as correlation
coefficient (CC) and the mutual information (MI). The feature which maximizes this information is chosen
by the algorithm
Density Driven Image Coding for Tumor Detection in mri ImageIOSRjournaljce
The significant of multi spectral band resolution is explored towards selection of feature coefficients based on its energy density. Toward the feature representiaon in transformed domain, multi wavelet transformations were used for finer spectral representation. However, due to a large feature count these features are not optimal under low resource computing system. In the recognition units, running with low resources a new coding approach of feature selection, considering the band spectral density is developed. The effective selection of feature element, based on its spectral density achieve two objective of pattern recognition, the feature coefficient representiaon is minimized, hence leading to lower resource requirement, and dominant feature representation, resulting in higher retrieval performance.
Propagation of Error Bounds due to Active Subspace ReductionMohammad
This document summarizes the propagation of error bounds due to active subspace reduction in computational models. It presents two algorithms for performing active subspace reduction: one that is gradient-free and reduces the response or state space, and one that is gradient-based and reduces the parameter space. It then develops a theorem for propagating error bounds across multiple reductions, both in the parameter and response spaces. Numerical experiments on an analytic function and a nuclear reactor pin cell model are used to validate the error bound approach.
A Novel Algorithm for Design Tree Classification with PCAEditor Jacotech
This document summarizes a research paper titled "A Novel Algorithm for Design Tree Classification with PCA". It discusses dimensionality reduction techniques like principal component analysis (PCA) that can improve the efficiency of classification algorithms on high-dimensional data. PCA transforms data to a new coordinate system such that the greatest variance by any projection of the data comes to lie on the first coordinate, called the first principal component. The paper proposes applying PCA and linear transformation on an original dataset before using a decision tree classification algorithm, in order to get better classification results.
This document summarizes a research paper titled "A Novel Algorithm for Design Tree Classification with PCA". It discusses dimensionality reduction techniques like principal component analysis (PCA) that can improve the efficiency of classification algorithms on high-dimensional data. PCA transforms data to a new coordinate system such that the greatest variance by any projection of the data comes to lie on the first coordinate, called the first principal component. The paper proposes applying PCA and linear transformation on an original dataset before using a decision tree classification algorithm, in order to get better classification results.
The document discusses an algorithm called Adaptive Multichannel Component Analysis (AMMCA) for separating image sources from mixtures using adaptively learned dictionaries. It begins by reviewing image denoising using learned dictionaries, then extends this to image separation from single mixtures. The key contribution is applying this approach to separating sources from multichannel mixtures by learning local dictionaries for each source during the separation process. The algorithm is described and simulated results are shown separating two images from a noisy mixture using the learned dictionaries. In conclusion, AMMCA is able to separate sources without prior knowledge of their sparsity domains by fusing dictionary learning into the separation process.
Abstract This paper presents a novel approach for the gesture recognition system using software. In this paper the real time image is taken and is compared with a training set of images and displays a matched training image. In this approach we have used skin detection techniques for detecting the skin threshold regions, Principle Component Analysis (PCA) algorithm and Linear Discriminant Analysis (LDA) for data compressing and analyzing and K-Nearest Neighbor (KNN), Support Vector Machine (SVM) classification for matching the appropriate training image to the real-time image. The software used is MATLAB. The hand gestures used are taken from the American Sign Language. Keywords— PCA algorithm, LDA algorithm, skin detection, KNN and SVM classification
This document provides an overview of knowledge representation techniques and object recognition. It discusses syntax and semantics in representation, as well as descriptions, features, grammars, languages, predicate logic, production rules, fuzzy logic, semantic nets, and frames. It then covers statistical and cluster-based pattern recognition methods, feedforward and backpropagation neural networks, unsupervised learning including Kohonen feature maps, and Hopfield neural networks. The goal is to represent knowledge in a way that enables object classification and decision-making.
This document provides an overview of knowledge representation techniques and object recognition. It discusses syntax and semantics in representation, as well as descriptions, features, grammars, languages, predicate logic, production rules, fuzzy logic, semantic nets, and frames. It then covers statistical and cluster-based pattern recognition methods, feedforward and backpropagation neural networks, unsupervised learning including Kohonen feature maps, and Hopfield neural networks. The goal is to represent knowledge in a way that enables object classification and decision-making.
This document discusses single object tracking and velocity determination. It begins with an introduction and objectives of the project which is to develop an algorithm for tracking a single object and determining its velocity in a sequence of video frames. It then provides details on preprocessing techniques like mean filtering, Gaussian smoothing and median filtering to reduce noise. It describes segmentation methods including histogram-based, single Gaussian background and frame difference approaches. Feature extraction methods like edges, bounding boxes and color are explained. Object detection using optical flow and block matching is covered. Finally, it discusses tracking and calculating velocity of the moving object. MATLAB is introduced as a technical computing language for solving these types of problems.
Gesture Recognition using Principle Component Analysis & Viola-Jones AlgorithmIJMER
Gesture recognition pertains to recognizing meaningful expressions of motion by a human,
involving the hands, arms, face, head, and/or body. It is of utmost importance in designing an intelligent
and efficient human–computer interface. The applications of gesture recognition are manifold, ranging
from sign language through medical rehabilitation to virtual reality. In this paper, we provide a survey on
gesture recognition with particular emphasis on hand gestures and facial expressions. Applications
involving wavelet transform and principal component analysis for face and hand gesture recognition on
digital images
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...IRJET Journal
This document discusses machine learning algorithms for image classification using five different classification schemes. It summarizes the mathematical models behind each classification algorithm, including Nearest Class Centroid classifier, Nearest Sub-Class Centroid classifier, k-Nearest Neighbor classifier, Perceptron trained using Backpropagation, and Perceptron trained using Mean Squared Error. It also describes two datasets used in the experiments - the MNIST dataset of handwritten digits and the ORL face recognition dataset. The performance of the five classification schemes are compared on these datasets.
A systematic image compression in the combination of linear vector quantisati...eSAT Publishing House
1) The document presents a method for image compression that combines linear vector quantization and discrete wavelet transform.
2) Linear vector quantization is used to generate codebooks and encode image blocks, achieving better PSNR and MSE than self-organizing maps.
3) The encoded blocks are then subjected to discrete wavelet transform. Low-low subbands are stored for reconstruction while other subbands are discarded.
4) Experimental results show the proposed method achieves higher PSNR and lower MSE than existing techniques, preserving both texture and edge information.
This document describes a novel statistical damage detection approach using unsupervised support vector machines (SVM). It aims to identify damage in structural components through vibration-based methods. The proposed approach builds a statistical model through unsupervised learning, avoiding the need for measurements from damaged structures. It is computationally efficient even with large numbers of features and does not suffer from local minima problems like artificial neural networks. Numerical simulations show the approach can accurately detect both the occurrence and location of damage.
This document describes a novel statistical damage detection approach using unsupervised support vector machines (SVM). It begins with an introduction to vibration-based damage detection methods and their limitations. It then provides an overview of SVM and describes how one-class SVM and support vector regression can be applied for unsupervised learning in damage detection, without requiring data from damaged structures. Feature selection is discussed to improve computational efficiency when dealing with large datasets. Numerical simulations are analyzed to examine the accuracy and scalability of the proposed approach.
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
Literature Review Basics and Understanding Reference Management.pptxDr Ramhari Poudyal
Three-day training on academic research focuses on analytical tools at United Technical College, supported by the University Grant Commission, Nepal. 24-26 May 2024
KuberTENes Birthday Bash Guadalajara - K8sGPT first impressionsVictor Morales
K8sGPT is a tool that analyzes and diagnoses Kubernetes clusters. This presentation was used to share the requirements and dependencies to deploy K8sGPT in a local environment.
Understanding Inductive Bias in Machine LearningSUTEJAS
This presentation explores the concept of inductive bias in machine learning. It explains how algorithms come with built-in assumptions and preferences that guide the learning process. You'll learn about the different types of inductive bias and how they can impact the performance and generalizability of machine learning models.
The presentation also covers the positive and negative aspects of inductive bias, along with strategies for mitigating potential drawbacks. We'll explore examples of how bias manifests in algorithms like neural networks and decision trees.
By understanding inductive bias, you can gain valuable insights into how machine learning models work and make informed decisions when building and deploying them.
Low power architecture of logic gates using adiabatic techniquesnooriasukmaningtyas
The growing significance of portable systems to limit power consumption in ultra-large-scale-integration chips of very high density, has recently led to rapid and inventive progresses in low-power design. The most effective technique is adiabatic logic circuit design in energy-efficient hardware. This paper presents two adiabatic approaches for the design of low power circuits, modified positive feedback adiabatic logic (modified PFAL) and the other is direct current diode based positive feedback adiabatic logic (DC-DB PFAL). Logic gates are the preliminary components in any digital circuit design. By improving the performance of basic gates, one can improvise the whole system performance. In this paper proposed circuit design of the low power architecture of OR/NOR, AND/NAND, and XOR/XNOR gates are presented using the said approaches and their results are analyzed for powerdissipation, delay, power-delay-product and rise time and compared with the other adiabatic techniques along with the conventional complementary metal oxide semiconductor (CMOS) designs reported in the literature. It has been found that the designs with DC-DB PFAL technique outperform with the percentage improvement of 65% for NOR gate and 7% for NAND gate and 34% for XNOR gate over the modified PFAL techniques at 10 MHz respectively.
We have compiled the most important slides from each speaker's presentation. This year’s compilation, available for free, captures the key insights and contributions shared during the DfMAy 2024 conference.
6th International Conference on Machine Learning & Applications (CMLA 2024)ClaraZara1
6th International Conference on Machine Learning & Applications (CMLA 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of on Machine Learning & Applications.
A review on techniques and modelling methodologies used for checking electrom...nooriasukmaningtyas
The proper function of the integrated circuit (IC) in an inhibiting electromagnetic environment has always been a serious concern throughout the decades of revolution in the world of electronics, from disjunct devices to today’s integrated circuit technology, where billions of transistors are combined on a single chip. The automotive industry and smart vehicles in particular, are confronting design issues such as being prone to electromagnetic interference (EMI). Electronic control devices calculate incorrect outputs because of EMI and sensors give misleading values which can prove fatal in case of automotives. In this paper, the authors have non exhaustively tried to review research work concerned with the investigation of EMI in ICs and prediction of this EMI using various modelling methodologies and measurement setups.
A review on techniques and modelling methodologies used for checking electrom...
Conference_paper.pdf
1. PREDICTION OF COVID-19 USING
MACHINE LEARNING TECHNIQUES
Narenraj Vivekanandan
Dept. of Electrical Engineering
National Institute of Technology
Calicut, India
naren.raj.vivek7@gmail.com
Mohamed Ashiq Rahman S
Dept. of Electrical Engineering
National Institute of Technology
Calicut, Inda
ashiqxq@gmail.com
Vedant Mahalle
Dept. of Electrical Engineering
National Institute of Technology
Calicut, India
vedantmahalle21@gmail.com
Sharath M Nair
Dept. of Electrical Engineering
National Institute of Technology
Calicut, India
sharathmappu99@gmail.com
Abstract—Numerous techniques have been proposed by
WHO and other esteemed medical authorities for the diagnosis of
the COVID-19 virus. The most popular diagnostic method is
Reverse transcription polymerase chain reaction (RT –PCR).
Other clinical diagnosis techniques involve antibody tests. There
has been other research focused on classifying the covid vs
non-covid classification using chest x-ray images. However, many
of these classification is done over images that account for
increased overfitting. We propose a different model that employs
wavelet entropy to extract features from and then classify the
chest x-ray images. The proposed technique extracts space
frequency features from chest x-ray images using Discrete
Wavelet Transform, the dimensionality of which is reduced using
Shannon entropy technique, and resulting vector is trained using
Standard machine learning classifiers such as Logistic
Regression, Support Vector Machine, Decision Tree classifier,
Gaussian Naïve bayes and Convolutional Neural Network
Keywords— wavelet transform, entropy, logistic regression,
naive bayes, decision tree, support vector machine
I. INTRODUCTION
Rapid and reliable diagnosing of COVID-19 is one of the
foremost challenges we face today. This is most important for
those who may be critical and need medical care. The main
effect of SARS-COV-2 or COVID-19 is that it affects the
lungs of the infected person. The most common effects of the
virus is that it causes severe respiratory illness and pneumonia.
These effects can be commonly diagnosed with the
examination of Chest X-Ray(CXR) images. Previous studies
have shown that machine learning models are much more
accurate and better in reading X-ray images than a human eye.
Diagnosing COVID-19 with CXR images is much more
reliable and rapid than RT-PCR test or Antigen tests. We have
built a machine learning model which can help the medical
community with speedy diagnosis of COVID-19 with the use
of CXR images using a pretrained model. We will use
Discrete Wavelet Transform(DWT) for feature extraction as
studies have shown that wavelet transforms are excellent in
detecting edges and distinguishing frequencies. We will
further use different classifiers to train our model such as
Logistic Regression, Support Vector Machine , Decision Tree
Classifier and Naive Bayes and study the results. Also, we will
examine the effect of using CNN to classify our images
without feature extraction.
II. BASIC CONCEPTS
A. Discrete Wavelet Transform
The discrete wavelet transform (DWT) is used to get the
multi-scale (frequency) representation of the function.Using
wavelets, the image data can be analyzed in multiple
resolutions. Wavelet transformation is better at capturing fine
details because of the high frequency components. The 1D
-DWT of signal x is calculated by passing it through a pair of
high and low pass filters (quadrature mirror filters) with
impulse response h, and g respectively
Fig 1. Filter representation of Wavelet Transform.
The Approximation coefficients is represented by
[k]g[2n ]
Y low = ∑
∞
k=−∞
x − k (2)
The detail coefficients are represented as
[k]h[2n ]
Y high = ∑
∞
k=−∞
x − k (3)
2. At every decomposed level since half of the frequency is
discarded, half of the samples can be discarded as well as per
the Nyquist criterion.
1 dimensional discrete wavelet transform (1D-DWT) can
be extended to (2D-DWT) by processing along the x and y
axis using low pass filters (expanded wavelets) and high pass
filters (shrunken wavelets). Four sub-band of images (HH1,
LH1, HL1, LL1) at each scale will be generated after the
level-1 decomposition. The A1 sub-band containing the
low-frequency components can be regarded as the
approximation component of the image.
Fig 2. 2-level wavelet decomposition
while the LH, HL, and HH sub-bands, which contain
relatively higher frequency portions of the image, have the
more detailed components of the image.
Working over the assumption that most of the image data is
contained in the LL1 sub-band, it can be further decomposed
to level-2 thus arriving at 7 sub-bands (HH1, HL2, LH2, HH2,
HL1, LH1, HH1)
B. Entropy
The major disadvantage of the Discrete wavelet transform
technique is the curse of dimensionality. Too many features
results in increased computation times and excessive storage
memory. To overcome this disadvantage, we have to reduce
the number of coefficients, thus we employ an additional
parameter, entropy, to reduce the dimension by averaging out
the inter-related variables while maintaining the sufficient
information. In information theory entropy is the minimum
limit to which you can compress an information without loss.
Shannon defined that the entropy H for a discrete random
variable X with values {x1, x2, … xn}and probability mass
function P(X) as:
H(X) = - (4)
log x
∑
n
i=0
xi b i
Shanon’s entropy thus quantifies the amount of
information available in a variable. It’s metric is defined as
the absolute minimum amount of storage required to
succinctly capture any information.
C. Feature Extraction
For a 256*256 image there can be 65536 coefficients
however with the inclusion of entropy parameter, the number
of features can be reduced to 7 entropy vectors with each
vector corresponding to a sub-band after 2-level 2D wavelet
transform of the image. This can be computationally efficient.
III. MACHINE LEARNING MODELS
A. Naive Bayes
Naive Bayes is a family of probabilistic algorithms that use
probability theory and the theorem of Bayes to predict an
event. They are probabilistic, meaning that they measure for a
given data the likelihood of each label, and then output the
label with the highest one. Using Bayes' Theorem, which
defines the likelihood of feature, is the way they get these
probabilities, based on previous knowledge of what could be
relevant to that feature.
Abstractly, Naive Bayes is a Conditional Probability
model: We are given a problem sample X to be classified,
where
{x , x , x ......, x }
X = 1 2 3 n (5)
Where X represents n features (independent variables).
The probability estimated from the model will be a dependent
class C with a small number of outcomes (Covid positive/
negative here) conditional on feature vector X.
(6)
(C |x , , x ......, x )
P K 1 x2 3 n
Here if a feature can take on a large number of values, or
the number of features n is large, then basing such a model on
Probability tables is impractical. Thus using Bayes’ Theorem,
the conditional probability can be reduced to
(C |x)
p k =
p(x)
p(C )p(x|C )
k K
(7)
Thus the posterior probability is formed combining both
sources of information, the prior and the likelihood. Since the
features are known beforehand, the denominator is a constant
and is not considered in practice.
Now considering the conditional independence of the
features i.e since each feature Xi is independent, the joint
model can be expressed as
(8)
(C | x , , x ......, x ) ∝p(C ) (x |C )
P K 1 x2 3 n k ∏
n
i=1
p i k
Where P(xi | Ck) can be estimated using the training
sample.
B. Logistic Regression
The name logistic regression comes from the logistic
function or the sigmoid function used as the activation
function. The sigmoid function has a range of 0 to 1 thus it is
widely used in models that require a probability estimate as an
output.
Logistic regression is a statistical model that in its basic
form uses a logistic function to model a binary dependent
variable. In regression the parameters corresponding to most
accurate probability is estimated.
Let X be an n*d dimensional matrix. Here n is the number
of samples and d is the number of features or independent
attributes, and y be a binary outcomes vector. y is a
n*1dimensional matrix which corresponds to the labels for
each 1*d data in X
3. A linear model to describing this problem would be of
form
(9)
W X
Z = T
+ B
(10)
(z)
y
︿
= a = σ
(11)
(a, y) loga 1 )log(1 )
L = − y + ( − y − a
Where a is the sigmoid of z and represents the probability
of a class to occur given a data in X and y is the ground truth
(0 or 1). L is the loss function which is a relationship between
y and a, and the objective of the regression is to estimate the
parameter vectors w and b to minimise the Loss function as
much as possible. This can be done using Gradient Descent.
In gradient descent, we reduce the parameters w and b by
dw and db until the optimal parameters are achieved. Here dw
is the derivative of the loss function with respect to the
parameter w and db is the derivative of the loss function with
respect to the parameter b. Here,
dz = a - y (12)
dw = x*dz (13)
db = dz (14)
w = w - *dw (15)
b = b - *db (16)
Where 𝜶 is the learning rate of the algorithm.
C. Decision Tree Classifier
Decision tree algorithm is from a class of supervised machine
learning algorithms. The goal of the classifier is to create an
optimal decision tree from the given set of features and labels
so that it can predict the label of a new set of features by
iterating down the decision tree.
A decision tree consists of a root node (which is the best
predictor) , a set of inner nodes and leaf nodes. Leaf nodes
correspond to different classes the dataset belongs to, whereas
the root node and the inner nodes correspond to the features
extracted from the dataset.
The performance of the classifier depends on how good the
tree is constructed from the training data. The process of
building a decision tree is recursive. It begins from the root
node and continues to split the dataset into many subsets
depending on the number of classes. The features which best
predicts a particular sub dataset takes the place in that
particular inner node in the tree.
A common metric to measure which feature is the best
predictor of a sub dataset is the Gini impurity of that sub
dataset. Gini impurity measures how often a random element
from the dataset would be mis-classfied if it was randomly
labeled according to the distribution of classes in the subset.
The Gini impurity can be calculated by summing the
probability of class i being chosen times the probability of
pi
misclassifying that item which is .
1 − pi
To compute the Gini impurity of a sub dataset with J classes
(p) (p ) (1 )
G = ∑
J
i=1
i ∑
k=i
/
pk = ∑
J
i=1
pi − pi = ∑
J
i−1
pi − ∑
J
i=1
pi
2
Hence, (p)
G = 1 − ∑
J
i=1
pi
2
(17)
Where can be estimated in each sub dataset.
pi
D. Support Vector Machine
The Support Vector Machine (SVM) is a machine learning
classifier that takes a multi-dimensional data vector and the
class/label they belong to and establishes a boundary called
the decision boundary between the various classes, so that it is
simple to identify new data by inspecting the boundary it falls
within.
However, depending on the parameters a maximum margin
classifier may not always lead to an optimal decision boundary
as, if there are errors on either side of the boundary the
boundary may be very close to some data points. Hence, it is
important to sometimes allow misclassifications to find the
optimal boundary. Such a classifier that allows some
misclassification to find the most optimal boundary with
maximum margin is called a soft margin classifier or a support
vector classifier.
Mathematically, the aim of support vector machine is to
minimize in relation with eq.(18) and subject to eq.(19)
|w|
2
1 2
X
Y = WT
+ B (18)
Y < , x − |
| i − w i > b ≤ ε (19)
Again, a linear support vector classifier may not always be
optimal in the case of a dataset with complex features. Hence
different kernel functions exist using which we can find the
maximal margin hyperplane. Some of the more common
kernels are linear kernels, polynomial kernels and RBF
kernels. Kernels like polynomial kernel work in higher
dimensions to find the best support vector classifier while
radial basis function (RBF) also known as Gaussian kernels
are functions that are based on the absolute distance from a
data point (r = ||x−xi||) . The RBF kernel between two data
points,x and x′ is defined by
(x, x ) e
K ′ = −γ||x−x ||
′
2
(20)
Where is the Euclidean distance, γ is a parameter
||x ||
− x′ 2
specified and K(x,x′) is given as a feature vector.
4. IV. CONVOLUTIONAL NEURAL NETWORKS
A Convolutional Neural Network is a deep learning neural
network that is used to analyze visual imagery. It consists of
several layers in the order: input layer, hidden
layers(convolution layers, pooling layers and fully connected
layers) and output layer. ConvNet learns the features by
applying appropriate kernel filters. As the parameters are
decreased and weights updated, the network is able to
generalise very well on the image dataset. Its work is to ensure
that the images are in form that is easily handled, without
compromising the features which are essential for obtaining an
accurate prediction.
The convolution operation is a mathematical operation
applied on the input images to capture the high-level features
such as edges.A Pooling Layer almost always follows a
Convolutional layer and is used to reduce spatial size of the
matrix. It also employs dimensionality reduction to efficiently
lower the computational power necessary for model training.
By applying the above techniques, we have a convolved
matrix which understands several features from the images
fed. We will now flatten the matrix and employ a neural
network for classification.The flattened matrix has values
which are non-linear combinations and in order to learn these
combinations and make accurate predictions, we use a
Fully-Connected layer which in this case is a multi-level
perceptron. Backpropagation is applied to every iteration of
training. After some epochs, the model classifies the image
into two classes using the Sigmoid Classifier.
V. CLASSIFICATION & COMPARISON
A. Dataset
We used the publicly available CovidX dataset Covid-Net
Open Source Initiative by Linda Wang, Alexander Wong from
Department of Systems Design Engineering, University of
Waterloo, Canada. This is a standard and labelled dataset. This
dataset contains 14904 Non-Covid images and 594 Covid
images
Fig 3. Covid -ve CXR images
The images were read and converted to integer representation
using cv2 module, the obtained values were scaled uniformly
to avoid zero values that may lead to division by zero
scenarios. The images were then transformed to a 7 feature
vector using DWT and entropy.
Fig 4: Covid +ve CXR images
B. Result and Analysis
Choosing the appropriate parameters is essential to
arriving at the best classification model, for which we used
hyper-parameter tuning techniques to validate our models at
different parameter values. The Naive Bayes classifier turned
out to be independent of the major parameters such as prior
probability, the Logistic regression performed better with the
penalty set as ‘l2’ which uses ridge method, and solver set as
‘newton-cg’ that uses second order derivatives to arrive at
optimization. The DTC performed the best with the criterion
parameter set to ‘entropy’ as compared to ‘gini’. The SVM
showed the best with parameter ‘C’ set to 63, this parameter is
inversely proportional to the proportion of mis-classification
allowed, in SVM the kernel was set to ‘RBF’ as expected -
allowing classification to work in infinite dimension, the
gamma, that defines the curvature of rbf kernel is set to 0.001
thus allowing less curvature.
We compared the features obtained from DWT+entropy
technique using Decision Tree classifier, Logistic Regression
Classifier, Naive Bayes Classifier and Support Vector
Machine; The classification parameters were obtained from a
method of Hyper-parameter tuning. Alternatively, we used the
image directly without any other feature extraction in the
Convolutional Neural Network based classifiers.
TABLE I. C
LASSIFICATION C
OMPARISON
Feature Precision
Score
Recall
Score
F1-Score Accuracy
CNN NA NA NA 83.44%
DWT+ENTRO
PY+SVM
0.9162 0.867 0.8854 99.13%
DWT+DTC 0.855 0.8538 0.8494 98.85%
DWT+LRC 0.909 0.4298 0.5556 97.59%
DWT+NBC 0.907 0.7932 0.8414 98.86%
5. The f1-score is chosen as the appropriate classification
metric since we were dealing with an imbalanced dataset.
As evident from the scores given in Table I. the support
vector machine classifier did the best job at classification, with
a mean f1-score of 0.8854. The support vector machine came
ahead in all other classification metrics as well.
The logistic regression performed worse with a ‘F1-score’
of 0.5556 which is marginally better than random prediction,
this underwhelming performance can be attributed to the
linearly inseparable nature of the feature set, which the logistic
regression cannot classify
VI. CONCLUSION
.In this paper, we compared ML classification algorithms
to accurately predict covid-19 using the feature set extracted
from wavelet entropy.
Although the entropy values and other hyper-parameters
used in the classification are difficult to interpret, the proposed
method using SVM has good classification results. The
classification metrics can be improved by training with more
images, and more robust hyperparameter tuning, alternatively
we can use techniques other than entropy as a dimensionality
reduction measure.
The model can be further improved to accommodate more
diseases that can be diagnosed using CXR images thus in
future we can improve the model to a multi-disease
classification model. .
ACKNOWLEDGMENT
The work was done under the guidance of Dr. Shihabudeen
K.V, Assistant professor at National Institute of Technology,
Calicut.
REFERENCES..
[1] Sun, Da-Wen. (2008). Computer Vision Technology for Food Quality
Evaluation. 10.1016/B978-0-12-373642-0.X5001-7.
[2] Zhou, Xing-Xing & Zhang, Yu-Dong & ji, Genlin & Yang, Jiquan &
Dong, Zhengchao & Wang, Shuihua & Zhang, Guangshuai & Phillips,
Preetha. (2016). Detection of abnormal MR brains based on wavelet
entropy and feature selection. IEEJ Transactions on Electrical and
Electronic Engineering. 11. n/a-n/a. 10.1002/tee.22226.
[3] Akshay Iyer, Akanksha Pandey, Dipti Pamnani, Karmanya Pathak and
Prof. Mrs. Jayshree Hajgude “Email Filtering and Analysis Using
Classification Algorithms” IJCSI International Journal of Computer
Science Issues, Vol. 11, Issue 4, No 1, July 2014
[4] Joaquim de Moura, Jorge Novo, Marcos Ortega. "Fully automatic deep
convolutional approaches for the analysis of Covid-19 using chest X-ray
images", Cold Spring Harbor Laboratory, 2020
[5] Sohaib Asif, Yi Wenhui. "Automatic Detection of COVID-19 Using
X-ray Images with Deep Convolutional Neural Networks and Machine
Learning", Cold Spring Harbor Laboratory, 2020
[6] Zhang, Yudong, Shuihua Wang, Preetha Phillips, Zhengchao Dong,
Genlin Ji, and Jiquan Yang. "Detection of Alzheimer's disease and mild
cognitive impairment based on structural volumetric MR images using
3D-DWT and WTA-KSVM trained by PSOTVAC", Biomedical Signal
Processing and Control, 2015.
[7] Jian-Ding Qiu. "Prediction of the Types of Membrane Proteins Based on
Discrete Wavelet Transform and Support Vector Machines", The Protein
Journal, 02/18/2010
[8] "Wavelet-entropy based detection of pathological brain in MRI
scanning", Computer Science and Applications, 2015.
[9]
[10] Maher Maalouf. "Logistic regression in data analysis: an overview",
International Journal of Data Analysis Techniques and Strategies, 2011