This document presents a new system for classifying heart sounds using fractional Fourier transform-based Mel-frequency spectral coefficients (FrFT-MFSC) and traditional machine learning classifiers. The system includes preprocessing, cardiac cycle segmentation, feature extraction using FrFT-MFSC, feature reduction, and classification stages. Several traditional classifiers including SVM, KNN and ensemble classifiers are used and their performances are compared. The system is tested on a publicly available heart sound dataset and achieves a classification score of 0.9200 using an SVM classifier, outperforming previous methods on the same dataset. The methodology and results are discussed to validate the potential of the proposed system for automated heart sound classification.
PHONOCARDIOGRAM HEART SOUND SIGNAL CLASSIFICATION USING DEEP LEARNING TECHNIQUEIRJET Journal
This document describes a study that aimed to classify heart sounds as normal or abnormal using deep learning techniques. The study used phonocardiogram (PCG) signals from online datasets to train convolutional neural network (CNN) and recurrent neural network (RNN) models. Feature extraction using mel-frequency cepstral coefficients was performed on the PCG signals before training the models. Experimental results showed that the CNN model achieved higher accuracy (90.6%) and lower loss than the RNN model, demonstrating that CNN is better suited for this heart sound classification task. The trained CNN model can classify new heart sound recordings with a confidence value indicating the likelihood of the sound being normal or abnormal.
A Study Based On Methods Used In Detection of Cardiac ArrhythmiaIRJET Journal
This document summarizes research on methods for detecting cardiac arrhythmia. It discusses how machine learning and deep learning techniques like convolutional neural networks (CNNs) and long short-term memory networks (LSTMs) have achieved higher accuracy than traditional feature-based machine learning methods for arrhythmia detection. The document reviews several papers that have used techniques like CNNs, LSTMs, oversampling, and focal loss to improve arrhythmia detection. It identifies limitations in relying too heavily on feature extraction and suggests that deep learning methods that learn features automatically can help address these limitations and improve accuracy.
This document summarizes research on using discrete cosine transform (DCT) to extract frequency domain features from electrocardiogram (ECG) signals for classifying cardiac arrhythmias. Features are extracted by computing the distance between RR waves. These frequency domain features are then classified using various soft computing techniques, including classification and regression trees, radial basis function networks, support vector machines, and multilayer perceptron neural networks. Experiments were conducted on the MIT-BIH arrhythmia database to evaluate the performance of these techniques for ECG-based arrhythmia classification.
This document presents research on classifying cardiac arrhythmias using frequency domain features extracted from electrocardiogram (ECG) signals. Features are extracted from ECG data using Discrete Cosine Transform to calculate the distance between RR waves. These frequency domain features are then classified using various machine learning algorithms, including Classification and Regression Trees, Radial Basis Function networks, Support Vector Machines, and Multilayer Perceptron Neural Networks. Experiments were conducted on the MIT-BIH arrhythmia database to evaluate the performance of the different classifiers.
The document discusses a hybrid CNN and LSTM network for heart disease prediction. It begins by noting heart disease is a leading cause of death worldwide and that traditional machine learning methods have achieved only 65-85% accuracy in prediction. The proposed method uses a convolutional neural network to extract features from heart disease data, and then an LSTM network to classify the data as normal or abnormal based on those features. When tested on heart disease data, the hybrid model achieved 89% accuracy, outperforming other machine learning algorithms like SVM, Naive Bayes and decision trees.
The clinical indication of arrhythmia identifies specific aberrant circumstances in heart pumping that may be detected using electrical impulses during conduction or by allowing a little amount of current to travel through the electrodes, disrupting the cardiac muscle's resistance. The electrocardiogram (ECG) is one of the most important instruments for detecting cardiac arrhythmia since it is the most least intrusive and effective procedure. Physically or visually inspecting the heart is time-consuming and difficult, hence the development of computer aided diagnosis (CAD) is being developed to aid clinical decision-making. In this suggested research, a convolutional neural network (CNN)-based approach is used to automate the heartbeat classification process in order to identify cardiac arrhythmia. The improved enhancement of CNN structure has been implemented in this suggested research. The feature maps are then subjected to the max pooling process. Finally, feature maps are generated by concatenating kernels of different sizes and delivering them as an input to the fully linked layers. The MIT BIH arrhythmia database is used to implement this approach, and the total average accuracy is 99.21%. The proof of the suggested study's efficiency and efficacy in identifying cardiac arrhythmia has also been done via an experimental comparison.
An optimal heart disease prediction using chaos game optimization‑based recur...BASMAJUMAASALEHALMOH
The document proposes a Chaos Game Optimization based Recurrent Neural Network (CGO-RNN) model for predicting heart disease. It uses Kernel Principal Component Analysis (KPCA) for dimensionality reduction and feature extraction. The CGO algorithm tunes the hyperparameters of the RNN model to improve accuracy. The model is evaluated on heart disease datasets and achieves prediction accuracies of 98.99-98.54%, demonstrating improved performance over other methods.
A REVIEW PAPER ON PULMONARY NODULE DETECTIONIRJET Journal
This document reviews different techniques for pulmonary nodule detection in CT scans using deep learning. It summarizes several papers that have used techniques like convolutional neural networks (CNNs), 3D CNNs, and customized mixed link networks to develop computer-aided diagnosis systems for detecting and classifying lung nodules. These papers report accuracy rates from 85.7% to 98.7% and sensitivities from 80.06% to 94% depending on the specific deep learning approach and dataset used. The document concludes by comparing the performance of these different papers.
PHONOCARDIOGRAM HEART SOUND SIGNAL CLASSIFICATION USING DEEP LEARNING TECHNIQUEIRJET Journal
This document describes a study that aimed to classify heart sounds as normal or abnormal using deep learning techniques. The study used phonocardiogram (PCG) signals from online datasets to train convolutional neural network (CNN) and recurrent neural network (RNN) models. Feature extraction using mel-frequency cepstral coefficients was performed on the PCG signals before training the models. Experimental results showed that the CNN model achieved higher accuracy (90.6%) and lower loss than the RNN model, demonstrating that CNN is better suited for this heart sound classification task. The trained CNN model can classify new heart sound recordings with a confidence value indicating the likelihood of the sound being normal or abnormal.
A Study Based On Methods Used In Detection of Cardiac ArrhythmiaIRJET Journal
This document summarizes research on methods for detecting cardiac arrhythmia. It discusses how machine learning and deep learning techniques like convolutional neural networks (CNNs) and long short-term memory networks (LSTMs) have achieved higher accuracy than traditional feature-based machine learning methods for arrhythmia detection. The document reviews several papers that have used techniques like CNNs, LSTMs, oversampling, and focal loss to improve arrhythmia detection. It identifies limitations in relying too heavily on feature extraction and suggests that deep learning methods that learn features automatically can help address these limitations and improve accuracy.
This document summarizes research on using discrete cosine transform (DCT) to extract frequency domain features from electrocardiogram (ECG) signals for classifying cardiac arrhythmias. Features are extracted by computing the distance between RR waves. These frequency domain features are then classified using various soft computing techniques, including classification and regression trees, radial basis function networks, support vector machines, and multilayer perceptron neural networks. Experiments were conducted on the MIT-BIH arrhythmia database to evaluate the performance of these techniques for ECG-based arrhythmia classification.
This document presents research on classifying cardiac arrhythmias using frequency domain features extracted from electrocardiogram (ECG) signals. Features are extracted from ECG data using Discrete Cosine Transform to calculate the distance between RR waves. These frequency domain features are then classified using various machine learning algorithms, including Classification and Regression Trees, Radial Basis Function networks, Support Vector Machines, and Multilayer Perceptron Neural Networks. Experiments were conducted on the MIT-BIH arrhythmia database to evaluate the performance of the different classifiers.
The document discusses a hybrid CNN and LSTM network for heart disease prediction. It begins by noting heart disease is a leading cause of death worldwide and that traditional machine learning methods have achieved only 65-85% accuracy in prediction. The proposed method uses a convolutional neural network to extract features from heart disease data, and then an LSTM network to classify the data as normal or abnormal based on those features. When tested on heart disease data, the hybrid model achieved 89% accuracy, outperforming other machine learning algorithms like SVM, Naive Bayes and decision trees.
The clinical indication of arrhythmia identifies specific aberrant circumstances in heart pumping that may be detected using electrical impulses during conduction or by allowing a little amount of current to travel through the electrodes, disrupting the cardiac muscle's resistance. The electrocardiogram (ECG) is one of the most important instruments for detecting cardiac arrhythmia since it is the most least intrusive and effective procedure. Physically or visually inspecting the heart is time-consuming and difficult, hence the development of computer aided diagnosis (CAD) is being developed to aid clinical decision-making. In this suggested research, a convolutional neural network (CNN)-based approach is used to automate the heartbeat classification process in order to identify cardiac arrhythmia. The improved enhancement of CNN structure has been implemented in this suggested research. The feature maps are then subjected to the max pooling process. Finally, feature maps are generated by concatenating kernels of different sizes and delivering them as an input to the fully linked layers. The MIT BIH arrhythmia database is used to implement this approach, and the total average accuracy is 99.21%. The proof of the suggested study's efficiency and efficacy in identifying cardiac arrhythmia has also been done via an experimental comparison.
An optimal heart disease prediction using chaos game optimization‑based recur...BASMAJUMAASALEHALMOH
The document proposes a Chaos Game Optimization based Recurrent Neural Network (CGO-RNN) model for predicting heart disease. It uses Kernel Principal Component Analysis (KPCA) for dimensionality reduction and feature extraction. The CGO algorithm tunes the hyperparameters of the RNN model to improve accuracy. The model is evaluated on heart disease datasets and achieves prediction accuracies of 98.99-98.54%, demonstrating improved performance over other methods.
A REVIEW PAPER ON PULMONARY NODULE DETECTIONIRJET Journal
This document reviews different techniques for pulmonary nodule detection in CT scans using deep learning. It summarizes several papers that have used techniques like convolutional neural networks (CNNs), 3D CNNs, and customized mixed link networks to develop computer-aided diagnosis systems for detecting and classifying lung nodules. These papers report accuracy rates from 85.7% to 98.7% and sensitivities from 80.06% to 94% depending on the specific deep learning approach and dataset used. The document concludes by comparing the performance of these different papers.
Survey analysis for optimization algorithms applied to electroencephalogramIJECEIAES
This paper presents a survey for optimization approaches that analyze and classify electroencephalogram (EEG) signals. The automatic analysis of EEG presents a significant challenge due to the high-dimensional data volume. Optimization algorithms seek to achieve better accuracy by selecting practical features and reducing unwanted features. Forty-seven reputable research papers are provided in this work, emphasizing the developed and executed techniques divided into seven groups based on the applied optimization algorithm particle swarm optimization (PSO), ant colony optimization (ACO), artificial bee colony (ABC), grey wolf optimizer (GWO), Bat, Firefly, and other optimizer approaches). The main measures to analyze this paper are accuracy, precision, recall, and F1-score assessment. Several datasets have been utilized in the included papers like EEG Bonn University, CHB-MIT, electrocardiography (ECG) dataset, and other datasets. The results have proven that the PSO and GWO algorithms have achieved the highest accuracy rate of around 99% compared with other techniques.
WAVELET SCATTERING TRANSFORM FOR ECG CARDIOVASCULAR DISEASE CLASSIFICATIONgerogepatton
Classifying the ECG dataset is the main technique for diagnosing heart disease. However, the focus of this
field is increasingly on prediction, with a growing dependence on machine learning techniques. This study
aimed to enhance the accuracy of cardiovascular disease classification using data from the PhysioNet
database by employing machine learning (ML). The study proposed several multi-class classification
models that accurately identify patterns within three classes: heart failure rhythm (HFR), normal heart
rhythm (NHR), and arrhythmia (ARR). This was accomplished by utilizing a database containing 162 ECG
signals. The study employed a variety of techniques, including frequency-time domain analysis, spectral
features, and wavelet scattering, to extract features and capture unique characteristics from the ECG
dataset. The SVM model produced a training accuracy of 97.1% and a testing accuracy of 92%. This work
provides a reliable, effective, and human error-free diagnostic tool for identifying heart disease.
Furthermore, it could prove to be a valuable resource for future medical research projects aimed at
improving the diagnosis and treatment of cardiovascular diseases
WAVELET SCATTERING TRANSFORM FOR ECG CARDIOVASCULAR DISEASE CLASSIFICATIONijaia
Classifying the ECG dataset is the main technique for diagnosing heart disease. However, the focus of this
field is increasingly on prediction, with a growing dependence on machine learning techniques. This study
aimed to enhance the accuracy of cardiovascular disease classification using data from the PhysioNet
database by employing machine learning (ML). The study proposed several multi-class classification
models that accurately identify patterns within three classes: heart failure rhythm (HFR), normal heart
rhythm (NHR), and arrhythmia (ARR). This was accomplished by utilizing a database containing 162 ECG
signals. The study employed a variety of techniques, including frequency-time domain analysis, spectral
features, and wavelet scattering, to extract features and capture unique characteristics from the ECG
dataset. The SVM model produced a training accuracy of 97.1% and a testing accuracy of 92%. This work
provides a reliable, effective, and human error-free diagnostic tool for identifying heart disease.
Furthermore, it could prove to be a valuable resource for future medical research projects aimed at
improving the diagnosis and treatment of cardiovascular diseases.
WAVELET SCATTERING TRANSFORM FOR ECG CARDIOVASCULAR DISEASE CLASSIFICATIONgerogepatton
Classifying the ECG dataset is the main technique for diagnosing heart disease. However, the focus of this
field is increasingly on prediction, with a growing dependence on machine learning techniques. This study
aimed to enhance the accuracy of cardiovascular disease classification using data from the PhysioNet
database by employing machine learning (ML). The study proposed several multi-class classification
models that accurately identify patterns within three classes: heart failure rhythm (HFR), normal heart
rhythm (NHR), and arrhythmia (ARR). This was accomplished by utilizing a database containing 162 ECG
signals. The study employed a variety of techniques, including frequency-time domain analysis, spectral
features, and wavelet scattering, to extract features and capture unique characteristics from the ECG
dataset. The SVM model produced a training accuracy of 97.1% and a testing accuracy of 92%. This work
provides a reliable, effective, and human error-free diagnostic tool for identifying heart disease.
Furthermore, it could prove to be a valuable resource for future medical research projects aimed at
improving the diagnosis and treatment of cardiovascular diseases.
DETECTION OF CRACKLES AND WHEEZES IN LUNG SOUND USING TRANSFER LEARNING hiij
In recent years, deep learning models have improved how well various diseases, particularly respiratory
ailments, can be diagnosed. In order to assist in offering a diagnosis of respiratory pathologies in digitally
recorded respiratory sounds, this research will provide an evaluation of the effectiveness of several deep
learning models connected with the raw lung auscultation sounds in detecting respiratory pathologies. We
will also determine which deep learning model is most appropriate for this purpose. With the development
of computer -systems that can collect and analyze enormous volumes of data, the medical profession is
establishing several non-invasive tools. This work attempts to develop a non-invasive technique for
identifying respiratory sounds acquired by a stethoscope and voice recording software via machine
learning techniques. This study suggests a trained and proven CNN-based approach for categorizing
respiratory sounds. A visual representation of each audio sample is constructed, allowing resource
identification for classification using methods like those used to effectively describe visuals. We used a
technique called Mel Frequency Cepstral Coefficients (MFCCs). Here, features are retrieved and
categorized via VGG16 (transfer learning) and prediction is accomplished using 5-fold cross-validation.
Employing various data splitting techniques, Respiratory Sound Database obtained cutting-edge results,
including accuracy of 95%, precision of 88%, recall score of 86%, and F1 score of 81 %. We trained and
tested the model using a sound database made by the International Conference on Biomedical and Health
Informatics (ICBHI) in 2017 and annotated by experts with a classification of the lung sound.
DETECTION OF CRACKLES AND WHEEZES IN LUNG SOUND USING TRANSFER LEARNING hiij
This document presents research on detecting crackles and wheezes in lung sounds using transfer learning. The researchers used the VGG16 convolutional neural network to extract features from lung sound spectrograms preprocessed using Mel Frequency Cepstral Coefficients. They trained and tested the model on the ICBHI 2017 Respiratory Sound Database, achieving 95% accuracy, 88% precision, 86% recall, and 81% F1 score. The researchers compared their results to other CNN and RNN models to determine the optimal model for lung sound classification.
PREDICTIVE ANALYTICS IN HEALTHCARE SYSTEM USING DATA MINING TECHNIQUEScscpconf
The health sector has witnessed a great evolution following the development of new computer technologies, and that pushed this area to produce more medical data, which gave birth to multiple fields of research. Many efforts are done to cope with the explosion of medical data on one hand, and to obtain useful knowledge from it on the other hand. This prompted researchers to apply all the technical innovations like big data analytics, predictive analytics, machine learning and learning algorithms in order to extract useful knowledge and help in making decisions. With the promises of predictive analytics in big data, and the use of machine learning
algorithms, predicting future is no longer a difficult task, especially for medicine because predicting diseases and anticipating the cure became possible. In this paper we will present an overview on the evolution of big data in healthcare system, and we will apply a learning algorithm on a set of medical data. The objective is to predict chronic kidney diseases by using Decision Tree (C4.5) algorithm.
Predicting heart failure using a wrapper-based feature selectionnooriasukmaningtyas
In the current health system, it is very difficult for medical practitioners/ physicians to diagnose the effectiveness of heart contraction. In this research, we proposed a machine learning model to predict heart contraction using an artificial neural network (ANN). We also proposed a novel wrapper-based feature selection utilizing a grey wolf optimization (GWO) to reduce the number of required input attributes. In this work, we compared the results achieved using our method and several conventional machine learning algorithms approaches such as support vector machine, decision tree, Knearest neighbor, naïve bayes, random forest, and logistic regression. Computational results show not only that much fewer features are needed, but also higher prediction accuracy can be achieved around 87%. This work has the potential to be applicable to clinical practice and become a supporting tool for doctors/physicians.
This document presents a novel deep learning approach for single-lead electrocardiogram (ECG) classification. The approach uses Restricted Boltzmann Machines (RBM) and Deep Belief Networks (DBN) for ECG classification after detecting ventricular and supraventricular heartbeats from single-lead ECG signals. Experimental results on the MIT-BIH database show the approach achieves high average recognition accuracies of 93.63% for ventricular ectopic beats and 95.57% for supraventricular ectopic beats at a low sampling rate of 114 Hz, outperforming traditional methods.
Attention gated encoder-decoder for ultrasonic signal denoisingIAESIJAI
Ultrasound imaging is one of the most widely used non-destructive testing methods. The transducer emits pulses that travel through the imaged samples and are reflected by echo-forming impedance. The resulting ultrasonic signals usually contain noise. Most of the traditional noise reduction algorithms require high skills and prior knowledge of noise distribution, which has a crucial impact on their performances. As a result, these methods generally yield a loss of information, significantly influencing the final data and deeply limiting both sensitivity and resolution of imaging devices in medical and industrial applications. In the present study, a denoising method based on an attention-gated convolutional autoencoder is proposed to fill this gap. To evaluate its performance, the suggested protocol is compared to widely used methods such as butterworth filtering (BF), discrete wavelet transforms (DWT), principal component analysis (PCA), and convolutional autoencoder (CAE) methods. Results proved that better denoising can be achieved especially when the original signal-to-noise ratio (SNR) is very low and the sound waves’ traces are distorted by noise. Moreover, the initial SNR was improved by up to 30 dB and the resulting Pearson correlation coefficient was maintained over 99% even for ultrasonic signals with poor initial SNR.
A novel framework for efficient identification of brain cancer region from br...IJECEIAES
Diagnosis of brain cancer using existing imaging techniques, e.g., Magnetic Resonance Imaging (MRI) is shrouded with various degrees of challenges. At present, there are very few significant research models focusing on introducing some novel and unique solutions towards such problems of detection. Moreover, existing techniques are found to have lesser accuracy as compared to other detection schemes. Therefore, the proposed paper presents a framework that introduces a series of simple and computationally cost-effective techniques that have assisted in leveraging the accuracy level to a very higher degree. The proposed framework takes the input image and subjects it to non-conventional segmentation mechanism followed by optimizing the performance using directed acyclic graph, Bayesian Network, and neural network. The study outcome of the proposed system shows the significantly higher degree of accuracy in detection performance as compared to frequently existing approaches.
The biomedical profession has gained importance due to the rapid and accurate diagnosis of clinical patients using computer-aided diagnosis (CAD) tools.
The diagnosis and treatment of Alzheimer’s disease (AD) using complementary multimodalities can improve the quality of life and mental state of patients.
In this study, we integrated a lightweight custom convolutional neural network
(CNN) model and nature-inspired optimization techniques to enhance the performance, robustness, and stability of progress detection in AD. A multi-modal
fusion database approach was implemented, including positron emission tomography (PET) and magnetic resonance imaging (MRI) datasets, to create a fused
database. We compared the performance of custom and pre-trained deep learning models with and without optimization and found that employing natureinspired algorithms like the particle swarm optimization algorithm (PSO) algorithm significantly improved system performance. The proposed methodology,
which includes a fused multimodality database and optimization strategy, improved performance metrics such as training, validation, test accuracy, precision, and recall. Furthermore, PSO was found to improve the performance of
pre-trained models by 3-5% and custom models by up to 22%. Combining different medical imaging modalities improved the overall model performance by
2-5%. In conclusion, a customized lightweight CNN model and nature-inspired
optimization techniques can significantly enhance progress detection, leading to
better biomedical research and patient care.
This research task develops a mobile healthcare analysis system (PHAS) which combines both easy ECG signal measurement and reliable analysis of heart rate variability for home care purpose. The PHAS is composed by a health care platform (HCP) and a data system analysis (DSA) module. The HCP consists of a self-developed two pole electrocardiography (ECG) measuring device and the DSA a data processing unit for detection and analysis of heart rate variability. For the DSA module, the adaptive R Peak detection algorithm is proposed to reliably detect the R peak of ECG for HRV analysis. A number of features are extracted from ECG signals. A data mining method is employed for HRV analysis to exploit the correlation between HRV and these features. Experiments are conducted by establishing a database of ECG signals measured from 29 subjects under rest and exercise condition. The results show the PHAS’s significant potential in mobile applications of personal daily health care.
ARTIFICIAL NEURAL NETWORKING.
FIRST STEP TO KNOWLEDGE IS TO KNOW THAT we are ignorant
Knowledge in medical field is characterized by uncertanity and vagueness
Historically as well as currently this fact remains a motivation for the development of medical decision support system are based on fuzzy logics
Greek philosopher visualized a basic model of brain function as early as 300 bc
Till date nervous system is not completely understood to human kind.
The document discusses machine learning methods for lung cancer detection using CT scans. It first provides background on lung cancer and the need for early detection. It then describes datasets used, including the LIDC-IDRI and Kaggle datasets containing labeled lung CT scans. Machine learning algorithms explored for segmentation include U-Net convolutional networks and for classification include logistic regression, naive Bayes, SVM, random forest and gradient boosting. Performance is evaluated based on sensitivity, specificity and other metrics. Overall results show machine learning methods achieving detection and classification performance comparable to radiologists while reducing false positives.
ijerst offers a fast publication schedule whilst maintaining rigorous peer review; the use of recommended electronic formats for article delivery expedites the process. Editorial Board or others working in the field as appropriate, to ensure they are likely to be of the level of interest and importance appropriate for the journal. International Journal of Engineering Research and Science & Technology (IJERST) is an international online journal in English published Quarterly. All submitted research articles are subjected to immediate rapid screening by the editors.
Chapter 5 applications of neural networksPunit Saini
Neural networks are being used experimentally in several medical applications, including modeling the cardiovascular system and diagnosing medical conditions. They can be used to detect diseases by learning from examples without needing a specific algorithm. Neural networks are also being explored for applications like implementing electronic noses for telemedicine. Researchers are working to build artificial brains more cheaply using field programmable gate arrays (FPGAs) on commercial boards, which could enable evolving millions of neural network modules at electronic speeds. Genetic algorithms are also being combined with neural networks to help optimize their structure and performance for tasks like object recognition.
Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...IRJET Journal
1) The document presents a technique for lung nodule feature extraction and classification using an Improved Neural Network Algorithm (INNA).
2) Texture features are extracted from CT lung images containing nodules using a Grey Level Co-occurrence Matrix based gradient approach.
3) The extracted features are used to classify lung nodules using INNA, which utilizes an enhanced backpropagation learning rule.
4) Simulation results show the proposed INNA technique achieves 98.99% accuracy in classifying cancer datasets, outperforming other techniques.
IRJET- A System to Detect Heart Failure using Deep Learning TechniquesIRJET Journal
This document proposes a system to detect heart failure using deep learning techniques. The system uses a boosted decision tree to initially detect the probability that a patient is prone to heart failure. If the probability is over 50%, the patient's ECG recordings are passed to a convolutional neural network (CNN) for more accurate detection of heart failure. The CNN is trained on a dataset of 60,000 ECG recordings. The system also aims to detect the subtype of heart failure using an SVM algorithm trained on data distinguishing systolic vs diastolic heart failure. The overall goal is to accurately detect heart failure at early stages to improve outcomes.
The MAGIC-5 lung CAD system uses three parallel approaches - region growing, virtual ants, and voxel-based neural networks - to detect and classify lung nodules in CT scans. It was validated on three databases and performs competitively with state-of-the-art systems. In the ANODE09 international competition, the three MAGIC-5 CADs achieved scores between 0.25-0.29 for detecting pulmonary nodules, outperforming other participating CAD systems. The MAGIC-5 project aims to support medical diagnosis and allow large-scale image analysis through developing models and algorithms for biomedical image analysis.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
More Related Content
Similar to Classification_of_heart_sounds_using_fra.pdf
Survey analysis for optimization algorithms applied to electroencephalogramIJECEIAES
This paper presents a survey for optimization approaches that analyze and classify electroencephalogram (EEG) signals. The automatic analysis of EEG presents a significant challenge due to the high-dimensional data volume. Optimization algorithms seek to achieve better accuracy by selecting practical features and reducing unwanted features. Forty-seven reputable research papers are provided in this work, emphasizing the developed and executed techniques divided into seven groups based on the applied optimization algorithm particle swarm optimization (PSO), ant colony optimization (ACO), artificial bee colony (ABC), grey wolf optimizer (GWO), Bat, Firefly, and other optimizer approaches). The main measures to analyze this paper are accuracy, precision, recall, and F1-score assessment. Several datasets have been utilized in the included papers like EEG Bonn University, CHB-MIT, electrocardiography (ECG) dataset, and other datasets. The results have proven that the PSO and GWO algorithms have achieved the highest accuracy rate of around 99% compared with other techniques.
WAVELET SCATTERING TRANSFORM FOR ECG CARDIOVASCULAR DISEASE CLASSIFICATIONgerogepatton
Classifying the ECG dataset is the main technique for diagnosing heart disease. However, the focus of this
field is increasingly on prediction, with a growing dependence on machine learning techniques. This study
aimed to enhance the accuracy of cardiovascular disease classification using data from the PhysioNet
database by employing machine learning (ML). The study proposed several multi-class classification
models that accurately identify patterns within three classes: heart failure rhythm (HFR), normal heart
rhythm (NHR), and arrhythmia (ARR). This was accomplished by utilizing a database containing 162 ECG
signals. The study employed a variety of techniques, including frequency-time domain analysis, spectral
features, and wavelet scattering, to extract features and capture unique characteristics from the ECG
dataset. The SVM model produced a training accuracy of 97.1% and a testing accuracy of 92%. This work
provides a reliable, effective, and human error-free diagnostic tool for identifying heart disease.
Furthermore, it could prove to be a valuable resource for future medical research projects aimed at
improving the diagnosis and treatment of cardiovascular diseases
WAVELET SCATTERING TRANSFORM FOR ECG CARDIOVASCULAR DISEASE CLASSIFICATIONijaia
Classifying the ECG dataset is the main technique for diagnosing heart disease. However, the focus of this
field is increasingly on prediction, with a growing dependence on machine learning techniques. This study
aimed to enhance the accuracy of cardiovascular disease classification using data from the PhysioNet
database by employing machine learning (ML). The study proposed several multi-class classification
models that accurately identify patterns within three classes: heart failure rhythm (HFR), normal heart
rhythm (NHR), and arrhythmia (ARR). This was accomplished by utilizing a database containing 162 ECG
signals. The study employed a variety of techniques, including frequency-time domain analysis, spectral
features, and wavelet scattering, to extract features and capture unique characteristics from the ECG
dataset. The SVM model produced a training accuracy of 97.1% and a testing accuracy of 92%. This work
provides a reliable, effective, and human error-free diagnostic tool for identifying heart disease.
Furthermore, it could prove to be a valuable resource for future medical research projects aimed at
improving the diagnosis and treatment of cardiovascular diseases.
WAVELET SCATTERING TRANSFORM FOR ECG CARDIOVASCULAR DISEASE CLASSIFICATIONgerogepatton
Classifying the ECG dataset is the main technique for diagnosing heart disease. However, the focus of this
field is increasingly on prediction, with a growing dependence on machine learning techniques. This study
aimed to enhance the accuracy of cardiovascular disease classification using data from the PhysioNet
database by employing machine learning (ML). The study proposed several multi-class classification
models that accurately identify patterns within three classes: heart failure rhythm (HFR), normal heart
rhythm (NHR), and arrhythmia (ARR). This was accomplished by utilizing a database containing 162 ECG
signals. The study employed a variety of techniques, including frequency-time domain analysis, spectral
features, and wavelet scattering, to extract features and capture unique characteristics from the ECG
dataset. The SVM model produced a training accuracy of 97.1% and a testing accuracy of 92%. This work
provides a reliable, effective, and human error-free diagnostic tool for identifying heart disease.
Furthermore, it could prove to be a valuable resource for future medical research projects aimed at
improving the diagnosis and treatment of cardiovascular diseases.
DETECTION OF CRACKLES AND WHEEZES IN LUNG SOUND USING TRANSFER LEARNING hiij
In recent years, deep learning models have improved how well various diseases, particularly respiratory
ailments, can be diagnosed. In order to assist in offering a diagnosis of respiratory pathologies in digitally
recorded respiratory sounds, this research will provide an evaluation of the effectiveness of several deep
learning models connected with the raw lung auscultation sounds in detecting respiratory pathologies. We
will also determine which deep learning model is most appropriate for this purpose. With the development
of computer -systems that can collect and analyze enormous volumes of data, the medical profession is
establishing several non-invasive tools. This work attempts to develop a non-invasive technique for
identifying respiratory sounds acquired by a stethoscope and voice recording software via machine
learning techniques. This study suggests a trained and proven CNN-based approach for categorizing
respiratory sounds. A visual representation of each audio sample is constructed, allowing resource
identification for classification using methods like those used to effectively describe visuals. We used a
technique called Mel Frequency Cepstral Coefficients (MFCCs). Here, features are retrieved and
categorized via VGG16 (transfer learning) and prediction is accomplished using 5-fold cross-validation.
Employing various data splitting techniques, Respiratory Sound Database obtained cutting-edge results,
including accuracy of 95%, precision of 88%, recall score of 86%, and F1 score of 81 %. We trained and
tested the model using a sound database made by the International Conference on Biomedical and Health
Informatics (ICBHI) in 2017 and annotated by experts with a classification of the lung sound.
DETECTION OF CRACKLES AND WHEEZES IN LUNG SOUND USING TRANSFER LEARNING hiij
This document presents research on detecting crackles and wheezes in lung sounds using transfer learning. The researchers used the VGG16 convolutional neural network to extract features from lung sound spectrograms preprocessed using Mel Frequency Cepstral Coefficients. They trained and tested the model on the ICBHI 2017 Respiratory Sound Database, achieving 95% accuracy, 88% precision, 86% recall, and 81% F1 score. The researchers compared their results to other CNN and RNN models to determine the optimal model for lung sound classification.
PREDICTIVE ANALYTICS IN HEALTHCARE SYSTEM USING DATA MINING TECHNIQUEScscpconf
The health sector has witnessed a great evolution following the development of new computer technologies, and that pushed this area to produce more medical data, which gave birth to multiple fields of research. Many efforts are done to cope with the explosion of medical data on one hand, and to obtain useful knowledge from it on the other hand. This prompted researchers to apply all the technical innovations like big data analytics, predictive analytics, machine learning and learning algorithms in order to extract useful knowledge and help in making decisions. With the promises of predictive analytics in big data, and the use of machine learning
algorithms, predicting future is no longer a difficult task, especially for medicine because predicting diseases and anticipating the cure became possible. In this paper we will present an overview on the evolution of big data in healthcare system, and we will apply a learning algorithm on a set of medical data. The objective is to predict chronic kidney diseases by using Decision Tree (C4.5) algorithm.
Predicting heart failure using a wrapper-based feature selectionnooriasukmaningtyas
In the current health system, it is very difficult for medical practitioners/ physicians to diagnose the effectiveness of heart contraction. In this research, we proposed a machine learning model to predict heart contraction using an artificial neural network (ANN). We also proposed a novel wrapper-based feature selection utilizing a grey wolf optimization (GWO) to reduce the number of required input attributes. In this work, we compared the results achieved using our method and several conventional machine learning algorithms approaches such as support vector machine, decision tree, Knearest neighbor, naïve bayes, random forest, and logistic regression. Computational results show not only that much fewer features are needed, but also higher prediction accuracy can be achieved around 87%. This work has the potential to be applicable to clinical practice and become a supporting tool for doctors/physicians.
This document presents a novel deep learning approach for single-lead electrocardiogram (ECG) classification. The approach uses Restricted Boltzmann Machines (RBM) and Deep Belief Networks (DBN) for ECG classification after detecting ventricular and supraventricular heartbeats from single-lead ECG signals. Experimental results on the MIT-BIH database show the approach achieves high average recognition accuracies of 93.63% for ventricular ectopic beats and 95.57% for supraventricular ectopic beats at a low sampling rate of 114 Hz, outperforming traditional methods.
Attention gated encoder-decoder for ultrasonic signal denoisingIAESIJAI
Ultrasound imaging is one of the most widely used non-destructive testing methods. The transducer emits pulses that travel through the imaged samples and are reflected by echo-forming impedance. The resulting ultrasonic signals usually contain noise. Most of the traditional noise reduction algorithms require high skills and prior knowledge of noise distribution, which has a crucial impact on their performances. As a result, these methods generally yield a loss of information, significantly influencing the final data and deeply limiting both sensitivity and resolution of imaging devices in medical and industrial applications. In the present study, a denoising method based on an attention-gated convolutional autoencoder is proposed to fill this gap. To evaluate its performance, the suggested protocol is compared to widely used methods such as butterworth filtering (BF), discrete wavelet transforms (DWT), principal component analysis (PCA), and convolutional autoencoder (CAE) methods. Results proved that better denoising can be achieved especially when the original signal-to-noise ratio (SNR) is very low and the sound waves’ traces are distorted by noise. Moreover, the initial SNR was improved by up to 30 dB and the resulting Pearson correlation coefficient was maintained over 99% even for ultrasonic signals with poor initial SNR.
A novel framework for efficient identification of brain cancer region from br...IJECEIAES
Diagnosis of brain cancer using existing imaging techniques, e.g., Magnetic Resonance Imaging (MRI) is shrouded with various degrees of challenges. At present, there are very few significant research models focusing on introducing some novel and unique solutions towards such problems of detection. Moreover, existing techniques are found to have lesser accuracy as compared to other detection schemes. Therefore, the proposed paper presents a framework that introduces a series of simple and computationally cost-effective techniques that have assisted in leveraging the accuracy level to a very higher degree. The proposed framework takes the input image and subjects it to non-conventional segmentation mechanism followed by optimizing the performance using directed acyclic graph, Bayesian Network, and neural network. The study outcome of the proposed system shows the significantly higher degree of accuracy in detection performance as compared to frequently existing approaches.
The biomedical profession has gained importance due to the rapid and accurate diagnosis of clinical patients using computer-aided diagnosis (CAD) tools.
The diagnosis and treatment of Alzheimer’s disease (AD) using complementary multimodalities can improve the quality of life and mental state of patients.
In this study, we integrated a lightweight custom convolutional neural network
(CNN) model and nature-inspired optimization techniques to enhance the performance, robustness, and stability of progress detection in AD. A multi-modal
fusion database approach was implemented, including positron emission tomography (PET) and magnetic resonance imaging (MRI) datasets, to create a fused
database. We compared the performance of custom and pre-trained deep learning models with and without optimization and found that employing natureinspired algorithms like the particle swarm optimization algorithm (PSO) algorithm significantly improved system performance. The proposed methodology,
which includes a fused multimodality database and optimization strategy, improved performance metrics such as training, validation, test accuracy, precision, and recall. Furthermore, PSO was found to improve the performance of
pre-trained models by 3-5% and custom models by up to 22%. Combining different medical imaging modalities improved the overall model performance by
2-5%. In conclusion, a customized lightweight CNN model and nature-inspired
optimization techniques can significantly enhance progress detection, leading to
better biomedical research and patient care.
This research task develops a mobile healthcare analysis system (PHAS) which combines both easy ECG signal measurement and reliable analysis of heart rate variability for home care purpose. The PHAS is composed by a health care platform (HCP) and a data system analysis (DSA) module. The HCP consists of a self-developed two pole electrocardiography (ECG) measuring device and the DSA a data processing unit for detection and analysis of heart rate variability. For the DSA module, the adaptive R Peak detection algorithm is proposed to reliably detect the R peak of ECG for HRV analysis. A number of features are extracted from ECG signals. A data mining method is employed for HRV analysis to exploit the correlation between HRV and these features. Experiments are conducted by establishing a database of ECG signals measured from 29 subjects under rest and exercise condition. The results show the PHAS’s significant potential in mobile applications of personal daily health care.
ARTIFICIAL NEURAL NETWORKING.
FIRST STEP TO KNOWLEDGE IS TO KNOW THAT we are ignorant
Knowledge in medical field is characterized by uncertanity and vagueness
Historically as well as currently this fact remains a motivation for the development of medical decision support system are based on fuzzy logics
Greek philosopher visualized a basic model of brain function as early as 300 bc
Till date nervous system is not completely understood to human kind.
The document discusses machine learning methods for lung cancer detection using CT scans. It first provides background on lung cancer and the need for early detection. It then describes datasets used, including the LIDC-IDRI and Kaggle datasets containing labeled lung CT scans. Machine learning algorithms explored for segmentation include U-Net convolutional networks and for classification include logistic regression, naive Bayes, SVM, random forest and gradient boosting. Performance is evaluated based on sensitivity, specificity and other metrics. Overall results show machine learning methods achieving detection and classification performance comparable to radiologists while reducing false positives.
ijerst offers a fast publication schedule whilst maintaining rigorous peer review; the use of recommended electronic formats for article delivery expedites the process. Editorial Board or others working in the field as appropriate, to ensure they are likely to be of the level of interest and importance appropriate for the journal. International Journal of Engineering Research and Science & Technology (IJERST) is an international online journal in English published Quarterly. All submitted research articles are subjected to immediate rapid screening by the editors.
Chapter 5 applications of neural networksPunit Saini
Neural networks are being used experimentally in several medical applications, including modeling the cardiovascular system and diagnosing medical conditions. They can be used to detect diseases by learning from examples without needing a specific algorithm. Neural networks are also being explored for applications like implementing electronic noses for telemedicine. Researchers are working to build artificial brains more cheaply using field programmable gate arrays (FPGAs) on commercial boards, which could enable evolving millions of neural network modules at electronic speeds. Genetic algorithms are also being combined with neural networks to help optimize their structure and performance for tasks like object recognition.
Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...IRJET Journal
1) The document presents a technique for lung nodule feature extraction and classification using an Improved Neural Network Algorithm (INNA).
2) Texture features are extracted from CT lung images containing nodules using a Grey Level Co-occurrence Matrix based gradient approach.
3) The extracted features are used to classify lung nodules using INNA, which utilizes an enhanced backpropagation learning rule.
4) Simulation results show the proposed INNA technique achieves 98.99% accuracy in classifying cancer datasets, outperforming other techniques.
IRJET- A System to Detect Heart Failure using Deep Learning TechniquesIRJET Journal
This document proposes a system to detect heart failure using deep learning techniques. The system uses a boosted decision tree to initially detect the probability that a patient is prone to heart failure. If the probability is over 50%, the patient's ECG recordings are passed to a convolutional neural network (CNN) for more accurate detection of heart failure. The CNN is trained on a dataset of 60,000 ECG recordings. The system also aims to detect the subtype of heart failure using an SVM algorithm trained on data distinguishing systolic vs diastolic heart failure. The overall goal is to accurately detect heart failure at early stages to improve outcomes.
The MAGIC-5 lung CAD system uses three parallel approaches - region growing, virtual ants, and voxel-based neural networks - to detect and classify lung nodules in CT scans. It was validated on three databases and performs competitively with state-of-the-art systems. In the ANODE09 international competition, the three MAGIC-5 CADs achieved scores between 0.25-0.29 for detecting pulmonary nodules, outperforming other participating CAD systems. The MAGIC-5 project aims to support medical diagnosis and allow large-scale image analysis through developing models and algorithms for biomedical image analysis.
Similar to Classification_of_heart_sounds_using_fra.pdf (20)
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
Using recycled concrete aggregates (RCA) for pavements is crucial to achieving sustainability. Implementing RCA for new pavement can minimize carbon footprint, conserve natural resources, reduce harmful emissions, and lower life cycle costs. Compared to natural aggregate (NA), RCA pavement has fewer comprehensive studies and sustainability assessments.
Redefining brain tumor segmentation: a cutting-edge convolutional neural netw...IJECEIAES
Medical image analysis has witnessed significant advancements with deep learning techniques. In the domain of brain tumor segmentation, the ability to
precisely delineate tumor boundaries from magnetic resonance imaging (MRI)
scans holds profound implications for diagnosis. This study presents an ensemble convolutional neural network (CNN) with transfer learning, integrating
the state-of-the-art Deeplabv3+ architecture with the ResNet18 backbone. The
model is rigorously trained and evaluated, exhibiting remarkable performance
metrics, including an impressive global accuracy of 99.286%, a high-class accuracy of 82.191%, a mean intersection over union (IoU) of 79.900%, a weighted
IoU of 98.620%, and a Boundary F1 (BF) score of 83.303%. Notably, a detailed comparative analysis with existing methods showcases the superiority of
our proposed model. These findings underscore the model’s competence in precise brain tumor localization, underscoring its potential to revolutionize medical
image analysis and enhance healthcare outcomes. This research paves the way
for future exploration and optimization of advanced CNN models in medical
imaging, emphasizing addressing false positives and resource efficiency.
International Conference on NLP, Artificial Intelligence, Machine Learning an...gerogepatton
International Conference on NLP, Artificial Intelligence, Machine Learning and Applications (NLAIM 2024) offers a premier global platform for exchanging insights and findings in the theory, methodology, and applications of NLP, Artificial Intelligence, Machine Learning, and their applications. The conference seeks substantial contributions across all key domains of NLP, Artificial Intelligence, Machine Learning, and their practical applications, aiming to foster both theoretical advancements and real-world implementations. With a focus on facilitating collaboration between researchers and practitioners from academia and industry, the conference serves as a nexus for sharing the latest developments in the field.
Batteries -Introduction – Types of Batteries – discharging and charging of battery - characteristics of battery –battery rating- various tests on battery- – Primary battery: silver button cell- Secondary battery :Ni-Cd battery-modern battery: lithium ion battery-maintenance of batteries-choices of batteries for electric vehicle applications.
Fuel Cells: Introduction- importance and classification of fuel cells - description, principle, components, applications of fuel cells: H2-O2 fuel cell, alkaline fuel cell, molten carbonate fuel cell and direct methanol fuel cells.
Comparative analysis between traditional aquaponics and reconstructed aquapon...bijceesjournal
The aquaponic system of planting is a method that does not require soil usage. It is a method that only needs water, fish, lava rocks (a substitute for soil), and plants. Aquaponic systems are sustainable and environmentally friendly. Its use not only helps to plant in small spaces but also helps reduce artificial chemical use and minimizes excess water use, as aquaponics consumes 90% less water than soil-based gardening. The study applied a descriptive and experimental design to assess and compare conventional and reconstructed aquaponic methods for reproducing tomatoes. The researchers created an observation checklist to determine the significant factors of the study. The study aims to determine the significant difference between traditional aquaponics and reconstructed aquaponics systems propagating tomatoes in terms of height, weight, girth, and number of fruits. The reconstructed aquaponics system’s higher growth yield results in a much more nourished crop than the traditional aquaponics system. It is superior in its number of fruits, height, weight, and girth measurement. Moreover, the reconstructed aquaponics system is proven to eliminate all the hindrances present in the traditional aquaponics system, which are overcrowding of fish, algae growth, pest problems, contaminated water, and dead fish.
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesChristina Lin
Traditionally, dealing with real-time data pipelines has involved significant overhead, even for straightforward tasks like data transformation or masking. However, in this talk, we’ll venture into the dynamic realm of WebAssembly (WASM) and discover how it can revolutionize the creation of stateless streaming pipelines within a Kafka (Redpanda) broker. These pipelines are adept at managing low-latency, high-data-volume scenarios.
2. 2 Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788
abnormal heart sound recordings. The PhysioNet/Computing in
Cardiology Challenge 2016 presented a dataset to develop, test
and compare several algorithms with conventional architectures
and deep learning to classify heart sounds. Many research groups
contributed new methods in response to this challenge [5–7].
Gokhale [8] introduced an algorithm that utilizes Hilbert enve-
lope and wavelet features with boosted trees ensemble classifier.
Goda et al. [9] used time-frequency domain features and support
vector machine (SVM) classifier. An algorithm was proposed by
Grzegorczyk et al. [10] to classify heart sound recording based
on deep neural networks. Conventional neural network and auto-
encoder deep neural network were used with forty-eight time
and frequency domain features. Tschannen et al. [11] introduced
another technique where deep features (generated by a wavelet-
based convolutional neural network) and time-frequency domain
features were both used with L2-SVM classifier. Homsi et al. [12]
presented an approach that use time, frequency, wavelet and sta-
tistical domain features with a nested set ensemble classifier that
included random forest, LogitBoost and cost-sensitive classifiers.
Langley et al. [13] utilized wavelet entropy to classify unsegmented
and short duration PCG signals where a wavelet entropy threshold
was determined from the training set then PCG signals with entropy
below the threshold were classified as abnormal. In Singh-Miller
et al. [14], spectral features with discriminative model based on ran-
dom forest regressor were applied for classification of heart sound
recordings. Vernekar et al. [15] demonstrated Markov features
with a weighted ensemble classifier that include four AdaBoost
ensemble classifiers and four artificial neural networks (ANN) to
classify PCG signals. Plesinger et al. [16] implemented a fuzzy logic
like approach with logical rules and probability assessment based
on histograms to classify heart sounds. Nabhan and Warriek [17]
attempted to improve the work in [12] by detection the outlier
signal and separated it from standard range signal by using an
interquartile range threshold. Abdollahpur et al. [18] proposed an
algorithm that assessed the signal quality of the segmented car-
diac cycle then a total of ninety features including time domain,
time-frequency, perceptual and Mel-frequency cepstral coefficient
(MFCC) were extracted from the correctly segmented cycles only.
The classification was performed using three feed-forward neural
networks followed by a voting system. Langley and Murray [19]
tried to improve the work in [13] where short and unsegmented
heart sounds recordings were classified using feature threshold
based classifier. Spectral amplitude and wavelet entropy features
were calculated using FFT and wavelet analysis. A decision tree was
then used to combine the spectral amplitude and wavelet entropy.
Maknickas and Maknickas [20] applied CNN to classify heart sound
records with Mel-frequency spectral coefficients (MFSC), difference
and second-order difference of the MFSC calculated and fed to CNN
as three dimensions for each frame. Whitaker et al. [21] proposed
combining sparse coding and time domain features for heart sounds
classification. Six SVM classifiers used. Abduh et al. [39] applied
fractional Fourier transform based Mel-frequency spectral coeffi-
cients (FrFT-MFSC) and stacked auto-encoder deep neural network
to classify heart sound records. In spite of the excellent performance
of the new method, there is always the concern of overfitting prob-
lem in such methods. Deep learning methods require large amounts
of training data that are difficult to meet by biomedical data sets,
and the current problem is no exception. Therefore, the estimation
of the many parameters associated with the deep learning neu-
ral network architecture using the much fewer training samples
available from our data set raises a major concern that the system
may not generalize well. Consequently, studies in this field have
attempted to propose augmentation or simulation methods to gen-
erate more data samples from the limited available ones. This raises
concerns about how realistic such synthetic data samples are and
whether this approach should be acceptable for medical diagnostic
systems. A number of regulatory organizations as AAMI (Associa-
tion for the Advancement of Medical Instrumentation), Medicines &
Healthcare products Regulatory Agency (MHRA), and British Stan-
dards Institute (BSI) held a joint workshop in 2019 to discuss
governance and regulation of emerging artificial intelligence and
machine learning technologies in healthcare and provided their
conclusions in an important position paper [40]. Among their rec-
ommendation was a particularly relevant one about the critical
importance of quantity and quality of data input to intelligent
systems. Even though such recommendations are not yet incorpo-
rated into active regulations, it seems that systems based on such
approaches using synthetic data will be much harder to be cleared
by regulatory agencies given the burden of proof of their validity.
Therefore, it is desirable to develop a system that is based on tradi-
tional classification with much fewer parameters to estimate and
would allow the limited available data set to be sufficient to train,
validate and test the system in a robust manner.
In this work, we develop a computer-aided auscultation sys-
tem based on the fractional Fourier transform based Mel-frequency
spectral coefficients features set developed by our group [39] com-
bined with traditional classifiers. The new system is described
in terms of its preprocessing, cardiac cycle segmentation, fea-
ture extraction, feature reduction and classification stages. The
proposed system with several variants of its implementation are
verified on the dataset of the PhysioNet/Computing in Cardiol-
ogy Challenge 2016 and compared to previous work on the same
dataset using the evaluation metric of this challenge.
2. Methodology
The components of the proposed computer-aided auscultation
system are shown in Fig. 1 and are detailed as follows.
2.1. Preprocessing
Ambient sounds, lung sound, internal body noise, cough and
stethoscope movement are the main interferences in heart sounds
recording and analysis. Therefore, effective filtration is of critical
important to enhance the heart sounds signal by reducing the influ-
ence of background noise and removing spike noise. In this work,
a two-stage preprocessing system is utilized. In the first stage, a
3rd-order Butterworth band-pass filter with corner frequencies of
15 and 800 Hz is used to select the useful bandwidth of the heart
sounds. In the second stage, spectral subtraction denoising method
was applied [22]. This method was reported to be highly effective
for background noise reduction in such challenging applications as
EEG signal denoising for brain-computer interface [23]. The advan-
tage of this technique is its adaptive noise estimation. The noise
power is estimated from the frequencies outside of the range of
frequencies of heart sounds then a weighted version of it is subse-
quently subtracted from the power spectrum of raw heart sounds
and used to reconstruct the denoised signal [22]. In this work,
spectral subtraction filtration was used with a weighting factor of
0.5.
2.2. Cardiac cycle segmentation
In this stage, each PCG signal is split into cardiac cycles. It is
essential for the recognition of the systolic or diastolic states, per-
mitting succeeding categorizing of abnormal states in this areas.
Many algorithms were implemented. Some was carried out by
the usage of a reference signal such as the ECG, the segmenta-
tion algorithms require to record the ECG in parallel. It will assist
to recognize heart sounds. Other algorithms do not use ECG as a
reference. In this work Springer’s improved version of Schmidt’s
segmentation algorithm [24] utilized to split PCG signal into cardiac
3. Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788 3
Fig. 1. Block diagram of the proposed approach for classification of heart sounds with an illustration of the signals obtained at each step (PCG: Phonocardiogram, FrFT:
Fractional Fourier transform. MFSC: Mel-frequency spectral coefficients).
cycles. Then, each complete cardiac cycle is used for processing. This
algorithm does not require ECG synchronization and uses a logis-
tic regression hidden semi-Markov model (HSMM) to estimate the
most likely sequence of states by incorporating information about
expected heart sound state durations. To overcome the problem of
variable time length of cardiac cycles (and hence size of their digital
signals) in subsequent processing steps, the size of all signals was
set to be the longest cardiac cycle found across all PCG recordings
(here, it was around 2 s). For cardiac cycles with shorter length, they
were zero-padded to that length. This ensured uniform frequency
resolution for all signals.
2.3. Feature extraction
In this stage, the segmented cardiac cycles are processed to
extract a set of features that best represents their salient char-
acteristics for optimal classification accuracy. Generally speaking,
since phonocardiogram signal is essentially similar to speech sig-
nal, popular feature representations currently used for speech
signals such as spectral and cepstral features can be potentially
useful in this application as well [25]. Among the most success-
ful of these are the Mel-frequency cepstral coefficients (MFCC)
transform the original PCG signal into a time-frequency rep-
resentation of the distribution of signal energy [4]. The MFCC
features correspond to the cepstrum of the log filter-bank energies.
The Mel-frequency scale is approximately linear for frequencies
below 1 kHz and logarithmic for frequencies above 1 kHz [27].
This is motivated by the fact that the human auditory system
becomes less frequency-selective as frequency increases above
1 kHz.
In this work, we use the fractional Fourier transform based
Mel-frequency spectral coefficients features (FrFT-MFSC) devel-
oped by our group [39]. These features are based on a modified
version of MFCC whereby Fractional Fourier transform is used
instead of discrete Fourier or the discrete cosine transforms to
allow for a flexible time-frequency expansion. The log-energy
was computed directly from the Mel-frequency spectral coeffi-
cients.
The Fractional Fourier Transform (FrFT) is the generalization
form of Fourier transform and indicates a rotation of a signal in
time-frequency plane. While the discrete Fourier or cosine trans-
forms can obtain the frequency components of a signal, they are
essentially decomposing the signal in terms of harmonic com-
ponents that are not localized in time. Therefore, they cannot
describe local variations in the signal as compared to a time-
frequency analysis method. The fractional Fourier transform offers
a generalization of the Fourier transform with the Fourier spec-
trum and the time domain signal are both special cases of this
transform. Hence, it allows a more flexible time-frequency rep-
resentation than other methods such as the spectrogram, Wigner
distribution or ambiguity function in that it can transform sig-
nals to any intermediate domain between time and frequency.
Hence, FrFT is suitable for non-stationary signal processing and
has wide applications in basic signal analysis and speech recog-
nition.
4. 4 Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788
The continuous form of FrFT with ath-order of a signal s(t) can be
defined within 0 ≤ |a| ≤ 2 through the linear operator as [28–31],
Fa
s
(wa) =
∞
−∞
Ka (wa, t) s (t) dt. (1)
Here, the kernel Ka (wa, t) is given by:
Ka (wa, t) =
⎧
⎨
⎩
k∅ exp(j
w2
a cot (∅) − 2watcsc (∅) + t2
cot(∅))
, if a /
= 0, ± 2,
ı (wa − t) , if a = 0,
ı (wa + t) , if a = ±2
(2)
where, ∅ = a
2
and k∅ =
exp(−j(
sgn(∅)
4
− ∅
2
)
√
|sin(∅)|
=
1 − j cot(∅) , and wa
means the variables in ath-order fractional Fourier transform. Sim-
ilar to the discrete Fourier transform (DFT), the discrete fractional
Fourier transform (DFrFT) matrix Fa
(m, n) is obtained as the dis-
crete version of Eq. (2) as,
Fa
(m, n) =
N−1
k=0
uk (m) exp(
−jka
2
) uk (n) . (3)
Here, u is discrete Hermite-Gaussian function and a is the fractional
order. The discrete fractional Fourier transform of a signal is just the
matrix vector multiplication of this transform matrix in Eq. (3) with
the signal vector.
Fig. 2 shows the steps of FrFT-MFSC extraction, which are
derived as follows:
1 The signal is pre-emphasized by a filter H(z)= (1 – 0.9 z−1) that
boosts the higher frequencies to balance the spectrum of sounds
steep roll-off in the high frequency region [38]. Then the signal
is framed and windowed into the selected frames length.
2 The fractional Fourier transform of the windowed signal is calcu-
lated.
3 The powers of the obtained spectrum are mapped into the Mel-
scale using triangular overlapping windows.
4 The logs of the powers are calculated at each Mel-frequency to
obtain the FrFT-MFSC values.
We perform normalization to make all features in the range
of [0,1] [26]. The Mel-scale is defined as follows where f is the
frequency in Hz:
Mel ( f ) = 2595 log10 1 +
f
700
. (4)
For FrFT-MFSC feature extraction, the processing parameters
were as follows: four fractional orders (a = 0.90,0.95,1.0,1.1). The
frame length parameter was chosen to cover the duration of each
of the fundamental heart sounds S1 and S2 (nominal values of
S1 = 122 ms and S2 = 114 ms). So, the chosen frame length is taken as
125 ms. The frame shift was taken close to 50 % of the frame length
as is commonly used in the literature and chosen to be 50 ms. Fig. 3
illustrates the frames for a single segmented cardiac cycle after
zero-padding to 2000 ms. The number of bands was taken as 20
based on experimentation. For each fractional order, a 20 (bands)
× 38 (frames) spectral coefficients matrix was computed. That is,
for each cardiac cycle, a total of 3040 features were obtained.
2.4. Feature dimensionality reduction
Principal component analysis (PCA) is used to convert possi-
bly correlated features into linearly uncorrelated ones with an
orthogonal transformation. It also identifies components that do
not contribute much to the representation (or variance or the data)
and hence can be removed. In this work, the percentage of variance
accounted for in PCA was set as 95 %.
2.5. Classification
Many factors affect the performance of traditional classifiers and
therefore should be considered during the mathematical model
building and as well as in the generalization for real data anal-
ysis. The most important factors include the distribution of data
in the feature space, balance of class data samples, heterogeneity,
diversity of training data and presence of noise. Furthermore, each
classifier has relative advantages and disadvantage compared to
others. For example, k-nearest neighbor (KNN) is a memory-based
learning technique [32] and therefore training and testing data
should both be available all the time. Also, decision tree classifiers
are preferred for noisy datasets [33]. On the other hand, boosted
ensemble classifier is reported to perform best for imbalanced data
[34].
In this work, we check the performance of different tradi-
tional classifiers on a set of features derived from dataset by
using cross-validation method. We carry out training to search for
the best classification model, including support vector machines
(using linear, quadratic, cubic, and Gaussian kernels), k-nearest
neighbor (using linear, cosine, cubic, and weighted distance met-
rics), and ensemble classification (bagged Trees, subspace KNN and
RUSBoosted tree) [35,36]. Table 1 summarizes the classification
parameters considered in this study.
Cross-validation and local holdout methods were used to train
and test the different classifiers. In the cross-validation experi-
ments, the feature vector data were divided into 5 folds. Each
fold was held out in turn for testing and the model was trained
for that fold using all data outside the fold. Then, each model
performance was assessed using the data inside the fold, and
the overall results are computed as the average over all folds.
In the local holdout method experiments, 80 % of feature vec-
tor data was used as training set and the remaining 20 % was
used for testing and both were chosen randomly. This 5-fold
cross-validation and 80-20 % holdout experiments were per-
formed 5 times by randomly selecting/dividing the given data
in order to estimate the mean performance variability of the
classifier.
2.6. Database description
In any pattern recognition problem, database selection plays
important role in model building and generalization. Database
should be large, diverse and understandable. So analysis and com-
parison of the different implemented algorithms will be easy.
The performance of the proposed system was verified using the
database of the PhysioNet/Computing in Cardiology Challenge 2016
[7]. The used dataset includes 3153 recordings. The sure labeled
data of this dataset includes 2868 recordings collected from six
databases. Normal patient records are 2249 whereas abnormal
patient records are 619, lasting from 5 s to just over 120 s. All
recordings have been resampled to 2000 Hz and have been pro-
vided as “.wav” format. They were recorded in different real-world
clinical and nonclinical environments and include recordings of
varying amounts of noise. The data were recorded from both nor-
mal subjects and pathological patients, and from both children and
adults. The same patient could have between 1 and 6 recordings
in the database. The data were recorded from different locations
on the body (including aortic area, pulmonic area, tricuspid area
and mitral area, among others). The data are clearly imbalanced
since the number of normal recordings are much larger than that
of abnormal recordings. The dataset was divided randomly into
two independent sets with 80 % for training and 20 % for test-
ing.
It should be noted that this dataset incorporates realistic chal-
lenges of noisy, imbalanced data in addition to different data
5. Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788 5
Fig. 2. Feature extraction using fractional Fourier transform Mel-frequency spectral coefficients with an illustration of the processed data at each step (PCG: Phonocardiogram,
FrFT: Fractional Fourier transform. MFSC: Mel-frequency spectral coefficients).
collection environments (that is, collection was done by different
research groups). Heterogeneity in the collection of the recordings
introduces differences with potential to make classifier training
more difficult. In particular, a classifier trained on one population
might perform poorly when applied to another.
2.7. Performance evaluation
To evaluate the performance of classification process, the con-
fusion matrix is computed with the abnormal cases as the positive
6. 6 Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788
Fig. 3. The frames used for a single segmented cardiac cycle after zero-padding.
Table 1
Parameters of the used classifiers.
Classifier Parameters
SVM
Linear Linear kernel, Sequential Minimal Optimization solver
Quadratic Quadratic kernel, polynomial, 2nd order, Sequential Minimal Optimization solver
Cubic Cubic kernel, polynomial, 3rd order, Sequential Minimal Optimization solver
Gaussian Gaussian kernel, polynomial, Sequential Minimal Optimization solver
Ensemble Classifier
Bagged Trees Combine Weights: Weighted Average, training cycles = 200
Subspace KNN Combine Weights: Weighted Average, training cycles = 200
RUSBoosted Tree Combine Weights: Weighted Average, training cycles = 200
KNN
Linear Euclidean distance, Distance weight = Equal, k = 10
Cosine Cosine distance, k = 10
Cubic Cubic distance, k = 10
Weighted Distance weight, k = 10
class and used to calculate the values for sensitivity, specificity and
accuracy as,
Sensitivity (Se) =
TP
TP + FN
, (5)
Specificity (Sp) =
TN
TN + FP
, (6)
Accuracy =
TP + TN
TP + TN + FP + FN
, (7)
where TP, TN, FP, and FN are the confusion matrix entries repre-
senting true positive, true negative, false positive and false negative
cases, respectively.
For imbalanced data, the error rate and the accuracy are not
appropriate to measure the classification performance because
they do not consider misclassification costs. As a result, they tend
to be strongly biased to favor the majority class and are sensitive
to class skews [34,39]. Therefore, an alternative evaluation score
based on the average between sensitivity and specificity was cho-
sen as the official evaluation metric for the PhysioNet/Computing
in Cardiology Challenge 2016 and defined as,
Score =
Se + Sp
2
. (8)
7. Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788 7
Table 2
FrFT- MFSC features classification results using 5-fold cross-validation train-test method showing the mean and standard deviation (SD) of 5 experiments.
Classifier Sensitivity Specificity Accuracy Score
SVM
Linear
Mean 0.6352 0.9429 0.8815 0.7890
SD 0.0009 0.0002 0.0002 0.0004
Quadratic
Mean 0.8284 0.9568 0.9312 0.8926
SD 0.0074 0.0014 0.0007 0.0032
Cubic
Mean 0.8657 0.9634 0.9439 0.9145
SD 0.0051 0.0012 0.0014 0.0027
Gaussian
Mean 0.7989 0.9569 0.9254 0.8779
SD 0.0110 0.0024 0.0019 0.0049
Ensemble
Bagged Trees
Mean 0.7387 0.9718 0.9254 0.8553
SD 0.0115 0.0015 0.0028 0.0059
Subspace KNN
Mean 0.8365 0.9703 0.9437 0.9034
SD 0.0094 0.0019 0.0022 0.0047
RUSBoosted Tree
Mean 0.7845 0.8502 0.8371 0.8173
SD 0.0131 0.0049 0.0048 0.0070
KNN
Linear
Mean 0.8277 0.9656 0.9381 0.8967
SD 0.0088 0.0011 0.0020 0.0045
Cosine
Mean 0.8012 0.9718 0.9378 0.8865
SD 0.0277 0.0084 0.0015 0.0097
Cubic
Mean 0.7305 0.9728 0.9245 0.8517
SD 0.0416 0.0064 0.0034 0.0177
Weighted
Mean 0.7901 0.9712 0.9351 0.8807
SD 0.0082 0.0010 0.0013 0.0038
Table 3
FrFT- MFSC features classification results using 80 % as training and 20 % as testing showing the mean and standard deviation (SD) of 5 experiments.
Classifier Sensitivity Specificity Accuracy Score
SVM
Linear
Mean 0.6292 0.9425 0.8807 0.7861
SD 0.0006 0.0014 0.0002 0.0003
Quadratic
Mean 0.8208 0.9305 0.9082 0.8757
SD 0.0024 0.0605 0.04802 0.0298
Cubic
Mean 0.8735 0.9666 0.9469 0.9200
SD 0.0018 0.0035 0.0006 0.0013
Gaussian
Mean 0.7879 0.9583 0.9239 0.8731
SD 0.0020 0.0011 0.0007 0.0012
Ensemble
Bagged Trees
Mean 0.7372 0.9709 0.9245 0.8540
SD 0.0033 0.0006 0.0008 0.0017
Subspace KNN
Mean 0.8453 0.9698 0.9447 0.9075
SD 0.0054 0.0008 0.0004 0.0024
RUSBoosted Tree
Mean 0.7907 0.8495 0.8386 0.8201
SD 0.0149 0.0097 0.0055 0.0043
KNN
Linear
Mean 0.8286 0.9659 0.9377 0.8972
SD 0.0013 0.0023 0.0002 0.0016
Cosine
Mean 0.7845 0.9770 0.9382 0.8808
SD 0.0036 0.0017 0.0003 0.0013
Cubic
Mean 0.7173 0.9775 0.9252 0.8474
SD 0.0074 0.0015 0.0020 0.0042
Weighted
Mean 0.7887 0.9716 0.9356 0.8802
SD 0.0012 0.0009 0.0003 0.0005
The results in this study present the values of all the above
measures.
3. Results and discussion
3.1. Experimental verification
The performance of the proposed system was verified using the
dataset of the PhysioNet/Computing in Cardiology Challenge 2016
described above [7]. Each of the PCG records was preprocessed
using the Butterworth band-pass filter of order 3 with corner fre-
quencies 15 and 800 Hz then enhanced further using the spectral
subtraction denoising with weight 0.5. Each record was segmented
into cardiac cycles resulting in 79,492 cardiac cycles. The longest
cardiac cycle found across all PCG recordings has a length of around
2 s. So, if a cardiac cycle had a length less than 2 s, the time series
was zero-padded to that length.
For FrFT-MFSC feature extraction, a total of 3040 features were
computed for each cardiac cycle. To improve the classification pro-
cess, all features vector values were normalized to interval [0,1].
Then, the feature reduction process was performed using PCA such
that the percentage of variance represented by the results is 95 %.
This process reduced the dimensionality of the feature space down
to 40 that are computed as weighted linear combination from all
3040 features.
Performance comparison was done for several traditional clas-
sifiers that included support vector machine (SVM), k-nearest
neighbor (KNN), bagged Trees ensemble classifier, subspace KNN
ensemble classifier, and RUSBoosted tree ensemble classifier with
parameters listed in Table 1.
3.2. Results
Table 2 summarizes the results of classification using 5-folds
cross validation train-test method. The cross-validation train-test
method appears to reduce the effect of over-fitting and helps
in model evaluation more effectively when using imbalanced
data. It should be noted that most classifiers in our experi-
8. 8 Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788
Table 4
Comparison of the performance metrics of different variants of the proposed
method.
Classifier Sensitivity Specificity Score
SVM - Cubic Kernel 0.8735 0.9666 0.9200
Subspace KNN Ensemble classifier 0.8453 0.9698 0.9075
KNN – Linear 0.8286 0.9659 0.8972
SVM -Quadratic Kernel 0.8284 0.9568 0.8926
KNN – Cosine 0.8012 0.9718 0.8865
KNN – Weighted 0.7901 0.9712 0.8807
SVM – Gaussian Kernel 0.7989 0.9569 0.8779
Ensemble classifier Bagged Trees 0.7387 0.9718 0.8553
KNN – Cubic Kernel 0.7305 0.9728 0.8517
Ensemble Classifier RUSBoosted Tree 0.7907 0.8495 0.8201
SVM - Linear Kernel 0.6352 0.9429 0.7890
ments achieved relatively balanced values for sensitivity and
specificity. The SVM classifier with cubic kernel achieved the
highest score value of 0.9145 with sensitivity and specificity of
0.8657 and 0.9634 respectively. The SVM classifier with cubic
kernel achieved the highest sensitivity of 0.8657. On the other
hand, SVM classifier with linear kernel achieved the lowest sen-
sitivity of 0.6352. The KNN classifier with cubic distance metric
classifier with achieved the highest specificity of 0.9728 while
RUSBoosted tree ensemble classifier the lowest specificity of
0.8502.
Table 3 summarizes the results of classification using local hold-
out train-test method. Most classifiers here also achieved relatively
balanced values for sensitivity and specificity. Also, the SVM clas-
sifier with cubic kernel achieved the highest score value of 0.9200
with sensitivity and specificity of 0.8735 and 0.9666 respectively.
The SVM classifier with cubic kernel achieved the highest sensitiv-
ity of 0.8735 while the SVM classifier with linear kernel achieved
the lowest sensitivity 0.6292. On the other hand, the KNN classi-
fier with cubic distance metric achieved the highest specificity of
0.9775, while RUSBoosted tree ensemble classifier produced the
lowest specificity of 0.8495.
The results of several variants of the proposed work are pre-
sented in Table 4. Also, the results reported in the previous work in
the literature are presented in Table 5.
3.3. Discussion
The SVM classifier with cubic kernel achieved the highest score
value of 0.9200 with sensitivity and specificity of 0.8735 and 0.9666
respectively. Hence, it provided the best estimate of the predictive
score in both methods and relatively balanced values for sensitiv-
ity and specificity. Performance assessments of classifiers in both
train-test methods were generally similar. The proposed system led
to significant improvemeernel achieved the highest score value of
0.9200 with sensitivity and specificity of 0.8735 and 0.9666 respec-
tively. Hence, it provided the best estimate of the predictive score
in both methods and relatively balanced values for sensitivity and
specificity. Performance assessments of classifiers in both train-
test methods were generally similar. The proposed system led to
significant improvements in classification performance using tradi-
tional classifiers, which indicates the robustness of the new system
and its ability to handle diverse PCG recordings and signal quality
conditions.
Table 4 shows several variants of the proposed method as they
rank by the official challenge score. It also shows the classifica-
tion methods used in each and also the sensitivity and specificity
values. The comparison of such variants should address not only
the score but also the balance between sensitivity and specificity
values. As can be observed, two variants of the proposed method
provide excellent performance metrics that outperform all previ-
ous methods applied on the same dataset in Table 5 except the
method by our group in [39] that uses deep learning. Furthermore,
other variants also appear to provide good performance metrics as
compared to the pervious methods applied to the same dataset.
Also, it should be noted that the sensitivity and specificity values
obtained for all entries from the new system are relatively close.
This indicate clear potential of the proposed system and indicate the
robustness of the used features which appear to work with various
traditional classification methods and parameters.
It is important to observe the effect of the class distribution of
the training data, which is a very important factor that determine
the quality of subsequent classification. In the challenge database,
the number of samples seems to be large but unfortunately these
samples are heavily imbalanced with much different numbers of
normal and abnormal recordings. This causes conventional classi-
fiers to often become biased toward the majority class and therefore
increase misclassification rate for the minority class. Therefore, the
performance of different classifiers was evaluated on the dataset by
using two validation methods. The first is the cross-validation train-
test method, which reduces the effect of over-fitting produced by
noisy records. The second is the local holdout method that simu-
lates real life analysis by building the model with 80 % of used data
and use the rest to evaluate the model.
Furthermore, the question comes of that possible data over-
fitting occurs in this study. Overfitting in the context of machine
learning refers to a system that models the details and noise in the
training data in a way that adversely affects performance on new
data. To investigate the extent of this problem in our study, the
results were reported in Tables 2 and 3 as the mean and standard
deviation of the outcome of 5-fold cross validation and 80-20 %
local holdout several experiments respectively. This approach fol-
lows the well-known random resampling technique that is used
to validate machine learning models. As can be observed, the stan-
dard deviations are less than 0.5 % of the mean value reported for all
results, which is considered small. The overall analysis of all results,
the mean of random variations in the cross-validation results was
0.786 % with median of 0.511 %, again both are fairly small. These
results indicate that the problem of overfitting may not have a sig-
nificant impact on the outcome of this study based on the small
variability in the different resampling experiments. Nevertheless,
it remains a concern to this study.
The results of the previous work are generally within a narrow
score metric range from 0.73 to 0.93 even they use different types
of features and classifiers as shown in Table 5. Also, it is noted
that the best scores were achieved with the most complex mod-
els. For example, Potes et al. [11] used final decision rule based
on combination of ensemble AdaBoost classifier and convolutional
neural network. Also, Homsi et al. [12] employed a nested set of
ensemble classifiers that includes cost-sensitive classifier (CSC),
LogitBoost (LB) and random forest (RF). Abduh et al. [39] applied
fractional Fourier transform based Mel-frequency spectral coeffi-
cients (FrFT-MFSC) and stacked auto-encoder deep neural network.
The proposed features still gave good result with the much simpler
classifier using SVM with cubic kernel. This indicates the potential
of the used features and their impact in simplifying the develop-
ment process of the classification system.
This study indicates the potential of FrFT-MFSC features for
phonocardiogram analysis where they represent the distribution of
PCG signal energy in a more effective manner even under different
noise and recording environment conditions. Applying FrFT-MFSC
features with different fractional orders provides a good tool to rep-
resent noisy PCG signal in different time-frequency planes where
they also preserve locality in both time and frequency. The use
of log-energy computed directly from the Mel-frequency spectral
coefficients maintains such time-frequency localization. The alter-
native yet more common use of discrete cosine transform for this
9. Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788 9
Table 5
Comparison of the performance metrics of the previous methods reported in the literature.
Study Classifier Sensitivity Specificity Score
Abduh et al [39]. Stacked Autoencoder Deep Neural Network 0.8930 0.9700 0.9315
Plesinger et al. [16] Fuzzy logic like approach 0.8690 0.9370 0.9030
Langley and Murray
[19]
Amplitude spectrum and wavelet entropy
threshold followed by decision tree
0.8690 0.9370 0.9030
Goda et al. [9] SVM 0.9308 0.8470 0.8870
Nabhan and Warriek
[17]
Nested ensemble of algorithm including
random forest, LogitBoost and cost-sensitive
classifier
0.8740 0.9140 0.8940
Abdollahpur et al. [18] Three feed-forward NNs 0.8883 0.8851 0.8867
Whitaker et al. [21] SVM 0.8867 0.8816 0.8841
Homsi et al. [12] Nested ensemble of algorithms including
random forest, LogitBoost and cost-sensitive
classifier
0.9440 0.8690 0.8840
Tschannen et al. [11] L2-SVM 0.9080 0.8320 0.8700
Potes et al. [3] Final decision rule based on AdaBoost
ensemble classifier and Convolutional neural
network
0.8800 0.8200 0.8500
Singh-Miller and
Singh-Miller [14]
Discriminative model based on random forest
regressor
0.8100 0.8900 0.8500
Potes et al. [3] Convolutional neural network (CNN) 0.7900 0.8600 0.8200
Vernekar et al. [15] Weighted ensemble classifier including four
AdaBoost ensemble and four ANN classifiers
0.7920 0.8430 0.8200
Potes et al. [3] AdaBoost- ensemble classifier 0.7000 0.8800 0.7900
Langley and Murray
[13]
Wavelet entropy threshold 0. 9500 0.6000 0.7800
Gokhale [8] Boosted trees ensemble classifier 0.7600
Grzegorczyk et al. [10] Conventional neural network and autoencoder 0.8300 0.6200 0.7300
computation projects the spectral energies into a new basis that
would not maintain such localization.
Noting that the dataset is significantly imbalanced, it is interest-
ing to observe from Table 4 that several variants of the proposed
method that performed very well used conventional classifiers such
as SVM and KNN that are not intended for imbalanced data. In
fact, we can see also from Table 5 that SVM was also reported to
perform well by several previous studies. This indicates that the
dataset has sufficiently limited statistical variability that allows
the characteristics of the underlying distribution of the abnor-
mal class to be learned by such classifiers from the available
number of samples. So, even though the use of ensemble clas-
sifiers is always recommended for imbalanced data in general
[37], the dataset used in our problem presents an interesting spe-
cial case as confirmed by our results and those from previous
work.
Although it may seem that the computation of 3040 features is a
lot to process, we aimed to analyze all this information to establish
first that they are useful. Moreover, given the relatively slow sam-
pling rate and the speed of today’s computers, processors or FPGAs,
it is practically possible to compute all such features in real-time
and at reasonable cost as well.
Even though the PCA does not provide binary in/out reduction
of features like the classical feature reduction approaches, one can
still gain insight into which features are more relevant by observing
their weights in the principal components that explain the 95 %
of data variance. For example, in our results, we found that 466
features had significant weights above 90 % of the maximum weight
in the principal component feature combination derived from PCA.
Fig. 4 shows the distribution of these significant features across
the different fractional orders. It indicates balanced importance for
features from all orders with slight advantage to order 1.0 then
0.95. Fig. 5 shows the distribution of the same features over the
different frames within the cardiac cycle. It clearly indicates that
the initial 600 ms part of the cardiac cycle is the most relevant to
the diagnosis. This range show more emphasis that coincides with
both the S1 and S2 parts of the PCG and to a lesser extent with the
systole part. So, the outcome from PCA gives further insight into
Fig. 4. The distribution of significant features in the principal component across
different values of the fractional order.
the physiological correlation of the set of features emphasized by
PCA based on weight that can be used to find out the most useful
of them.
3.4. Study limitations
Some difficulties were encountered in dealing with the data that
are common to all similar studies. First, the variations in heart rate
lead to temporal record length variations. This was addressed via
zero-padding of all recordings to the length of the longest one. This
maintains the same sampling rate while harmonizing frequency
resolution among all records, which is critical to our method given
its reliance on frequency domain features. Second, the inter-patient
differences made it challenging to build a model that generalizes
well across patients. This remains a challenge in biomedical sig-
nal measurement in general and there is not much to do to reduce
10. 10 Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788
Fig. 5. The distribution of significant features in the principal component across
different frames within the cardiac cycle.
such differences. Finally, differences introduced by heterogeneity
in the collection of the recordings can render a classifier trained on
one population much less effective when applied to another. Here,
even though the features proposed seem to capture the salient
features among all such populations as evident from the robust
performance among populations, the problem of overfitting still
cannot be ignored and therefore is considered among the potential
limitations of the present study. Such limitations are open research
points for future research.
4. Conclusions
A new approach is proposed to classify heart sounds. The main
contribution of this approach is the combination of the features
based on fractional Fourier transform based Mel-frequency spec-
tral coefficients and traditional classifiers that offer simplicity and
robustness against overfitting. The description of the proposed
methodology and its implementation details are presented. The
results of experimental verification using the database of the Phy-
sioNet/Computing in Cardiology Challenge 2016 indicate that the
presented approach and several variants of it performed well with
an evaluation score reaching 0.92 in the best results obtained using
a support vector machine classifier with cubic kernel. These results
confirm the robust performance obtained by the new system com-
pared to the previous work applied on the same data. This indicates
the potential of the new system for clinical utility.
Appendix A. Supplementary data
Supplementary material related to this article can be found,
in the online version, at doi:https://doi.org/10.1016/j.bspc.2019.
101788.
Declaration of Competing Interest
The authors have declared no conflict of interest.
References
[1] WHO, Cardiovascular Diseases World Statistics on WHO, 2017, last updated
May. www.who.int/mediacentre/factsheets/fs317/en/ (Accessed October 6,
2019).
[2] C. Jian, G. Xingming, X. Shouzhong, Study on the signification and method of
heart sound recognition, Foreign Med. Biomed. Eng. Fascicle 27 (2) (2004)
87–89.
[3] C. Potes, S. Parvaneh, A. Rahman, B. Conroy, Ensemble of feature-based and
deep learning-based classifiers for detection of abnormal heart sounds, Proc.
2016 Computing in Cardiology Conference (CinC) (2016) 621–624.
[4] J. Rubin, R. Abreu, A. Ganguli, S. Nelaturi, I. Matei, K. Sricharan, Classifying
heart sound recordings using deep convolutional neural networks and
mel-frequency cepstral coefficients, 2016 Computing in Cardiology
Conference (CinC) (2016) 813–816.
[5] G.D. Clifford, C. Liu, B. Moody, D. Springer, I. Silva, Q. Li, et al., Classification of
normal/abnormal heart sound recordings: the PhysioNet/Computing in
cardiology challenge 2016, Proc. 2016 Computing in Cardiology Conference
(CinC) (2016) 609–612.
[6] G.D. Clifford, C. Liu, B. Moody, J. Millet, S. Schmidt, Q. Li, et al., Recent advances
in heart sound analysis, Physiol. Meas. 38 (2017) E10.
[7] C. Liu, D. Springer, Q. Li, B. Moody, R.A. Juan, F.J. Chorro, et al., An open access
database for the evaluation of heart sound algorithms, Physiol. Meas. 37 (12)
(2016) 2181–2213.
[8] T. Gokhale, Machine learning based identification of pathological heart
sounds, 2016 Computing in Cardiology Conference (CinC) (2016) 553–556.
[9] M.A. Goda, P. Hajas, Morphological determination of pathological PCG signals
by time and frequency domain analysis, Proc. 2016 Computing in Cardiology
Conference (CinC) (2016) 1133–1136.
[10] I. Grzegorczyk, M. Soliński, A. Łepek Michałand Perka, J. Rosiński, J. Rymko, K.
Stkepień, et al., PCG classification using a neural network approach, Proc.
2016 Computing in Cardiology Conference (CinC) (2016) 1129–1132.
[11] M. Tschannen, T. Kramer, G. Marti, M. Heinzmann, T. Wiatowski, Heart sound
classification using deep structured features, Proc. 2016 Computing in
Cardiology Conference (CinC) (2016) 565–568.
[12] M.N. Homsi, N. Medina, M. Hernandez, N. Quintero, G. Perpiñan, A. Quintana,
et al., Automatic heart sound recording classification using a nested set of
ensemble algorithms, Proc. 2016 Computing in Cardiology Conference (CinC)
(2016) 817–820.
[13] P. Langley, A. Murray, Abnormal heart sounds detected from short duration
unsegmented phonocardiograms by wavelet entropy, Proc. 2016 Computing
in Cardiology Conference (CinC) (2016) 545–548.
[14] N.E. Singh-Miller, N. Singh-Miller, Using spectral acoustic features to identify
abnormal heart sounds, Proc. 2016 Computing in Cardiology Conference
(CinC) (2016) 557–560.
[15] S. Vernekar, S. Nair, D. Vijaysenan, R. Ranjan, A novel approach for
classification of normal/abnormal phonocardiogram recordings using
temporal signal analysis and machine learning, Proc. 2016 Computing in
Cardiology Conference (CinC) (2016) 1141–1144.
[16] F. Plesinger, I. Viscor, J. Halamek, J. Jurco, P. Jurak, Heart sounds analysis using
probability assessment, Physiol. Meas. 38 (8) (2017) 1685–1700.
[17] H.M. Nabhan, P. Warrick, Ensemble methods with outliers for
phonocardiogram classification, Physiol. Meas. 38 (8) (2017) 1631–1644.
[18] M. Abdollahpur, A. Ghaffari, S. Ghiasi, M.J. Mollakazemi, Detection of
pathological heart sounds, Physiol. Meas. 38 (8) (2017) 1616–1630.
[19] P. Langley, A. Murray, Heart sound classification from unsegmented
phonocardiograms, Physiol. Meas. 38 (8) (2017) 1658–1670.
[20] V. Maknickas, A. Maknickas, Recognition of normal-abnormal
phonocardiographic signals using deep convolutional neural networks and
mel-frequency spectral coefficients, Physiol. Meas. 38 (8) (2017) 1671–1684.
[21] B.M. Whitaker, P.B. Suresha, C. Liu, G. Clifford, D. Anderson, Combining sparse
coding and time-domain features for heart sound classification, Physiol. Meas.
38 (8) (2017) 1701–1729.
[22] S. Boll, Suppression of acoustic noise in speech using spectral subtraction,
IEEE Trans. Acoust. 27 (2) (1979) 113–120.
[23] M.J. Alhaddad, M.I. Kamel, M.M. Makary, H. Hargas, Y.M. Kadah, Spectral
subtraction denoising preprocessing block to improve P300-based
brain-computer interfacing, Biomed. Eng. Online 13 (1) (2014) 13–36.
[24] D.B. Springer, L. Tarassenko, G.D. Clifford, Logistic Regression-HSMM-Based
heart sound segmentation, IEEE Trans. Biomed. Eng. 63 (4) (2016) 822–832.
[25] D.P.W. Ellis, PLP, RASTA, MFCC And Inversion in Matlab, 2012, last updated.
www.ee.columbia.edu/∼dpwe/resources/matlab/rastamat/ (Accessed
October 16, 2017).
[26] Z. Abduh, M.A. Wahed, Y.M. Kadah, Robust computer-aided detection of
pulmonary nodules from chest computed tomography, J. Med. Imaging
Health Inform. 6 (3) (2016) 693–699.
[27] D. O’Shaughnessy, Speech Communications: Human and Machine, 2nd ed.,
Wiley-IEEE Press, 1999.
[28] A.C. McBride, F.H. Kerr, On Namias’s fractional Fourier transforms, IMA J. Appl.
Math. 39 (2) (1987) 159–175.
[29] L.B. Almeida, The fractional Fourier transform and time-frequency
representations, Ieee Trans. Signal Process. 42 (11) (1994) 3084–3091.
[30] C. Candan, M.A. Kutay, H.M. Ozaktas, The discrete fractional Fourier
transform, IEEE Trans. Signal Process. 48 (5) (2000) 1329–1337.
[31] S.-C. Pei, M.-H. Yeh, C.-C. Tseng, Discrete fractional Fourier transform based on
orthogonal projections, IEEE Trans. Signal Process. 47 (5) (1999) 1335–1348.
[32] A. Clark, C. Fox, S. Lappin, Handbook of Computational Linguistics and Natural
Language Processing, Wiley-Blackwell, 2010.
[33] J.R. Quinlan, Induction of decision trees, Mach. Learn. 1 (1) (1986) 81–106.
[34] M. Galar, A. Fernandez, E. Barrenechea, H. Bustince, F. Herrera, A review on
ensembles for the class imbalance problem: bagging-, boosting-, and
11. Z. Abduh, E.A. Nehary, M. Abdel Wahed et al. / Biomedical Signal Processing and Control 57 (2020) 101788 11
hybrid-based approaches, IEEE Trans. Syst. Man Cybern. Part C 42 (4) (2012)
463–484.
[35] S. Russell, P. Norvig, Artificial Intelligence: A Modern Approach, 3rd edition,
Prentice Hall, Englewood Cliffs, New Jersey, 2010.
[36] S. Sayad, An Introduction to Data Science, Last Updated, 2017 (Accessed
October 16, 2017) http://www.saedsayad.com/further readings.htm.
[37] V. López, A. Fernández, S. Garc’ia, V. Palade, F. Herrera, An insight into
classification with imbalanced data: empirical results and current trends on
using data intrinsic characteristics, Inf. Sci. 250 (2013) 113–141.
[38] K.S. Rao, S.G. Koolagudi, Robust emotion recognition using spectral and
prosodic features, Springer Briefs Speech Technol. (2013) 109.
[39] Z. Abduh, E.A. Nehary, M.A. Wahed, Y.M. Kadah, Classification of Heart Sounds
Using Fractional Fourier Transform Based Mel-Frequency Spectral
Coefficients and Stacked Autoencoder Deep Neural Network, J. Med. Imaging
Health Inform. 9 (1) (2019) 1–8.
[40] The emergence of artificial intelligence and machine learning algorithms in
healthcare: recommendations to support governance and regulation, in:
Position Paper Prepared by BSI and AAMI. The AAMI/BSI Workshop on
Artificial Intelligence and Machine Learning Algorithms in Health Technology,
Arlington, Virginia, May, 2019.