Developed Project with 3 more colleagues for Pneumonia Detection from Chest X-ray images using Convolutional Neural Network. Used confusion matrix, Recall, Precision for check the model performance on testing Data
Chest X-ray Pneumonia Classification with Deep LearningBaoTramDuong2
This document discusses using deep learning models to classify chest x-ray images as either normal or pneumonia. It obtained a dataset of over 5,800 pediatric chest x-rays from a Chinese hospital. Various deep learning models were explored, including multilayer perceptrons, convolutional neural networks, and transfer learning with VGG16, which achieved 92% validation accuracy. The document recommends future work such as distinguishing between viral and bacterial pneumonia and combining models with SVM. It also discusses recommendations to reduce childhood pneumonia prevalence.
Pneumonia Classification using Transfer LearningTushar Dalvi
Pneumonia can be life-threatening for people with weak immune systems, in which the alveoli filled with fluid that makes it hard to pass oxygen throughout the bloodstream. Detecting pneumonia is from a chest X-ray is not only expansive but also time-consuming for normal people. Throughout this research introduced a machine learning technique to classify pneumonia from Chest X-ray Images. Most of the medical datasets having class imbalance issues in the dataset. The Data augmentation technique used to reduce the class imbalance from the dataset, Horizontal Flip, width shift and height shift techniques used to complete the augmentation technique. Used VGG19 as a base architecture and ImageNet weights added for the transfer learning approach, also Removing initial layers and adding
some more dense layers helped to discover new possibilities. After testing the proposed model on testing data, we are able to achieve 98% recall and 82% of precision. As compare with state of the art technique, the proposed method able to achieve high
recall but that compromises with Precision.
classify images from the CIFAR-10 dataset. The dataset consists of airplanes, dogs, cats, and other objects.we'll preprocess the images, then train a convolutional neural network on all the samples. The images need to be normalized and the labels need to be one-hot encoded.
PPT on BRAIN TUMOR detection in MRI images based on IMAGE SEGMENTATION khanam22
The document presents three methods for tumor detection in MRI images: 1) K-means clustering with watershed algorithm, 2) Optimized K-means using genetic algorithm, and 3) Optimized C-means using genetic algorithm. It evaluates each method, finding that C-means clustering with genetic algorithm most accurately detects tumors by assigning data points to multiple clusters and finding the optimal solution in less time. The proposed approach successfully detects tumors with high accuracy, identifies the tumor area and internal structure, and provides a colorized output image.
A comprehensive tutorial on Convolutional Neural Networks (CNN) which talks about the motivation behind CNNs and Deep Learning in general, followed by a description of the various components involved in a typical CNN layer. It explains the theory involved with the different variants used in practice and also, gives a big picture of the whole network by putting everything together.
Next, there's a discussion of the various state-of-the-art frameworks being used to implement CNNs to tackle real-world classification and regression problems.
Finally, the implementation of the CNNs is demonstrated by implementing the paper 'Age ang Gender Classification Using Convolutional Neural Networks' by Hassner (2015).
Machine Learning - Breast Cancer DiagnosisPramod Sharma
Machine learning is helping in making smart decisions faster. In this presentation measurements carried out on FNAC was analysed. The results were validated using 20 percent of the data. The data used for POC is from UCI Repository/
Developed Project with 3 more colleagues for Pneumonia Detection from Chest X-ray images using Convolutional Neural Network. Used confusion matrix, Recall, Precision for check the model performance on testing Data
Chest X-ray Pneumonia Classification with Deep LearningBaoTramDuong2
This document discusses using deep learning models to classify chest x-ray images as either normal or pneumonia. It obtained a dataset of over 5,800 pediatric chest x-rays from a Chinese hospital. Various deep learning models were explored, including multilayer perceptrons, convolutional neural networks, and transfer learning with VGG16, which achieved 92% validation accuracy. The document recommends future work such as distinguishing between viral and bacterial pneumonia and combining models with SVM. It also discusses recommendations to reduce childhood pneumonia prevalence.
Pneumonia Classification using Transfer LearningTushar Dalvi
Pneumonia can be life-threatening for people with weak immune systems, in which the alveoli filled with fluid that makes it hard to pass oxygen throughout the bloodstream. Detecting pneumonia is from a chest X-ray is not only expansive but also time-consuming for normal people. Throughout this research introduced a machine learning technique to classify pneumonia from Chest X-ray Images. Most of the medical datasets having class imbalance issues in the dataset. The Data augmentation technique used to reduce the class imbalance from the dataset, Horizontal Flip, width shift and height shift techniques used to complete the augmentation technique. Used VGG19 as a base architecture and ImageNet weights added for the transfer learning approach, also Removing initial layers and adding
some more dense layers helped to discover new possibilities. After testing the proposed model on testing data, we are able to achieve 98% recall and 82% of precision. As compare with state of the art technique, the proposed method able to achieve high
recall but that compromises with Precision.
classify images from the CIFAR-10 dataset. The dataset consists of airplanes, dogs, cats, and other objects.we'll preprocess the images, then train a convolutional neural network on all the samples. The images need to be normalized and the labels need to be one-hot encoded.
PPT on BRAIN TUMOR detection in MRI images based on IMAGE SEGMENTATION khanam22
The document presents three methods for tumor detection in MRI images: 1) K-means clustering with watershed algorithm, 2) Optimized K-means using genetic algorithm, and 3) Optimized C-means using genetic algorithm. It evaluates each method, finding that C-means clustering with genetic algorithm most accurately detects tumors by assigning data points to multiple clusters and finding the optimal solution in less time. The proposed approach successfully detects tumors with high accuracy, identifies the tumor area and internal structure, and provides a colorized output image.
A comprehensive tutorial on Convolutional Neural Networks (CNN) which talks about the motivation behind CNNs and Deep Learning in general, followed by a description of the various components involved in a typical CNN layer. It explains the theory involved with the different variants used in practice and also, gives a big picture of the whole network by putting everything together.
Next, there's a discussion of the various state-of-the-art frameworks being used to implement CNNs to tackle real-world classification and regression problems.
Finally, the implementation of the CNNs is demonstrated by implementing the paper 'Age ang Gender Classification Using Convolutional Neural Networks' by Hassner (2015).
Machine Learning - Breast Cancer DiagnosisPramod Sharma
Machine learning is helping in making smart decisions faster. In this presentation measurements carried out on FNAC was analysed. The results were validated using 20 percent of the data. The data used for POC is from UCI Repository/
- Pakistan has the highest incidence of breast cancer in Asia, with over 83,000 new cases reported annually. Breast cancer is the leading cause of cancer mortality among females in Pakistan.
- Early detection of breast cancer significantly improves survival rates, with a 100% 5-year survival rate for cancers detected early. However, Pakistan currently lacks a system for widespread breast cancer screening.
- Artificial intelligence can help by assisting oncologists in diagnosing breast cancer. A machine learning model trained on breast cancer data achieved over 96% accuracy in predicting malignant tumors, which could help detect cancers earlier.
This document describes a system for detecting brain tumors in MRI images using image segmentation. It discusses how existing manual detection of tumors is difficult due to noise and requires many days. The proposed system applies preprocessing like filtering and grayscale conversion. It then uses image segmentation techniques to detect tumor edges and boundaries. Features are extracted and classification is used to differentiate between normal and tumor images, helping doctors detect tumors earlier. The system is implemented in MATLAB and aims to overcome difficulties in early tumor detection.
This document describes a project on age and gender detection using deep learning and convolutional neural networks. The objectives are to pretrain a model using the UTKFace dataset to detect age and gender from facial images. The methodology involves data preprocessing like grayscale conversion, resizing, normalization. A CNN model is built with convolutional blocks and fully connected layers. The model is compiled with binary cross entropy loss for gender classification and mean absolute error for age detection. The trained model is tested and deployed using streamlit. Real-world use cases discussed are advertising based on targeted audiences and an Android app that detects age from photos.
This document provides an introduction to deep learning. It defines artificial intelligence, machine learning, data science, and deep learning. Machine learning is a subfield of AI that gives machines the ability to improve performance over time without explicit human intervention. Deep learning is a subfield of machine learning that builds artificial neural networks using multiple hidden layers, like the human brain. Popular deep learning techniques include convolutional neural networks, recurrent neural networks, and autoencoders. The document discusses key components and hyperparameters of deep learning models.
details about brain tumor
literature survey on many reference papers related to brain tumor detection using various techniques
our proposed novel methodology for brain tumor detection
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSINGDharshika Shreeganesh
Image processing is an active research area in which medical image processing is a highly challenging field. Medical imaging
techniques are used to image the inner portions of the human body for medical diagnosis. Brain tumor is a serious life altering
disease condition. Image segmentation plays a significant role in image processing as it helps in the extraction of suspicious regions
from the medical images. In this paper we have proposed segmentation of brain MRI image using K-means clustering algorithm
followed by morphological filtering which avoids the misclustered regions that can inevitably be formed after segmentation of the brain MRI image for detection of tumor location.
This document provides an overview of deep learning in neural networks. It defines deep learning as using artificial neural networks with multiple levels that learn higher-level concepts from lower-level ones. It describes how deep learning networks have many layers that build improved feature spaces, with earlier layers learning simple features that are combined in later layers. Deep learning networks are categorized as unsupervised or supervised, or hybrids. Common deep learning architectures like deep neural networks, deep belief networks, convolutional neural networks, and deep Boltzmann machines are also described. The document explains why GPUs are useful for deep learning due to their throughput-oriented design that speeds up model training.
The presentation is made on CNN's which is explained using the image classification problem, the presentation was prepared in perspective of understanding computer vision and its applications. I tried to explain the CNN in the most simple way possible as for my understanding. This presentation helps the beginners of CNN to have a brief idea about the architecture and different layers in the architecture of CNN with the example. Please do refer the references in the last slide for a better idea on working of CNN. In this presentation, I have also discussed the different types of CNN(not all) and the applications of Computer Vision.
Lung Cancer Detection Using Convolutional Neural NetworkIRJET Journal
This document describes a study that uses a convolutional neural network (CNN) to classify lung cancer in CT scans. The CNN model is trained on a dataset of 1018 patient CT scans containing annotations of lung nodules as benign or malignant. The CNN architecture includes convolution layers to extract features, max pooling layers to reduce computations, dropout layers to prevent overfitting, and fully connected layers to classify scans. The model achieves a 65% accuracy on the training set at detecting cancer in new CT scans. The CNN is integrated into a web application to allow doctors to efficiently analyze scans for lung cancer.
Transfer Learning for the Detection and Classification of traditional pneumon...Yusuf Brima
A presentation of my MSc in Mathematical Sciences thesis at the African Institute of Mathematical Sciences (AIMS), Rwanda. This presentation explores the application of Deep Transfer Learning towards the diagnosis and classification of traditional pneumonia and pneumonia induced from COVID-19 using chest X-ray images.
Deep learning is a type of machine learning that uses neural networks inspired by the human brain. It has been successfully applied to problems like image recognition, speech recognition, and natural language processing. Deep learning requires large datasets, clear goals, computing power, and neural network architectures. Popular deep learning models include convolutional neural networks and recurrent neural networks. Researchers like Geoffry Hinton and companies like Google have advanced the field through innovations that have won image recognition challenges. Deep learning will continue solving harder artificial intelligence problems by learning from massive amounts of data.
Performance analysis of automated brain tumor detection from MR imaging and CT scan using basic image processing techniques based on various hard and soft computing has been performed in our work. Moreover, we applied six traditional classifiers to detect brain tumor in the images. Then we applied CNN for brain tumor detection to include deep learning method in our work. We compared the result of the traditional one having the best accuracy (SVM) with the result of CNN. Furthermore, our work presents a generic method of tumor detection and extraction of its various features.
Automatic gender and age classification has become quite relevant in the rise of social media platforms. However, the existing methods have not been completely successful in achieving this. Through this project, an attempt has been made to determine the gender and age based on a frame of the person. This is done by using deep learning, OpenCV which is capable of processing the real-time frames. This frame is given as input and the predicted gender and age are given as output. It is difficult to predict the exact age of a person using one frame due the facial expressions, lighting, makeup and so on so for this purpose various age ranges are taken, and the predicted age falls in one of them. The Adience dataset is used as it is a benchmark for face photos and includes various real-world imaging conditions like noise, lighting etc.
Convolutional neural network (CNN / ConvNet's) is a part of Computer Vision. Machine Learning Algorithm. Image Classification, Image Detection, Digit Recognition, and many more. https://technoelearn.com .
Application of-image-segmentation-in-brain-tumor-detectionMyat Myint Zu Thin
This document discusses applications of image segmentation in brain tumor detection. It begins by defining brain tumors and different types. It then discusses various image segmentation methods that can be used for brain tumor segmentation, including k-means clustering, region-based watershed algorithm, region growing, and active contour methods. It demonstrates how these methods can be implemented in Python for segmenting tumors from MRI images. The document also discusses computer-aided diagnosis systems and the roles of artificial intelligence and machine learning in medical image analysis and cancer diagnosis using image processing.
Machine Learning - Convolutional Neural NetworkRichard Kuo
The document provides an overview of convolutional neural networks (CNNs) for visual recognition. It discusses the basic concepts of CNNs such as convolutional layers, activation functions, pooling layers, and network architectures. Examples of classic CNN architectures like LeNet-5 and AlexNet are presented. Modern architectures such as Inception and ResNet are also discussed. Code examples for image classification using TensorFlow, Keras, and Fastai are provided.
Image classification with Deep Neural NetworksYogendra Tamang
This document discusses image classification using deep neural networks. It provides background on image classification and convolutional neural networks. The document outlines techniques like activation functions, pooling, dropout and data augmentation to prevent overfitting. It summarizes a paper on ImageNet classification using CNNs with multiple convolutional and fully connected layers. The paper achieved state-of-the-art results on ImageNet in 2010 and 2012 by training CNNs on a large dataset using multiple GPUs.
This document discusses different types of electromagnetic radiation and their uses in digital image processing. It covers gamma rays, X-rays, ultraviolet rays, visible and infrared rays, microwaves, and radio bands. Applications described include medical imaging techniques like MRI, industrial inspection, astronomy, remote sensing, and law enforcement applications like license plate and fingerprint recognition. Radar imaging is also discussed as a key application using microwaves.
The document describes a student project that aims to classify and detect lung disorders from chest x-rays using generative adversarial networks (GANs). It provides an overview of the problem statement, literature review on existing methodologies, and the proposed system architecture. The problem is that early diagnosis of lung diseases is important but current methods are time-consuming. The project aims to develop a hybrid GAN model to overcome issues in existing models and accurately detect lung disorders from x-rays. It will classify diseases like pneumonia and lung cancer.
Pneumonia Detection Using Convolutional Neural Network WritersIRJET Journal
This document describes a study that developed a convolutional neural network (CNN) model to detect and classify pneumonia from chest x-ray images without any training. The model was able to extract relevant features from chest x-ray images and determine if a patient has pneumonia. The CNN model achieved high accuracy for pneumonia classification compared to other advanced methods that rely on transfer learning or manual feature engineering. The study used a dataset of 5,500 chest x-ray images and performed image processing techniques like background removal and cropping before extracting features with the CNN and classifying images.
- Pakistan has the highest incidence of breast cancer in Asia, with over 83,000 new cases reported annually. Breast cancer is the leading cause of cancer mortality among females in Pakistan.
- Early detection of breast cancer significantly improves survival rates, with a 100% 5-year survival rate for cancers detected early. However, Pakistan currently lacks a system for widespread breast cancer screening.
- Artificial intelligence can help by assisting oncologists in diagnosing breast cancer. A machine learning model trained on breast cancer data achieved over 96% accuracy in predicting malignant tumors, which could help detect cancers earlier.
This document describes a system for detecting brain tumors in MRI images using image segmentation. It discusses how existing manual detection of tumors is difficult due to noise and requires many days. The proposed system applies preprocessing like filtering and grayscale conversion. It then uses image segmentation techniques to detect tumor edges and boundaries. Features are extracted and classification is used to differentiate between normal and tumor images, helping doctors detect tumors earlier. The system is implemented in MATLAB and aims to overcome difficulties in early tumor detection.
This document describes a project on age and gender detection using deep learning and convolutional neural networks. The objectives are to pretrain a model using the UTKFace dataset to detect age and gender from facial images. The methodology involves data preprocessing like grayscale conversion, resizing, normalization. A CNN model is built with convolutional blocks and fully connected layers. The model is compiled with binary cross entropy loss for gender classification and mean absolute error for age detection. The trained model is tested and deployed using streamlit. Real-world use cases discussed are advertising based on targeted audiences and an Android app that detects age from photos.
This document provides an introduction to deep learning. It defines artificial intelligence, machine learning, data science, and deep learning. Machine learning is a subfield of AI that gives machines the ability to improve performance over time without explicit human intervention. Deep learning is a subfield of machine learning that builds artificial neural networks using multiple hidden layers, like the human brain. Popular deep learning techniques include convolutional neural networks, recurrent neural networks, and autoencoders. The document discusses key components and hyperparameters of deep learning models.
details about brain tumor
literature survey on many reference papers related to brain tumor detection using various techniques
our proposed novel methodology for brain tumor detection
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSINGDharshika Shreeganesh
Image processing is an active research area in which medical image processing is a highly challenging field. Medical imaging
techniques are used to image the inner portions of the human body for medical diagnosis. Brain tumor is a serious life altering
disease condition. Image segmentation plays a significant role in image processing as it helps in the extraction of suspicious regions
from the medical images. In this paper we have proposed segmentation of brain MRI image using K-means clustering algorithm
followed by morphological filtering which avoids the misclustered regions that can inevitably be formed after segmentation of the brain MRI image for detection of tumor location.
This document provides an overview of deep learning in neural networks. It defines deep learning as using artificial neural networks with multiple levels that learn higher-level concepts from lower-level ones. It describes how deep learning networks have many layers that build improved feature spaces, with earlier layers learning simple features that are combined in later layers. Deep learning networks are categorized as unsupervised or supervised, or hybrids. Common deep learning architectures like deep neural networks, deep belief networks, convolutional neural networks, and deep Boltzmann machines are also described. The document explains why GPUs are useful for deep learning due to their throughput-oriented design that speeds up model training.
The presentation is made on CNN's which is explained using the image classification problem, the presentation was prepared in perspective of understanding computer vision and its applications. I tried to explain the CNN in the most simple way possible as for my understanding. This presentation helps the beginners of CNN to have a brief idea about the architecture and different layers in the architecture of CNN with the example. Please do refer the references in the last slide for a better idea on working of CNN. In this presentation, I have also discussed the different types of CNN(not all) and the applications of Computer Vision.
Lung Cancer Detection Using Convolutional Neural NetworkIRJET Journal
This document describes a study that uses a convolutional neural network (CNN) to classify lung cancer in CT scans. The CNN model is trained on a dataset of 1018 patient CT scans containing annotations of lung nodules as benign or malignant. The CNN architecture includes convolution layers to extract features, max pooling layers to reduce computations, dropout layers to prevent overfitting, and fully connected layers to classify scans. The model achieves a 65% accuracy on the training set at detecting cancer in new CT scans. The CNN is integrated into a web application to allow doctors to efficiently analyze scans for lung cancer.
Transfer Learning for the Detection and Classification of traditional pneumon...Yusuf Brima
A presentation of my MSc in Mathematical Sciences thesis at the African Institute of Mathematical Sciences (AIMS), Rwanda. This presentation explores the application of Deep Transfer Learning towards the diagnosis and classification of traditional pneumonia and pneumonia induced from COVID-19 using chest X-ray images.
Deep learning is a type of machine learning that uses neural networks inspired by the human brain. It has been successfully applied to problems like image recognition, speech recognition, and natural language processing. Deep learning requires large datasets, clear goals, computing power, and neural network architectures. Popular deep learning models include convolutional neural networks and recurrent neural networks. Researchers like Geoffry Hinton and companies like Google have advanced the field through innovations that have won image recognition challenges. Deep learning will continue solving harder artificial intelligence problems by learning from massive amounts of data.
Performance analysis of automated brain tumor detection from MR imaging and CT scan using basic image processing techniques based on various hard and soft computing has been performed in our work. Moreover, we applied six traditional classifiers to detect brain tumor in the images. Then we applied CNN for brain tumor detection to include deep learning method in our work. We compared the result of the traditional one having the best accuracy (SVM) with the result of CNN. Furthermore, our work presents a generic method of tumor detection and extraction of its various features.
Automatic gender and age classification has become quite relevant in the rise of social media platforms. However, the existing methods have not been completely successful in achieving this. Through this project, an attempt has been made to determine the gender and age based on a frame of the person. This is done by using deep learning, OpenCV which is capable of processing the real-time frames. This frame is given as input and the predicted gender and age are given as output. It is difficult to predict the exact age of a person using one frame due the facial expressions, lighting, makeup and so on so for this purpose various age ranges are taken, and the predicted age falls in one of them. The Adience dataset is used as it is a benchmark for face photos and includes various real-world imaging conditions like noise, lighting etc.
Convolutional neural network (CNN / ConvNet's) is a part of Computer Vision. Machine Learning Algorithm. Image Classification, Image Detection, Digit Recognition, and many more. https://technoelearn.com .
Application of-image-segmentation-in-brain-tumor-detectionMyat Myint Zu Thin
This document discusses applications of image segmentation in brain tumor detection. It begins by defining brain tumors and different types. It then discusses various image segmentation methods that can be used for brain tumor segmentation, including k-means clustering, region-based watershed algorithm, region growing, and active contour methods. It demonstrates how these methods can be implemented in Python for segmenting tumors from MRI images. The document also discusses computer-aided diagnosis systems and the roles of artificial intelligence and machine learning in medical image analysis and cancer diagnosis using image processing.
Machine Learning - Convolutional Neural NetworkRichard Kuo
The document provides an overview of convolutional neural networks (CNNs) for visual recognition. It discusses the basic concepts of CNNs such as convolutional layers, activation functions, pooling layers, and network architectures. Examples of classic CNN architectures like LeNet-5 and AlexNet are presented. Modern architectures such as Inception and ResNet are also discussed. Code examples for image classification using TensorFlow, Keras, and Fastai are provided.
Image classification with Deep Neural NetworksYogendra Tamang
This document discusses image classification using deep neural networks. It provides background on image classification and convolutional neural networks. The document outlines techniques like activation functions, pooling, dropout and data augmentation to prevent overfitting. It summarizes a paper on ImageNet classification using CNNs with multiple convolutional and fully connected layers. The paper achieved state-of-the-art results on ImageNet in 2010 and 2012 by training CNNs on a large dataset using multiple GPUs.
This document discusses different types of electromagnetic radiation and their uses in digital image processing. It covers gamma rays, X-rays, ultraviolet rays, visible and infrared rays, microwaves, and radio bands. Applications described include medical imaging techniques like MRI, industrial inspection, astronomy, remote sensing, and law enforcement applications like license plate and fingerprint recognition. Radar imaging is also discussed as a key application using microwaves.
The document describes a student project that aims to classify and detect lung disorders from chest x-rays using generative adversarial networks (GANs). It provides an overview of the problem statement, literature review on existing methodologies, and the proposed system architecture. The problem is that early diagnosis of lung diseases is important but current methods are time-consuming. The project aims to develop a hybrid GAN model to overcome issues in existing models and accurately detect lung disorders from x-rays. It will classify diseases like pneumonia and lung cancer.
Pneumonia Detection Using Convolutional Neural Network WritersIRJET Journal
This document describes a study that developed a convolutional neural network (CNN) model to detect and classify pneumonia from chest x-ray images without any training. The model was able to extract relevant features from chest x-ray images and determine if a patient has pneumonia. The CNN model achieved high accuracy for pneumonia classification compared to other advanced methods that rely on transfer learning or manual feature engineering. The study used a dataset of 5,500 chest x-ray images and performed image processing techniques like background removal and cropping before extracting features with the CNN and classifying images.
The document describes using a convolutional neural network with the VGG16 architecture to classify lung cancer CT scan images into 4 classes: large cell carcinoma, squamous cell carcinoma, adenocarcinoma, and normal lungs. The model is trained on a dataset of 1000 CT scan images from Kaggle and achieves an AUC of 0.94, indicating high accuracy in identifying different types of lung cancer. This CNN model with pre-trained VGG16 weights provides an effective approach for classifying lung cancer images and could help enable early diagnosis and treatment of lung cancer.
Class imbalance is a pervasive issue in the field of disease classification from
medical images. It is necessary to balance out the class distribution while training a model. However, in the case of rare medical diseases, images from affected
patients are much harder to come by compared to images from non-affected
patients, resulting in unwanted class imbalance. Various processes of tackling
class imbalance issues have been explored so far, each having its fair share of
drawbacks. In this research, we propose an outlier detection based image classification technique which can handle even the most extreme case of class imbalance. We have utilized a dataset of malaria parasitized and uninfected cells. An
autoencoder model titled AnoMalNet is trained with only the uninfected cell images at the beginning and then used to classify both the affected and non-affected
cell images by thresholding a loss value. We have achieved an accuracy, precision, recall, and F1 score of 98.49%, 97.07%, 100%, and 98.52% respectively,
performing better than large deep learning models and other published works.
As our proposed approach can provide competitive results without needing the
disease-positive samples during training, it should prove to be useful in binary
disease classification on imbalanced datasets.
Performance Comparison Analysis for Medical Images Using Deep Learning Approa...IRJET Journal
This document discusses and compares several deep learning approaches for analyzing medical images, specifically chest x-rays. It first provides an abstract that outlines comparing existing technologies for analyzing chest x-rays using deep learning. It then reviews literature on models like convolutional neural networks (CNN), fully convolutional networks (FCN), lookup-based convolutional neural networks (LCNN), and deep cascade of convolutional neural networks (DCCNN) that have been applied to tasks like image segmentation, classification, and quality assessment of medical images. The document compares the performance of these models on different medical image datasets based on accuracy metrics.
Pneumonia Detection Using Deep Learning and Transfer LearningIRJET Journal
This document presents research on using deep learning and transfer learning techniques to detect pneumonia from chest x-ray images. The researchers trained several models, including CNNs, DenseNet, VGG-16, ResNet, and InceptionNet on a dataset of chest x-rays labeled as normal or pneumonia. The models achieved accuracy in detecting pneumonia of 89.6-97%, depending on the specific model. The researchers found that deep learning approaches like these have significant potential to improve the accuracy and efficiency of pneumonia diagnosis compared to traditional methods. Overall, the study demonstrated promising results for using machine and deep learning to classify medical images and detect health conditions like pneumonia.
APPLICATION OF CNN MODEL ON MEDICAL IMAGEIRJET Journal
The document discusses using convolutional neural network (CNN) models to detect diseases from medical images such as chest X-rays. It describes how CNN models can be trained on large labeled datasets of chest X-rays to learn patterns and features that indicate diseases. The document then evaluates several CNN architectures - including VGG-16, ResNet, DenseNet, and InceptionNet - for classifying chest X-rays as normal or infected. It finds these models achieve high accuracy, with metrics like accuracy over 89% and AUC over 0.94. In conclusion, deep learning models show promising results for automated disease detection from medical images.
Cervical Cancer Detection: An Enhanced Approach through Transfer Learning and...IRJET Journal
This document presents research on using the DenseNet169 deep learning model for cervical cancer detection. The researcher trained and tested the model on a large cervical cell image dataset from Kaggle. Through data preprocessing like augmentation and normalization, and transfer learning by fine-tuning a DenseNet pre-trained on ImageNet, the model achieved 95.27% accuracy in classifying five cervical cell types. Evaluation of the model showed high average precision, recall, and F1-score, demonstrating its ability to correctly classify different cervical cell images. The research highlights the potential of deep learning models for automating cervical cancer screening and improving early detection.
This document summarizes Mansi Chowkkar's MSc research project on detecting breast cancer from histopathological images using deep learning and transfer learning. The research aims to classify images as malignant or benign with high accuracy and efficiency. It implements CNN and DenseNet-121 models, with the latter using transfer learning with pre-trained ImageNet weights. The research achieved 90.9% test accuracy with CNN and 88.03% accuracy with transfer learning. Related work discusses deep learning applications in healthcare, image processing techniques, prior research on DenseNet, and the use of transfer learning for medical image classification with limited data.
Enhancing Pneumonia Detection: A Comparative Study of CNN, DenseNet201, and V...IRJET Journal
This document presents a comparative study evaluating the performance of three deep learning models - a custom CNN, DenseNet201, and VGG16 - in classifying chest X-ray images to detect pneumonia. The CNN model achieved the best performance with 80% accuracy, comparable to human radiologists. A chest X-ray dataset was used to train and evaluate the models. VGG16 consistently outperformed the other models, though all models showed potential for improving pneumonia diagnosis through rapid and accurate analysis of medical images using deep learning techniques.
This document presents a model to detect and classify brain tumors using watershed algorithm for image segmentation and convolutional neural networks (CNN). The model takes MRI images as input, pre-processes the images by converting them to grayscale and removing noise, then uses watershed algorithm for image segmentation and CNN for tumor classification. The CNN architecture achieves classification of three tumor types. Previous related works that also used deep learning methods for brain tumor detection and classification are discussed. The proposed system methodology involves inputting MRI images, pre-processing, segmentation using watershed algorithm, and classification of tumorous vs non-tumorous cells using CNN.
Despite the availability of radiology devices in some health care centers, thorax diseases are considered as one of the most common health problems, especially in rural areas. By exploiting the power of the Internet of things and specific platforms to analyze a large volume of medical data, the health of a patient could be improved earlier. In this paper, the proposed model is based on pre-trained ResNet-50 for diagnosing thorax diseases. Chest x-ray images are cropped to extract the rib cage part from the chest radiographs. ResNet-50 was re-train on Chest x-ray14 dataset where a chest radiograph images are inserted into the model to determine if the person is healthy or not. In the case of an unhealthy patient, the model can classify the disease into one of the fourteen chest diseases. The results show the ability of ResNet-50 in achieving impressive performance in classifying thorax diseases.
The current big challenge facing radiologists in healthcare is the automatic detection and classification of masses in breast mammogram images. In the last few years, many researchers have proposed various solutions to this problem. These solutions are effectively dependent and work on annotated breast image data. But these solutions fail when applied to unlabeled and non-annotated breast image data. Therefore, this paper provides the solution to this problem with the help of a neural network that considers any kind of unlabeled data for its procedure. In this solution, the algorithm automatically extracts tumors in images using a segmentation approach, and after that, the features of the tumor are extracted for further processing. This approach used a double thresholding-based segmentation technique to obtain a perfect location of the tumor region, which was not possible in existing techniques in the literature. The experimental results also show that the proposed algorithm provides better accuracy compared to the accuracy of existing algorithms in the literature.
This document describes a study that used deep learning models to diagnose Alzheimer's disease using MRI scans. Specifically, it used convolutional neural networks (CNNs) trained on MRI data from the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset. The study achieved an F1-score of 89% for classifying Alzheimer's vs normal cognition. It also used a CycleGAN for data augmentation, which generated additional synthetic MRI images and improved the CNN classification accuracy to 95% F1-score. The document provides background on Alzheimer's diagnosis, CNNs, GANs including CycleGAN, related work applying machine learning to Alzheimer's diagnosis, and the methodology used in this study for data preprocessing, model training, and evaluating the impact of GAN
Breast cancer classification with histopathological image based on machine le...IJECEIAES
This document summarizes a study that used pre-trained convolutional neural network (CNN) models to classify breast cancer histopathology images as benign or malignant. Five CNN architectures - ResNet-50, VGG-19, Inception-V3, AlexNet, and ResNet-50 as a feature extractor combined with random forest and k-nearest neighbors classifiers - were evaluated on the publicly available BreakHis dataset. The ResNet-50 network achieved the highest test accuracy of 97% for classifying the breast cancer images.
BREAST CANCER DIAGNOSIS USING MACHINE LEARNING ALGORITHMS –A SURVEYijdpsjournal
Breast cancer has become a common factor now-a-days. Despite the fact, not all general hospitals
have the facilities to diagnose breast cancer through mammograms. Waiting for diagnosing a breast
cancer for a long time may increase the possibility of the cancer spreading. Therefore a computerized
breast cancer diagnosis has been developed to reduce the time taken to diagnose the breast cancer and
reduce the death rate. This paper summarizes the survey on breast cancer diagnosis using various machine
learning algorithms and methods, which are used to improve the accuracy of predicting cancer. This survey
can also help us to know about number of papers that are implemented to diagnose the breast cancer.
This document summarizes a study that developed an attention-based deep learning model to detect diabetic retinopathy from fundus images. The proposed model used an InceptionV3 architecture with additional convolutional layers and an attention mechanism. On a dataset of 35,000 fundus images, the model achieved 94.3% validation accuracy, an improvement over previous models. Activation heatmaps showed the model learned important retinal features. The study demonstrated that deep learning can effectively detect diabetic retinopathy from images and may help with early diagnosis.
This document is a research project submission for a Master's in Data Analytics. It proposes using a convolutional neural network to classify multiple types of skin lesions from dermoscopic images. Specifically, it will use a modified VGG19 network, replacing the classification layers with global average pooling, dropout, and two fully connected layers with softmax activation. The model will be tested on the ISIC 2018 dataset containing over 10,000 images across 7 lesion classes, after preprocessing the images. The goal is to assist dermatologists in identifying lesion types.
Multiclass skin lesion classification with CNN and Transfer LearningYashIyengar
A multiclass classifier has been developed using transfer learning approach to detect seven different types of skin lesions. A modified VGG19 architecture is used for this purpose. The proposed model has achieved 78 % validation accuracy and 74 % test accuracy.
This document summarizes a research project that aims to analyze gamer behavior profiles by using clustering algorithms on game telemetry data. The project involves collecting game data, preprocessing the data, applying clustering algorithms like K-means, DBSCAN, K-medoids and Gaussian Mixture Models to segment players into groups based on their gameplay behavior. It will then evaluate the results using metrics like Davies-Bouldin Index, Silhouette Score and Sum of Squared Errors to validate the clusters. The goal is to help game developers better understand player tendencies to improve games and increase popularity.
Big Data Analysis of Second hand Car SalesYashIyengar
This document summarizes a student's research project on using big data technologies and linear regression to analyze second-hand vehicle sales data from Craig's List. The student cleaned and preprocessed the data, loaded it into databases, and performed analysis using Pig, Hive, and MapReduce. Regression analysis was then used to build a model to predict used car prices based on various attributes in the data.
1) Millennials spend more on eating out than previous generations due to convenience and higher incomes. They prefer quick service restaurants that provide fast, affordable meals.
2) McDonald's is a highly successful global fast food chain that has used strategies like customizing menus for local tastes, franchising, reducing operational costs, and extensive advertising to maintain its dominant market position worldwide.
3) By implementing technologies like kiosks and automation, McDonald's has increased efficiency while reducing costs and workforce, further boosting its profits.
Performance Comparison of HBase and CassandraYashIyengar
The document compares the performance of Hbase and Cassandra databases using YCSB (Yahoo! Cloud Serving Benchmark). It summarizes the key characteristics of each database, including that Hbase is master-based while Cassandra is masterless. The document then describes testing each database with YCSB Workloads A, B, and C at record counts of 100,000; 250,000; and 500,000 to compare their performance under different conditions.
The document is a project submission sheet for a student named Yash Balaji Iyengar. It includes details of the student's program of study, the module and lecturer, as well as information about the project such as the title "Statistics Continuous Assessment 2", word count of 1718, and due date of April 7, 2019. The student certifies that all work is their own or properly cited. Instructions are provided for project submission.
In depth Analysis of Suicide and its factorsYashIyengar
This document describes a project analyzing factors related to suicide rates. The project involves collecting data on suicide rates, unemployment, GDP, cigarette consumption, depression rates, and drug use from various sources. A data warehouse and business intelligence model is developed using a bottom-up approach to understand relationships between suicide rates and socioeconomic factors. Three key business questions are identified relating suicide rates to GDP, unemployment, and other health factors. Six data sources are described, including WHO, Our World in Data, Kaggle, OECD, and Wikipedia. The sources provide structured and unstructured data on suicide rates, depression, drug use, GDP, and unemployment for multiple countries over time.
Global Situational Awareness of A.I. and where its headedvikram sood
You can see the future first in San Francisco.
Over the past year, the talk of the town has shifted from $10 billion compute clusters to $100 billion clusters to trillion-dollar clusters. Every six months another zero is added to the boardroom plans. Behind the scenes, there’s a fierce scramble to secure every power contract still available for the rest of the decade, every voltage transformer that can possibly be procured. American big business is gearing up to pour trillions of dollars into a long-unseen mobilization of American industrial might. By the end of the decade, American electricity production will have grown tens of percent; from the shale fields of Pennsylvania to the solar farms of Nevada, hundreds of millions of GPUs will hum.
The AGI race has begun. We are building machines that can think and reason. By 2025/26, these machines will outpace college graduates. By the end of the decade, they will be smarter than you or I; we will have superintelligence, in the true sense of the word. Along the way, national security forces not seen in half a century will be un-leashed, and before long, The Project will be on. If we’re lucky, we’ll be in an all-out race with the CCP; if we’re unlucky, an all-out war.
Everyone is now talking about AI, but few have the faintest glimmer of what is about to hit them. Nvidia analysts still think 2024 might be close to the peak. Mainstream pundits are stuck on the wilful blindness of “it’s just predicting the next word”. They see only hype and business-as-usual; at most they entertain another internet-scale technological change.
Before long, the world will wake up. But right now, there are perhaps a few hundred people, most of them in San Francisco and the AI labs, that have situational awareness. Through whatever peculiar forces of fate, I have found myself amongst them. A few years ago, these people were derided as crazy—but they trusted the trendlines, which allowed them to correctly predict the AI advances of the past few years. Whether these people are also right about the next few years remains to be seen. But these are very smart people—the smartest people I have ever met—and they are the ones building this technology. Perhaps they will be an odd footnote in history, or perhaps they will go down in history like Szilard and Oppenheimer and Teller. If they are seeing the future even close to correctly, we are in for a wild ride.
Let me tell you what we see.
The Building Blocks of QuestDB, a Time Series Databasejavier ramirez
Talk Delivered at Valencia Codes Meetup 2024-06.
Traditionally, databases have treated timestamps just as another data type. However, when performing real-time analytics, timestamps should be first class citizens and we need rich time semantics to get the most out of our data. We also need to deal with ever growing datasets while keeping performant, which is as fun as it sounds.
It is no wonder time-series databases are now more popular than ever before. Join me in this session to learn about the internal architecture and building blocks of QuestDB, an open source time-series database designed for speed. We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
End-to-end pipeline agility - Berlin Buzzwords 2024Lars Albertsson
We describe how we achieve high change agility in data engineering by eliminating the fear of breaking downstream data pipelines through end-to-end pipeline testing, and by using schema metaprogramming to safely eliminate boilerplate involved in changes that affect whole pipelines.
A quick poll on agility in changing pipelines from end to end indicated a huge span in capabilities. For the question "How long time does it take for all downstream pipelines to be adapted to an upstream change," the median response was 6 months, but some respondents could do it in less than a day. When quantitative data engineering differences between the best and worst are measured, the span is often 100x-1000x, sometimes even more.
A long time ago, we suffered at Spotify from fear of changing pipelines due to not knowing what the impact might be downstream. We made plans for a technical solution to test pipelines end-to-end to mitigate that fear, but the effort failed for cultural reasons. We eventually solved this challenge, but in a different context. In this presentation we will describe how we test full pipelines effectively by manipulating workflow orchestration, which enables us to make changes in pipelines without fear of breaking downstream.
Making schema changes that affect many jobs also involves a lot of toil and boilerplate. Using schema-on-read mitigates some of it, but has drawbacks since it makes it more difficult to detect errors early. We will describe how we have rejected this tradeoff by applying schema metaprogramming, eliminating boilerplate but keeping the protection of static typing, thereby further improving agility to quickly modify data pipelines without fear.
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Aggregage
This webinar will explore cutting-edge, less familiar but powerful experimentation methodologies which address well-known limitations of standard A/B Testing. Designed for data and product leaders, this session aims to inspire the embrace of innovative approaches and provide insights into the frontiers of experimentation!
Predictably Improve Your B2B Tech Company's Performance by Leveraging DataKiwi Creative
Harness the power of AI-backed reports, benchmarking and data analysis to predict trends and detect anomalies in your marketing efforts.
Peter Caputa, CEO at Databox, reveals how you can discover the strategies and tools to increase your growth rate (and margins!).
From metrics to track to data habits to pick up, enhance your reporting for powerful insights to improve your B2B tech company's marketing.
- - -
This is the webinar recording from the June 2024 HubSpot User Group (HUG) for B2B Technology USA.
Watch the video recording at https://youtu.be/5vjwGfPN9lw
Sign up for future HUG events at https://events.hubspot.com/b2b-technology-usa/
Codeless Generative AI Pipelines
(GenAI with Milvus)
https://ml.dssconf.pl/user.html#!/lecture/DSSML24-041a/rate
Discover the potential of real-time streaming in the context of GenAI as we delve into the intricacies of Apache NiFi and its capabilities. Learn how this tool can significantly simplify the data engineering workflow for GenAI applications, allowing you to focus on the creative aspects rather than the technical complexities. I will guide you through practical examples and use cases, showing the impact of automation on prompt building. From data ingestion to transformation and delivery, witness how Apache NiFi streamlines the entire pipeline, ensuring a smooth and hassle-free experience.
Timothy Spann
https://www.youtube.com/@FLaNK-Stack
https://medium.com/@tspann
https://www.datainmotion.dev/
milvus, unstructured data, vector database, zilliz, cloud, vectors, python, deep learning, generative ai, genai, nifi, kafka, flink, streaming, iot, edge
State of Artificial intelligence Report 2023kuntobimo2016
Artificial intelligence (AI) is a multidisciplinary field of science and engineering whose goal is to create intelligent machines.
We believe that AI will be a force multiplier on technological progress in our increasingly digital, data-driven world. This is because everything around us today, ranging from culture to consumer products, is a product of intelligence.
The State of AI Report is now in its sixth year. Consider this report as a compilation of the most interesting things we’ve seen with a goal of triggering an informed conversation about the state of AI and its implication for the future.
We consider the following key dimensions in our report:
Research: Technology breakthroughs and their capabilities.
Industry: Areas of commercial application for AI and its business impact.
Politics: Regulation of AI, its economic implications and the evolving geopolitics of AI.
Safety: Identifying and mitigating catastrophic risks that highly-capable future AI systems could pose to us.
Predictions: What we believe will happen in the next 12 months and a 2022 performance review to keep us honest.
1. Diagnosing Pneumonia from Chest X-Rays Using Neural Networks
Tushar Dalvi Shantanu Deshpande Yash Iyangar Ashish Soni
x18134301 x18125514 x18124739 x18136664
Abstract—Disease diagnosis with radiology is a common prac-
tice in the medical domain but requires doctors to correctly
interpret the results from the images. Over the years due to
the increase in the number of patients and the low availability of
doctors, there was a need for a new method to identify and detect
disease in a patient. In recent years machine learning techniques
have been very effective in image-based disease diagnosis. In
this paper, a method has been proposed to diagnose Pneumonia
using transfer learning with VGG19 model. For pre-processing
online image, augmentation has been used to reduce the class
imbalance in the training data, With this methodology. For pre-
training the VGG19 model, Imagenet weights have been used
and three layers have been stripped off the model and 2 dense
layers have been added with Relu as an activation function in
one layer and softmax in the other layer. We were able to achieve
a recall of 0.96 and a precision of 0.87 with such a setting.
Index Terms—Pneumonia detection, CNN, Image Classification
I. INTRODUCTION
Radiology is one of the many branches of medicine that
basically focuses on making use of medical images for the
detection, diagnosis and characterization of the disease. One of
the main responsibilities of a Radiologist is to view a medical
image and based on the image producing a written report of
the observed findings [1]. Medical experts have made use of
this technique for around several years in order to visualize as
well as explore abnormalities and fractures in the body organs.
In recent years, due to the advancement in the technology
field, we have seen an up-gradation in the healthcare systems.
This has helped the doctors in better diagnosis accuracy and
reducing the fatality rate. The chest is the most important
part of the body as it contains the respiration organs which
are responsible for sustaining important life function of the
body. The count of people being diagnosed with a chest
disease globally is in millions. Some of the diseases that are
diagnosed using chest x-ray images are lung cancer, heart
diseases, Pneumonia, bronchitis, fractures etc.
Pneumonia is one such kind of chest disease wherein there
is an inflammation or infection in the lungs and it is most
commonly caused by a virus or a bacteria. Another cause due
to which a person can acquire Pneumonia is through inhaling
vomit or other foreign substances. If a person is diagnosed
with this disease, the lungs air sacs get filled with mucus,
pus, and few other liquids which do not allow the lungs to
function properly. This causes an obstruction in the path of
oxygen to reach the blood and the cells of the body in an
effective manner. Based on the data provided by the World
Health Organization (WHO), 2.4 million persons die every
year due to Pneumonia. [2]
One of the most tedious tasks for the radiologists is the
classification of X-ray abnormalities to detect and diagnose
chest diseases. As a result of this, several algorithms had
been proposed by researchers to perform this task accurately.
In these couple of decades, computer-aided diagnosis (CAD)
has been developed through various studies to interpret useful
information and assist doctors in getting a meaningful insight
about an X-ray. Several research works were carried out in
past by using artificial intelligence methodologies. Based on
the learnings from artificial intelligence methodologies, we can
also develop an image processing system that will use the X-
ray images and will detect the diseases in the primary stages
with improved accuracy. Some of the neural network method-
ologies that have been effectively used by the researchers in
replacing the traditional pattern recognition methods for the
diagnosis of diseases are probabilistic neural network (PNN),
multi-layer neural network (MLNN), generalized regression
neural network (GRNN), learning vector quantization (LVQ),
and radial basis function. These deep neural networks have
showcased increased accuracy in terms of image classification
and have thereby motivated the researchers to explore more
artificial intelligent techniques.
The convolutional Neural Network is another frequently
used deep learning architecture due to its power of extraction
of various different level features from the images. [3] By go-
ing through the relevant research studies, we hereby propose a
deep convolutional neural network (CNN) in order to improve
the accuracy of the diagnosis of chest disease, particularly
Pneumonia.
II. RESEARCH QUESTION
Can the Recall be improved for Chest X-ray Pneumonia
Detection using CNN based VGG19 model as compared to
the state of the art technique?
III. LITERATURE REVIEW
Owing to the difficulty in manually classifying the X-ray
images for a disease, various studies have evolved around the
use of Computer-aided diagnostics for obtaining meaningful
X-ray information and thereby assist physicians to gain a
quantitative insight of the disease. In this section, we will
review the pieces of literature that supports the current research
work.
In [4], the author examined three methods, Backpropagation
NN, Competitive NN, Convolutional NN. In BPNN, the errors
at output layers are propagated back to the network layer
whereas CpNN is a two-layer NN based on a supervised
algorithm. CNN has the strongest network as it can include
many hidden layers and performs convolutions and subsam-
pling in order to obtain low to a high level of features
from the input data. Also, it has three layers; convolutional
2. layer, subsampling/pooling layer, full connection layer. Results
showed that CNN took more time and number of learning
iterations however achieved the highest recognition rate and
was also able to get a better generalization power over BPNN
and CpNN.
In the above study, the output in terms of precision, com-
putation time was not as efficient due to no prior image
pre-processing performed. In [5], histogram equalization is
performed for image pre-processing and the images are divided
as well as normalized, further, cross-validation is conducted on
test data. The proposed model achieved an accuracy of 95.3%
on test data.
Further to the outstanding performance of CNN in [4],
another study [6] was conducted that evaluated performance
of three CNN architectures, Sequential CNN, Residual CNN
and Inception CNN. In the residual, loss of information is
prevented by passing information from earlier network layers
downstream thus solving the problem of representational bot-
tlenecks. Six residual blocks have been used and performance
measured based on following metrics, accuracy, precision,
AUC, specificity, recall. Non-linearity is introduced by using
a Rectified Linear Unit (ReLU) layer. The VGC16 model out-
performed all the other models. In [7], like [5], data has been
normalized by dividing all pixels by 255 and thus transformed
them to floating points. CNN model has compared with MLP
wherein each node except the input node is a neuron that
uses non-linear activation function. The technique involved
multiplying inputs by weight and then add them to bias and
dropout technique has been used to avoid overfitting and
reduce the training time. CNN outperformed MLP based on
evaluation metrics like cross-validation and confusion matrix.
A different approach can be seen in [8] where the network
layer utilizes the final layer as the convolutional implementa-
tion of a fully connected layer that allows a 40-fold speedup.
For dealing with imbalanced label distributions, a two-phase
training procedure has been proposed. Performance of three
different optimizers has been compared in [9], Adam opti-
mizer, momentum optimizer and stochastic gradient descent.
Filters are applied to the input image and the matrices thus
formed are known as feature maps. Filters of 5x5 size used
to extract features. In order to prevent overfitting, half of
the neurons are disabled using the dropout technique. Adams
optimizer outperformed the other two optimizers. Utilization
of ResNet-50 for embedding the features has been done in
[10] and CNN model thus built is able to get overall AUC of
0.816.
A different and novel approach is visible in [11] where
the author used Transfer Learning to train the model. The
advantage of TL is that a model can be trained even with
limited data. A dataset of 5232 images has been used and
the model achieved accuracy of 92.8% with sensitivity and
specificity of 93.2% and 90.1% respectively. In [12], the pre-
processing steps involved increasing the images and rotating
them by 40 degrees. By following a similar model building
process as [11], the author obtained the accuracy of 95.31%
and is better than [11].
Two different CNN techniques were used in [13] and [14],
the GoogleNet and AlexNet. During the pre-processing stage,
multiple rotations at 90, 180 and 270 degrees have been
performed and able to get AUC of 0.99. A rather different
approach than CNN can be seen in [15] where the author
used a technique called Multiple-instance learning (MIL). The
research overcomes some of the drawbacks like no proper
estimation of positive instances by reducing the number of
iterations. The slow configuration is avoided by avoiding the
reclassification and the AUC achieved is 0.86%. In [16],
the author tackled the issue of class imbalance using image
augmentation methods like rotating and flipping the images.
In the last dense layer, the sigmoid function is used to avoid
the problem of overfitting. Like [9], Adam optimizer has been
used to reduce losses and batch size is set to 64 whereas epoch
set to 100. The recall obtained is 96.7% and is better than state
of art techniques. A 121 layer convolutional neural network
has been trained on a large dataset of 100,000 X-Ray images
with 14 diseases in [17]. Weights initialized using pre-trained
weights from ImageNet and the model is trained using Adam.
An extension of the algorithm is undertaken to detect multiple
diseases and it is found that this model outperforms the state
of the art.
A deep neural network algorithm, Mask-RCNN is used
in [18]. This method uses lung opacity as a feature for
binary classification and Pneumonia detection. The final output
obtained is an ensemble of two models. The base network
is trained with COCO weights based on ResNet50 and
ResNet101 models to extract features from the images. An
improved result due to an ensemble approach is 0.21. In [19],
for extracting representative and discriminative features from
the Chest X-ray images for the effective classification into
various body parts, CNN’s were explored and the capability
of CNN in terms of capturing the image structure from feature
maps is portrayed and hand-engineered features were easily
outperformed by CNNs.
The usefulness of transfer learning [11] is further stated
in [20] where the authors used a pre-trained ChexNet model
with 121-dense layers and applied it in 4 blocks. The output
of pretrained model is transferred to the new model and Adam
optimizer is used. The model was trained on multiple blocks
and multiple layers and best validation accuracy obtained was
90.38% for the model with 6 layers. The use of demographic
features like age, weight and height is used in addition to
image features in [21] and the balanced dataset is used for
training. Five different algorithms have been used viz. In-
ceptionResNetV2, DenseNet121, InceptionV3, ResNet50 and
VGG19. The batch size used was 16 and Stochastic Gradient
Descent was chosen as optimizer function. VGG19 model
outperformed other models with AUC 0.9714 for training and
0.9213 for testing data.
Like [20] and [11], transfer learning is used in [22], how-
ever, in this study due to the high resolution of CXR images,
an extra convolution layer is added. Proposed system showed
enhanced screening performance. A multi-CNN model is used
in [23] which consists of three CNN components. The output
3. of each CNN component is a probability which classifies
the image as normal or abnormal. The outcome of all three
components is calculated by using the association fusion rule.
The fusion rule consists of eight cases which lead to the final
classification of the image. The model has been trained using
the association fusion rule for classification of the images.
Their model has achieved an accuracy of 96%.
IV. METHODOLOGY
From above Literature Review, we can say that in the medi-
cal domain it is hard for a layman to identify a disease by just
looking at the images of the body part. To identify disease from
x-rays, MRI Scans well-trained doctors or properly trained
computer-based system are required because a layman doesn’t
possess the knowledge of the disease like the doctor. Even
though we are building a model to detect Pneumonia we would
want our model to detect different diseases with an adequate
amount of training examples and model weights. So we will be
using the CRISP-DM methodology [24]. It stands for Cross
Industry Standard Process for Data Mining. It has six steps
out of which we will be using Business Understanding, Data
Understanding, Data Preparation, Modeling and Evaluation.
We won’t be deploying our model for real-world use as it is
a model developed for this project and further methodologies
need to be tested.
1) Setup: In this study all the data is image formatted, so
we need to process the images to identify the patterns and
based on that patterns need to classify the different classes.
To perform all these tasks, we need a proper hardware system
which can handle image processing task without system fail-
ure. Also, the requirement of proper GPU is must to train the
model within minimum time. Python Programming Language
has been used for this research, but due to lack of hardware
processing power, normal laptops or systems might be not able
to handle the model building process smoothly. To overcome
this issue, for this study we are using cloud service-based
platform named Google Collaboratory provided by Google. It
is a free Jupyter notebook environment that requires no prior
setup and runs entirely on cloud [25]. In this environment 12
GB of ram and 50 GB of storage space is provided initially
as freemium service, also Google provides Free Tesla K80
Graphical processor of about 12 GB for fast computing. To
perform this study data is uploaded on the Google drive
for maximum time availability. And all the task which are
mentioned in the next section is performed on Google Colab
notebook.
V. DATA UNDERSTANDING AND PRE-PROCESSING
While training the neural network, we face a common
problem that not enough data is available to train the model
to maximize the capability of the neural network [26]. In
the medical domain data sets having more class imbalance
as compared to a dataset of other domains. The data which
we have gathered for our study is collected from Kaggle and is
already divided into training, validation and testing sets. So we
don’t have to create the partition for the data. As mentioned
above the data set is highly imbalanced, training data contains
a total of 5216 images out of which 1341 are images of normal
lungs and 3875 are images of Pneumonia infected lungs, which
means for every image of normal lungs we have 3 images of
Pneumonia infected lungs.
Fig. 1. Normal Vs Pneumonia Cases.
In such cases, the evaluation cannot rely only on metrics
like Accuracy. There are many techniques available to reduce
the recurring problem of class imbalance in the data which
includes random undersampling, random oversampling and
Synthetic Minority Over-sampling Technique(SMOTE). But
there is a flaw in using these methods as there is a high
probability of information loss as we have to detect Pneumonia
and ignoring any Pneumonia case from the dataset will lead
to loss of information for training purpose. Hence in data
pre-processing Data augmentation is a technique used to deal
with the class imbalance in the data without the need to
collect new data. In this method the images of the class with
fewer numbers in the data are duplicated by flipping, rotating,
padding the available images of that class. There are two types
of augmentation techniques one is offline and another is online,
offline create a new dataset in the mentioned directory and it
is time-consuming. But in Online Augmentation with the help
of Keras, we can create a batch of images and that image gets
duplicated and used as input, also online augmentation doesnt
need that much space and it consumes less computational
power. For code reproducibility we have set the initial seed to
100 with the function random.seet(100). In augmentation, we
have performed rescaling, rotation, width shift, height shift,
and horizontal flip using the ImageDataGenerator function
from the Keras package [27]. In Rescaling, image is divided by
255 and converted into floating value. Also, images are rotated
by 20 degrees. Width and height is shifted by 0.2 fractions
of the total width, and with the horizontal flip parameter
set to true, the images are flipped horizontally. The images
in the dataset had varying size in height and width, which
was not proper to provide as an input to the model, hence
all images were downsized to 128 x 128 pixel by using
the flow from directory function and also the shuffle flag
parameter was set to true, so while training phase model
4. should pick the random image form the data. Providing all
images at the same time will take a lot of computational power
to analyze images hence we set the batch size to 16 images per
batch and set the class mode as categorical. With the help of
the above steps, we were able to reduce the class imbalance in
the dataset. For testing and validation data set only rescaling
was performed during augmentation and also batch size was
set to 16 images.
VI. MODELING
As this study aims to diagnose Pneumonia from Chest
X-Ray images, a supervised learning technique is suitable
for further study. Previous studies like [16] and [6] used
Convolutional Neural Network and got promising results.
A study conducted by [21] is used as our base paper to
continue our research. As mention earlier in section IV-1
we are using Python programming language to complete this
study, the main reason to use python is, it has large amount
of libraries like NumPy, Matplotlib, and TensorFlow used to
carry out various task, and with cloud environment like Google
Collaboratory, and Google drive helps to keep personal system
available for another task. To build a model it requires suitable
computational power, with reference to [21] study, we are
using VGG19 model for further process. Here VGG stands
for the Visual Geometry group from the Oxford University
who developed the VGG16 model for an image recognition
competition and won it. The number ’16’ and ’19’ stands for
the number of layers in the CNN model. This CNN model
was developed and trained on multiple images for image
classification. It consists of 3x3 layers of convolution stacked
onto each other with the activation function as Rectifier Linear
Unit (ReLU), to reduce the volumetric size max-pooling layers
are used and are followed by fully connected layers again
with ReLU as their activation function. It can classify images
from almost one thousand different categories and can take
an input image with dimensions 224x224 pixels. The model’s
application in this project is explained in the next section.
1) Model Building: Building a model or selecting a model
is one of the crucial tasks in data mining techniques. Choosing
the right model not only enhance the result but also reduce the
time required to train the model. In this study, we use transfer
learning to build out CNN model, in transfer learning a pre-
built model store the knowledge which is gained while solving
one problem and use and apply this knowledge on another
related program [28]. In our study, we use MobileNetV2
as a base model for transfer learning. MobileNets improve
the performance of mobile models on multiple tasks. We
have created a custom model with the help of VGG19.
As mentioned above it is a renowned Convolutional Neural
Network Architecture for object recognition task developed
and trained by Oxford’s renowned Visual Geometry Group
[29]. In this study VGG19() function from the Keras package
imports VGG19 model. Here we use a novel approach by
excluding the top 3 layers of the model, add two dense layers
and we provide the input images with the dimensions of 128
x 128 x 3. Also, pretraining is performed on the Imagenet
weights by allocating Imagenet attribute to the weights. Lastly,
we set False flag for trainable attribute so while training the
model no new weight will be updated. After importing VGG19
model we call the GlobalAveragePooling2D() function which
performs global average pooling and calculates the average
map of features from the previous CNN layer to reduce the
effect of overfitting. Two dense layers are added to the model
with activation function as ReLU and softmax respectively.
The reason behind using ReLU activation function is, it is
non-linear and doesn’t produce back propagation errors like
the Sigmoid function. It also speeds the model building process
for a large Neural Networks. Softmax activation function helps
to get output in 0 to 1 range, therefore the output from softmax
falls under the probability distribution function and helps in
classifying multiple classes as it outputs the probability for
each class between 0 and 1. Also, while using a dense layer
512 units have been provided for the output space [30]. The
Dropout rate is set to 0.7 to reduce the effect of overfitting.
2) Model Training: In the previous section, we build a
custom model by adding 2 dense layers in the imported
VGG19 model. But the model is configured for better per-
formance and to minimize the loss function with the use
of the compile function. While configuring compile function
we need to use optimize our model to reduce training loss.
In this study, Adam optimizer has been used because it
computes the individual learning rates for different parameter.
Secondly, for the loss function, categorical crossentropy is
used. categorical crossentropy is a loss function that is used
for categorization of single label that means the single example
from our dataset should only belong to a single class. Hence
our model will not classify the both Normal and Pneumonia
class for a single image and for the metrics we have used
accuracy for evaluation of the model on the training data.
The model training phase is the most time-consuming phase
in the entire data mining process. This phase takes time to
perform depending on the size of the dataset and the number
of iteration. To configure this process fit generator() function
from Keras is used to fit the entire training data into the
memory, by setting the Epochs to 20 and verbose to 1.
The main task of the fit generator function is to discard the
previous data which was already used for training purpose
and create a new one for the next process. This task repeats
the same process with the number of epochs that have been
provided. Model training takes a lot of computing cost and
its a heavy process when it comes to performing image-based
data classification. Hence we run this process on Google collab
with 12 GB of Ram and with Tesla K80 graphical processor
with 12 GB graphical memory [31]. Even with this hardware
configuration, it took around 9 hrs to complete 20 Epochs.
While performing this training we store the log of each epoch
to evaluate the performance of the model which is explained
in the next section.
VII. EVALUATION
In the previous section, we created our model and trained it
on the training dataset, during the training period, logs were
5. recorded to check the model performance. In this section, we
will check the model performance from those logs. Evaluation
technique helps to compare the generated result and expected
result. Initially, with the help of logs, we will evaluate the
accuracy and loss while training the model.
As we can see in the figure 4 loss on the Validation data
is varying in all the epochs and gradually increasing, at the
very first epoch loss reduces from 0.5 to 0.4 approximately.
Till it reaches the 19th epoch it increases to 0.8 with lots of
variations. In the last epoch loss on the validation data falls to
0.64. The loss of the training data constantly decreases. This
contrast in the loss plot obtained by running the model with
20 epochs on both training and validation data suggests that
the model is not overfitting the training data.
Fig. 2. Loss Plot
Accuracy plot tells us about how many images of both
classes are being correctly predicted by the model. It increases
for the training data gradually over the period of 20 epochs to
approximately 0.92, but in case of validation dataset accuracy
sharply increases from 0.81 to 0.875 and sharply decreases
to .81 again and remains constant till 12 epochs and then
again sharply varies till the last epoch. Even over here we
can observe that both the training accuracy and validation
accuracy do not flow in the same direction that is they do not
increase constantly and together, this means that the model is
not overfitting the training data’s information.
The model’s performance is evaluated over the test data
using the evaluate generator() function from the Keras pack-
age. This function checks accuracy on the testing dataset,
in which maximum generator queue is set to 10 and set
use multiprocessing to true. As in result, we can see that
loss on the test dataset is around 33% and accuracy is
approximately 89%. But evaluate generator function initially
Fig. 3. Accuracy Plot
predict the output using training data and then evaluate the
performance by comparing against the test dataset which we
called as accuracy measure.
As mentioned in section V our dataset is from Health care
domain and to calculate the proper prediction we need to
consider recall and precision rather than accuracy. To calculate
the recall and precision we need to provide the dataset which is
not yet introduced to our model. For that, we need to perform
prediction on the testing dataset. But before that, we need to
identify the classes from test dataset to compare the result with
our prediction. With the use of load model() function from the
Keras package, the saved model is loaded. And with the help of
CV2 library, all images from testing dataset have been resized
to 128 x 128 pixel to reduce the computational cost. Also
with the help of CV2 library, each image has been labelled
with their respective class which is 1 for Pneumonia infected
lungs image and 0 for normal lungs image. This step helps
to compare the actual image’s class with predicted image’s
classes. To initiate the prediction process predict() function
has been used. The model’s performance on testing data is
tested and a batch size of 16 and with the argmax function
model returns the indices of the maximum values along the
rows as it is set to 1. To plot the result confusion matrix is used
with matplotlib library and precision and recall are calculated
from the output which is discussed in the Result section.
VIII. RESULTS
The performance of our classification model is described
through a table known as the confusion matrix, as given below:
The basic terms to be interpreted from our confusion matrix
are:
True Positives (TP) : The number cases where the model
predicted yes and the person do have Pneumonia, i.e. 376.
6. Fig. 4. Confusion Matrix
True Negatives (TN) : the number of cases where the model
predicted no and the person does not have Pneumonia. i.e.
177. False positives (FP) : Model predicted yes, but actually
they do not have Pneumonia. i.e. 57. False Negatives (FN) :
Model predicted no, but actually they do have Pneumonia. i.e.
34.
How often our model is correct when it predicts yes, is
given by precision and it is calculated as:
Precision = TP / (TP + FP) i.e. 0.87
Also, for our study, we will be focusing on Recall as a
metric and expect it to be high. Recall would be a good
measure of fit for our case as our goal is to correctly classify
all the Pneumonia cases to save the patient’s life. How many
images the model has classified correctly out of all the positive
classes is given by Recall and it should be as high as possible.
Recall “
TP
pTP ` FNq
As for our case, achieved Recall is 0.96. A high recall is
expected but whenever there is a high recall, the precision
reduces. This means an increase in the number of false
positive. This means the model is wrongly classifying normal
lung images as Pneumonia infected lung images. Thus there
is a trade-off between recall and precision which needs to be
maintained and depends on the domain and industry standards.
IX. CONCLUSION AND FUTURE WORK
In this study, a methodology has been proposed to diagnose
Pneumonia from chest x-ray images. The dataset selected
had a high number of Pneumonia images and comparative
less number of normal images. The imbalance ratio was
3 is to 1, so the image augmentation procedure was used
to reduce the class imbalance. Due to time constraints and
limited availability of resources other pre-processing methods
like pixel brightness transformation and image segmentation
could not be performed we suggest these techniques to be
implemented in the future. The VGG19 model was used and
three layers of the model were removed and two dense layers
were added. Building a CNN model from scratch requires high
computational capacity and as there was a shortage of time,
we would suggest building a CNN model from scratch and
fine-tuning the parameters for the future works. We were able
to achieve a decent output with 0.96 recall, with a different
approach, the result can be improved.
REFERENCES
1 Kohli, M., Prevedello, L. M., Filice, R. W., and Geis, J. R., “Implementing
machine learning in radiology practice and research,” American Journal
of Roentgenology, vol. 208, no. 4, pp. 754–760, 2017.
2 Er, O., Yumusak, N., and Temurtas, F., “Chest diseases diagnosis
using artificial neural networks,” Expert Systems with Applications,
vol. 37, no. 12, pp. 7648–7655, 2010. [Online]. Available: http:
//dx.doi.org/10.1016/j.eswa.2010.04.078
3 Bengio, Y., Lamblin, P., Popovici, D., and Larochelle, H., “Greedy
layer-wise training of deep networks,” in Advances in Neural Information
Processing Systems 19, Sch¨olkopf, B., Platt, J. C., and Hoffman, T., Eds.
MIT Press, 2007, pp. 153–160. [Online]. Available: http://papers.nips.cc/
paper/3048-greedy-layer-wise-training-of-deep-networks.pdf
4 Tom`e, D., Monti, F., Baroffio, L., Bondi, L., Tagliasacchi, M.,
and Tubaro, S., “Deep convolutional neural networks for pedestrian
detection,” vol. 2018, 2015. [Online]. Available: http://arxiv.org/abs/1510.
03608tz%u0Ahttp://dx.doi.org/10.1016/j.image.2016.05.007
5 Saraiva, A., Ferreira, N., Lopes de Sousa, L., Costa, N., Sousa, J.,
Santos, D., Valente, A., and Soares, S., “Classication of Images of
Childhood Pneumonia using Convolutional Neural Networks,” pp. 112–
119, 2019.
6 Rajaraman, S., Candemir, S., Kim, I., Thoma, G., and Antani, S., “Visual-
ization and Interpretation of Convolutional Neural Network Predictions in
Detecting Pneumonia in Pediatric Chest Radiographs,” Applied Sciences,
vol. 8, no. 10, p. 1715, 2018.
7 Hand, D. J., “Measuring classifier performance: A coherent alternative to
the area under the ROC curve,” Machine Learning, vol. 77, no. 1, pp.
103–123, 2009.
8 Havaei, M., Davy, A., Warde-Farley, D., Biard, A., Courville, A.,
Bengio, Y., Pal, C., Jodoin, P. M., and Larochelle, H., “Brain tumor
segmentation with Deep Neural Networks,” Medical Image Analysis,
vol. 35, pp. 18–31, 2017. [Online]. Available: http://dx.doi.org/10.1016/j.
media.2016.05.004
9 Hooda, R., Sofat, S., Kaur, S., Mittal, A., and Meriaudeau, F., “Deep-
learning: A potential method for tuberculosis detection using chest ra-
diography,” Proceedings of the 2017 IEEE International Conference on
Signal and Image Processing Applications, ICSIPA 2017, pp. 497–502,
2017.
10 Guan, Q. and Huang, Y., “Multi-label chest X-ray image classification via
category-wise residual attention learning,” Pattern Recognition Letters,
no. xxxx, 2018. [Online]. Available: https://doi.org/10.1016/j.patrec.2018.
10.027
11 Kermany, D. S., Goldbaum, M., Cai, W., Valentim, C. C., Liang, H.,
Baxter, S. L., McKeown, A., Yang, G., Wu, X., Yan, F., Dong, J.,
Prasadha, M. K., Pei, J., Ting, M., Zhu, J., Li, C., Hewett, S., Dong, J.,
Ziyar, I., Shi, A., Zhang, R., Zheng, L., Hou, R., Shi, W., Fu, X., Duan, Y.,
Huu, V. A., Wen, C., Zhang, E. D., Zhang, C. L., Li, O., Wang, X.,
Singer, M. A., Sun, X., Xu, J., Tafreshi, A., Lewis, M. A., Xia, H., and
Zhang, K., “Identifying Medical Diagnoses and Treatable Diseases by
Image-Based Deep Learning,” Cell, vol. 172, no. 5, pp. 1122–1131.e9,
2018. [Online]. Available: https://doi.org/10.1016/j.cell.2018.02.010
12 Stephen, O., Sain, M., Maduh, U. J., and Jeong, D. U., “An Efficient Deep
Learning Approach to Pneumonia Classification in Healthcare,” Journal
of Healthcare Engineering, vol. 2019, 2019.
13 Lakhani, P. and Sundaram, B., “Radiol.2017162326,” vol. 000, no. 0,
2017.
7. 14 Gu, X., Pan, L., Liang, H., and Yang, R., “Classification of Bacterial and
Viral Childhood Pneumonia Using Deep Learning in Chest Radiography,”
pp. 88–93, 2018.
15 Melendez, J., Van Ginneken, B., Maduskar, P., Philipsen, R. H., Re-
ither, K., Breuninger, M., Adetifa, I. M., Maane, R., Ayles, H., and
S´anchez, C. I., “A novel multiple-instance learning-based approach to
computer-aided detection of tuberculosis on chest X-rays,” IEEE Trans-
actions on Medical Imaging, vol. 34, no. 1, pp. 179–192, 2015.
16 Liang, G. and Zheng, L., “A transfer learning method with deep
residual network for pediatric pneumonia diagnosis,” Computer Methods
and Programs in Biomedicine, no. xxxx, 2019. [Online]. Available:
https://doi.org/10.1016/j.cmpb.2019.06.023
17 Rajpurkar, P., Irvin, J., Zhu, K., Yang, B., Mehta, H., Duan, T.,
Ding, D., Bagul, A., Langlotz, C., Shpanskaya, K., Lungren, M. P.,
and Ng, A. Y., “CheXNet: Radiologist-Level Pneumonia Detection on
Chest X-Rays with Deep Learning,” pp. 3–9, 2017. [Online]. Available:
http://arxiv.org/abs/1711.05225
18 Jaiswal, A. K., Tiwari, P., Kumar, S., Gupta, D., Khanna, A., and
Rodrigues, J. J., “Identifying pneumonia in chest X-rays: A deep learning
approach,” Measurement: Journal of the International Measurement
Confederation, vol. 145, pp. 511–518, 2019. [Online]. Available:
https://doi.org/10.1016/j.measurement.2019.05.076
19 Srinivas, M., Roy, D., and Mohan, C. K., “Discriminative feature ex-
traction from X-ray images using deep convolutional neural networks,”
ICASSP, IEEE International Conference on Acoustics, Speech and Signal
Processing - Proceedings, vol. 2016-May, pp. 917–921, 2016.
20 Pardamean, B., Cenggoro, T. W., Rahutomo, R., Budiarto, A.,
and Karuppiah, E. K., “Transfer learning from chest x-ray pre-
trained convolutional neural network for learning mammogram
data,” Procedia Computer Science, vol. 135, no. Breast cancer:
basic and clinical research 9 2015, pp. 400–407, 2018.
[Online]. Available: https://app.dimensions.ai/details/publication/pub.
1106388589andhttps://doi.org/10.1016/j.procs.2018.08.190
21 Heo, S. J., Kim, Y., Yun, S., Lim, S. S., Kim, J., Nam, C. M., Park, E. C.,
Jung, I., and Yoon, J. H., “Deep learning algorithms with demographic
information help to detect tuberculosis in chest radiographs in annual
workers’ health examination data,” International Journal of Environmental
Research and Public Health, vol. 16, no. 2, 2019.
22 Hwang, S., Kim, H.-E., Jeong, J., and Kim, H.-J., “A novel approach for
tuberculosis screening based on deep convolutional neural networks,” 03
2016, p. 97852W.
23 Kieu, P. N., Tran, H. S., Le, T. H., Le, T., and Nguyen, T. T. T.,
“Applying multi-cnns model for detecting abnormal problem on chest
x-ray images,” 2018 10th International Conference on Knowledge and
Systems Engineering (KSE), pp. 300–305, 2018.
24 Wirth, R. and Hipp, J., “CRISP-DM : Towards a Standard Process Model
for Data Mining,” Proceedings of the Fourth International Conference on
the Practical Application of Knowledge Discovery and Data Mining, no.
24959, pp. 29–39, 1995.
25 Carneiro, T., Da Nobrega, R. V. M., Nepomuceno, T., Bian, G. B., De
Albuquerque, V. H. C., and Filho, P. P. R., “Performance Analysis of
Google Colaboratory as a Tool for Accelerating Deep Learning Applica-
tions,” IEEE Access, vol. 6, pp. 61 677–61 685, 2018.
26 Lemley, J., Bazrafkan, S., and Corcoran, P., “Smart Augmentation Learn-
ing an Optimal Data Augmentation Strategy,” IEEE Access, vol. 5, pp.
5858–5869, 2017.
27 Fujita, K., Kobayashi, M., and Nagao, T., “Data Augmentation using Evo-
lutionary Image Processing,” 2018 International Conference on Digital
Image Computing: Techniques and Applications, DICTA 2018, pp. 1–6,
2019.
28 Ling Shao, Fan Zhu, and Xuelong Li, “Transfer Learning for Visual
Categorization: A Survey,” IEEE Transactions on Neural Networks and
Learning Systems, vol. 26, no. 5, pp. 1019–1034, 2014.
29 Shaha, M. and Pawar, M., “Transfer Learning for Image Classification,”
Proceedings of the 2nd International Conference on Electronics, Commu-
nication and Aerospace Technology, ICECA 2018, no. Iceca, pp. 656–660,
2018.
30 Lau, M. M. and Lim, K. H., “Review of adaptive activation function
in deep neural network,” 2018 IEEE EMBS Conference on Biomedical
Engineering and Sciences, IECBES 2018 - Proceedings, pp. 686–690,
2019.
31 Garg, A., Gupta, D., Saxena, S., and Sahadev, P. P., “Validation of Random
Dataset Using an Efficient CNN Model Trained on MNIST Handwritten
Dataset,” 2019 6th International Conference on Signal Processing and
Integrated Networks, SPIN 2019, pp. 602–606, 2019.