This document summarizes Mansi Chowkkar's MSc research project on detecting breast cancer from histopathological images using deep learning and transfer learning. The research aims to classify images as malignant or benign with high accuracy and efficiency. It implements CNN and DenseNet-121 models, with the latter using transfer learning with pre-trained ImageNet weights. The research achieved 90.9% test accuracy with CNN and 88.03% accuracy with transfer learning. Related work discusses deep learning applications in healthcare, image processing techniques, prior research on DenseNet, and the use of transfer learning for medical image classification with limited data.
Pneumonia Classification using Transfer LearningTushar Dalvi
Pneumonia can be life-threatening for people with weak immune systems, in which the alveoli filled with fluid that makes it hard to pass oxygen throughout the bloodstream. Detecting pneumonia is from a chest X-ray is not only expansive but also time-consuming for normal people. Throughout this research introduced a machine learning technique to classify pneumonia from Chest X-ray Images. Most of the medical datasets having class imbalance issues in the dataset. The Data augmentation technique used to reduce the class imbalance from the dataset, Horizontal Flip, width shift and height shift techniques used to complete the augmentation technique. Used VGG19 as a base architecture and ImageNet weights added for the transfer learning approach, also Removing initial layers and adding
some more dense layers helped to discover new possibilities. After testing the proposed model on testing data, we are able to achieve 98% recall and 82% of precision. As compare with state of the art technique, the proposed method able to achieve high
recall but that compromises with Precision.
Early detection of breast cancer using mammography images and software engine...TELKOMNIKA JOURNAL
The breast cancer has affected a wide region of women as a particular case. Therefore, different researchers have focused on the early detection of this disease to overcome it in efficient way. In this paper, an early breast cancer detection system has been proposed based on mammography images. The proposed system adopts deep-learning technique to increase the accuracy of detection. The convolutional neural network (CNN) model is considered for preparing the datasets of training and test. It is important to note that the software engineering process model has been adopted in constructing the proposed algorithm. This is to increase the reliably, flexibility and extendibility of the system. The user interfaces of the system are designed as a website used at country side general purpose (GP) health centers for early detection to the disease under lacking in specialist medical staff. The obtained results show the efficiency of the proposed system in terms of accuracy up to more than 90% and decrease the efforts of medical staff as well as helping the patients. As a conclusion, the proposed system can help patients by early detecting the breast cancer at far places from hospital and referring them to nearest specialist center.
A novel framework for efficient identification of brain cancer region from br...IJECEIAES
Diagnosis of brain cancer using existing imaging techniques, e.g., Magnetic Resonance Imaging (MRI) is shrouded with various degrees of challenges. At present, there are very few significant research models focusing on introducing some novel and unique solutions towards such problems of detection. Moreover, existing techniques are found to have lesser accuracy as compared to other detection schemes. Therefore, the proposed paper presents a framework that introduces a series of simple and computationally cost-effective techniques that have assisted in leveraging the accuracy level to a very higher degree. The proposed framework takes the input image and subjects it to non-conventional segmentation mechanism followed by optimizing the performance using directed acyclic graph, Bayesian Network, and neural network. The study outcome of the proposed system shows the significantly higher degree of accuracy in detection performance as compared to frequently existing approaches.
Pneumonia Classification using Transfer LearningTushar Dalvi
Pneumonia can be life-threatening for people with weak immune systems, in which the alveoli filled with fluid that makes it hard to pass oxygen throughout the bloodstream. Detecting pneumonia is from a chest X-ray is not only expansive but also time-consuming for normal people. Throughout this research introduced a machine learning technique to classify pneumonia from Chest X-ray Images. Most of the medical datasets having class imbalance issues in the dataset. The Data augmentation technique used to reduce the class imbalance from the dataset, Horizontal Flip, width shift and height shift techniques used to complete the augmentation technique. Used VGG19 as a base architecture and ImageNet weights added for the transfer learning approach, also Removing initial layers and adding
some more dense layers helped to discover new possibilities. After testing the proposed model on testing data, we are able to achieve 98% recall and 82% of precision. As compare with state of the art technique, the proposed method able to achieve high
recall but that compromises with Precision.
Early detection of breast cancer using mammography images and software engine...TELKOMNIKA JOURNAL
The breast cancer has affected a wide region of women as a particular case. Therefore, different researchers have focused on the early detection of this disease to overcome it in efficient way. In this paper, an early breast cancer detection system has been proposed based on mammography images. The proposed system adopts deep-learning technique to increase the accuracy of detection. The convolutional neural network (CNN) model is considered for preparing the datasets of training and test. It is important to note that the software engineering process model has been adopted in constructing the proposed algorithm. This is to increase the reliably, flexibility and extendibility of the system. The user interfaces of the system are designed as a website used at country side general purpose (GP) health centers for early detection to the disease under lacking in specialist medical staff. The obtained results show the efficiency of the proposed system in terms of accuracy up to more than 90% and decrease the efforts of medical staff as well as helping the patients. As a conclusion, the proposed system can help patients by early detecting the breast cancer at far places from hospital and referring them to nearest specialist center.
A novel framework for efficient identification of brain cancer region from br...IJECEIAES
Diagnosis of brain cancer using existing imaging techniques, e.g., Magnetic Resonance Imaging (MRI) is shrouded with various degrees of challenges. At present, there are very few significant research models focusing on introducing some novel and unique solutions towards such problems of detection. Moreover, existing techniques are found to have lesser accuracy as compared to other detection schemes. Therefore, the proposed paper presents a framework that introduces a series of simple and computationally cost-effective techniques that have assisted in leveraging the accuracy level to a very higher degree. The proposed framework takes the input image and subjects it to non-conventional segmentation mechanism followed by optimizing the performance using directed acyclic graph, Bayesian Network, and neural network. The study outcome of the proposed system shows the significantly higher degree of accuracy in detection performance as compared to frequently existing approaches.
Developed Project with 3 more colleagues for Pneumonia Detection from Chest X-ray images using Convolutional Neural Network. Used confusion matrix, Recall, Precision for check the model performance on testing Data
BREAST CANCER DIAGNOSIS USING MACHINE LEARNING ALGORITHMS –A SURVEYijdpsjournal
Breast cancer has become a common factor now-a-days. Despite the fact, not all general hospitals
have the facilities to diagnose breast cancer through mammograms. Waiting for diagnosing a breast
cancer for a long time may increase the possibility of the cancer spreading. Therefore a computerized
breast cancer diagnosis has been developed to reduce the time taken to diagnose the breast cancer and
reduce the death rate. This paper summarizes the survey on breast cancer diagnosis using various machine
learning algorithms and methods, which are used to improve the accuracy of predicting cancer. This survey
can also help us to know about number of papers that are implemented to diagnose the breast cancer.
SVM &GA-CLUSTERING BASED FEATURE SELECTION APPROACH FOR BREAST CANCER DETECTIONijscai
Mortality leading among women in developed countries is breast cancer. Breast cancer is women's second most prominent cause of cancer mortality worldwide. In recent decades, women's high prevalence of breast cancer has risen dramatically. This paper discussed several data analysis methods used to detect breast cancer early. Breast cancer diagnosis distinguishes benign and malignant breast lumps. Using data processing tools, we tackled this disease analysis. Data mining is an important step of library discovery where intelligent methods are used to detect patterns. Several clinical breast cancer studies were conducted using soft computing and machine learning techniques. Sometimes their algorithms are easier, easier, or more comprehensive than others. This research is focused on genetic programming and machine learning algorithms to reliably identify benign and malignant breast cancer. This study aimed to optimise the testing algorithm. We used genetic programming methods to choose classification machines' best features and parameter values. Data mining is an important step of library discovery where intelligent methods are used to detect patterns. We are analysing data accessible from the U.C.I. deep-learning data set in Wisconsin. In this experiment, we equate four Weka clustering strategies with genetic clustering. A comparison of results reveals that sequential minimal optimization (S.M.O.) is better than I.B.K. and B.F. Tree processes, i.e. 97.71%.
An Analysis of The Methods Employed for Breast Cancer Diagnosis IJORCS
Breast cancer research over the last decade has been tremendous. The ground breaking innovations and novel methods help in the early detection, in setting the stages of the therapy and in assessing the response of the patient to the treatment. The prediction of the recurrent cancer is also crucial for the survival of the patient. This paper studies various techniques used for the diagnosis of breast cancer. Different methods are explored for their merits and de-merits for the diagnosis of breast lesion. Some of the methods are yet unproven but the studies look very encouraging. It was found that the recent use of the combination of Artificial Neural Networks in most of the instances gives accurate results for the diagnosis of breast cancer and their use can also be extended to other diseases.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Comparison of Image Segmentation Algorithms for Brain Tumor DetectionIJMTST Journal
This paper deals with the implementation of Simple Algorithms for detection of size and shape of tumor in brain using MRI images. Generally, CT scan or MRI that is directed into intracranial cavity produces a complete image of brain. This image is visually examined by the physician for detection & diagnosis of brain tumor. However this method of detection resists the accurate determination of stage & size of tumor. To avoid that, this project uses computer aided method for segmentation (detection) of brain tumor by applying Fuzzy C-Means, K-Means, Gaussian Kernel and Pillar K-means algorithms. This segmentation process includes a new mechanism for clustering the elements of high-resolution images in order to improve precision and reduce computation time. The system applies FCM, Gaussian kernel and K-means clustering to the image later optimized by Pillar Algorithm. It designates the initial centroids’ positions by calculating the Euclidian distance metric between each data point and all previous centroids. Then it selects data points which have the maximum distance as new initial centroids. This algorithm distributes all initial centroids according to the maximum accumulated distance metric. In addition, it also reduces the time for analysis. At the end of the process the tumor is extracted from the MRI image and its exact position and the shape is also determined. This paper evaluates the proposed approach for Brain tumor detection by comparing with K-means, Fuzzy C means, Gaussian Kernel and manually segmented algorithms. The experimental results clarify the effectiveness of proposed approach to improve the segmentation quality in aspects of precision and computational time.
COVID-19 detection from scarce chest X-Ray image data using few-shot deep lea...Shruti Jadon
In the current COVID-19 pandemic situation, there is an urgent need to screen infected patients quickly and accurately. Using deep learning models trained on chest X-ray images can become an efficient method for screening COVID-19 patients in these situations. Deep learning approaches are already widely used in the medical community. However, they require a large amount of data to be accurate.
SVM &GA-CLUSTERING BASED FEATURE SELECTION APPROACH FOR BREAST CANCER DETECTIONijscai
Mortality leading among women in developed countries is breast cancer. Breast cancer is women's second most prominent cause of cancer mortality worldwide. In recent decades, women's high prevalence of breast cancer has risen dramatically. This paper discussed several data analysis methods used to detect breast cancer early. Breast cancer diagnosis distinguishes benign and malignant breast lumps. Using data processing tools, we tackled this disease analysis. Data mining is an important step of library discovery where intelligent methods are used to detect patterns. Several clinical breast cancer studies were conducted using soft computing and machine learning techniques. Sometimes their algorithms are easier, easier, or more comprehensive than others. This research is focused on genetic programming and machine
learning algorithms to reliably identify benign and malignant breast cancer. This study aimed to optimise the testing algorithm. We used genetic programming methods to choose classification machines' best features and parameter values. Data mining is an important step of library discovery where intelligent methods are used to detect patterns. We are analysing data accessible from the U.C.I. deep-learning data
set in Wisconsin. In this experiment, we equate four Weka clustering strategies with genetic clustering. A comparison of results reveals that sequential minimal optimization (S.M.O.) is better than I.B.K. and B.F. Tree processes, i.e. 97.71%.
Glioblastomas brain tumour segmentation based on convolutional neural network...IJECEIAES
Brain tumour segmentation can improve diagnostics efficiency, rise the prediction rate and treatment planning. This will help the doctors and experts in their work. Where many types of brain tumour may be classified easily, the gliomas tumour is challenging to be segmented because of the diffusion between the tumour and the surrounding edema. Another important challenge with this type of brain tumour is that the tumour may grow anywhere in the brain with different shape and size. Brain cancer presents one of the most famous diseases over the world, which encourage the researchers to find a high-throughput system for tumour detection and classification. Several approaches have been proposed to design automatic detection and classification systems. This paper presents an integrated framework to segment the gliomas brain tumour automatically using pixel clustering for the MRI images foreground and background and classify its type based on deep learning mechanism, which is the convolutional neural network. In this work, a novel segmentation and classification system is proposed to detect the tumour cells and classify the brain image if it is healthy or not. After collecting data for healthy and non-healthy brain images, satisfactory results are found and registered using computer vision approaches. This approach can be used as a part of a bigger diagnosis system for breast tumour detection and manipulation.
Twin support vector machine using kernel function for colorectal cancer detec...journalBEEI
Nowadays, machine learning technology is needed in the medical field. therefore, this research is useful for solving problems in the medical field by using machine learning. Many cases of colorectal cancer are diagnosed late. When colorectal cancer is detected, the cancer is usually well developed. Machine learning is an approach that is part of artificial intelligence and can detect colorectal cancer early. This study discusses colorectal cancer detection using twin support vector machine (SVM) method and kernel function i.e. linear kernels, polynomial kernels, RBF kernels, and gaussian kernels. By comparing the accuracy and running time, then we will know which method is better in classifying the colorectal cancer dataset that we get from Al-Islam Hospital, Bandung, Indonesia. The results showed that polynomial kernels has better accuracy and running time. It can be seen with a maximum accuracy of twin SVM using polynomial kernels 86% and 0.502 seconds running time.
Breast Cancer Detection from Mammography Images Using Machine Learning Algorithms (U-Net Segmentation and Dense Net Classifier implementation are in progress)
Breast cancer histological images nuclei segmentation and optimized classifi...IJECEIAES
Breast cancer incidences have grown worldwide during the previous few years. The histological images obtained from a biopsy of breast tissues are regarded as being the highest accurate approach to determine whether any cells exhibit symptoms of cancer. The visible position of nuclei inside the image is achieved through the use of instance segmentation, nevertheless, this work involves nucleus segmentation and features classification of the predicted nucleus for the achievement of best accuracy. The extracted features map using the feature pyramid network has been modified using segmenting objects by locations (SOLO) convolution with grasshopper optimization for multiclass classification. A breast cancer multiclassification technique based on a suggested deep learning algorithm was examined to achieve the accuracy of 99.2% using a huge database of ICIAR 2018, demonstrating the method’s efficacy in offering an important weapon for breast cancer multi-classification in a medical setting. The segmentation accuracy achieved is 88.46%.
Developed Project with 3 more colleagues for Pneumonia Detection from Chest X-ray images using Convolutional Neural Network. Used confusion matrix, Recall, Precision for check the model performance on testing Data
BREAST CANCER DIAGNOSIS USING MACHINE LEARNING ALGORITHMS –A SURVEYijdpsjournal
Breast cancer has become a common factor now-a-days. Despite the fact, not all general hospitals
have the facilities to diagnose breast cancer through mammograms. Waiting for diagnosing a breast
cancer for a long time may increase the possibility of the cancer spreading. Therefore a computerized
breast cancer diagnosis has been developed to reduce the time taken to diagnose the breast cancer and
reduce the death rate. This paper summarizes the survey on breast cancer diagnosis using various machine
learning algorithms and methods, which are used to improve the accuracy of predicting cancer. This survey
can also help us to know about number of papers that are implemented to diagnose the breast cancer.
SVM &GA-CLUSTERING BASED FEATURE SELECTION APPROACH FOR BREAST CANCER DETECTIONijscai
Mortality leading among women in developed countries is breast cancer. Breast cancer is women's second most prominent cause of cancer mortality worldwide. In recent decades, women's high prevalence of breast cancer has risen dramatically. This paper discussed several data analysis methods used to detect breast cancer early. Breast cancer diagnosis distinguishes benign and malignant breast lumps. Using data processing tools, we tackled this disease analysis. Data mining is an important step of library discovery where intelligent methods are used to detect patterns. Several clinical breast cancer studies were conducted using soft computing and machine learning techniques. Sometimes their algorithms are easier, easier, or more comprehensive than others. This research is focused on genetic programming and machine learning algorithms to reliably identify benign and malignant breast cancer. This study aimed to optimise the testing algorithm. We used genetic programming methods to choose classification machines' best features and parameter values. Data mining is an important step of library discovery where intelligent methods are used to detect patterns. We are analysing data accessible from the U.C.I. deep-learning data set in Wisconsin. In this experiment, we equate four Weka clustering strategies with genetic clustering. A comparison of results reveals that sequential minimal optimization (S.M.O.) is better than I.B.K. and B.F. Tree processes, i.e. 97.71%.
An Analysis of The Methods Employed for Breast Cancer Diagnosis IJORCS
Breast cancer research over the last decade has been tremendous. The ground breaking innovations and novel methods help in the early detection, in setting the stages of the therapy and in assessing the response of the patient to the treatment. The prediction of the recurrent cancer is also crucial for the survival of the patient. This paper studies various techniques used for the diagnosis of breast cancer. Different methods are explored for their merits and de-merits for the diagnosis of breast lesion. Some of the methods are yet unproven but the studies look very encouraging. It was found that the recent use of the combination of Artificial Neural Networks in most of the instances gives accurate results for the diagnosis of breast cancer and their use can also be extended to other diseases.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Comparison of Image Segmentation Algorithms for Brain Tumor DetectionIJMTST Journal
This paper deals with the implementation of Simple Algorithms for detection of size and shape of tumor in brain using MRI images. Generally, CT scan or MRI that is directed into intracranial cavity produces a complete image of brain. This image is visually examined by the physician for detection & diagnosis of brain tumor. However this method of detection resists the accurate determination of stage & size of tumor. To avoid that, this project uses computer aided method for segmentation (detection) of brain tumor by applying Fuzzy C-Means, K-Means, Gaussian Kernel and Pillar K-means algorithms. This segmentation process includes a new mechanism for clustering the elements of high-resolution images in order to improve precision and reduce computation time. The system applies FCM, Gaussian kernel and K-means clustering to the image later optimized by Pillar Algorithm. It designates the initial centroids’ positions by calculating the Euclidian distance metric between each data point and all previous centroids. Then it selects data points which have the maximum distance as new initial centroids. This algorithm distributes all initial centroids according to the maximum accumulated distance metric. In addition, it also reduces the time for analysis. At the end of the process the tumor is extracted from the MRI image and its exact position and the shape is also determined. This paper evaluates the proposed approach for Brain tumor detection by comparing with K-means, Fuzzy C means, Gaussian Kernel and manually segmented algorithms. The experimental results clarify the effectiveness of proposed approach to improve the segmentation quality in aspects of precision and computational time.
COVID-19 detection from scarce chest X-Ray image data using few-shot deep lea...Shruti Jadon
In the current COVID-19 pandemic situation, there is an urgent need to screen infected patients quickly and accurately. Using deep learning models trained on chest X-ray images can become an efficient method for screening COVID-19 patients in these situations. Deep learning approaches are already widely used in the medical community. However, they require a large amount of data to be accurate.
SVM &GA-CLUSTERING BASED FEATURE SELECTION APPROACH FOR BREAST CANCER DETECTIONijscai
Mortality leading among women in developed countries is breast cancer. Breast cancer is women's second most prominent cause of cancer mortality worldwide. In recent decades, women's high prevalence of breast cancer has risen dramatically. This paper discussed several data analysis methods used to detect breast cancer early. Breast cancer diagnosis distinguishes benign and malignant breast lumps. Using data processing tools, we tackled this disease analysis. Data mining is an important step of library discovery where intelligent methods are used to detect patterns. Several clinical breast cancer studies were conducted using soft computing and machine learning techniques. Sometimes their algorithms are easier, easier, or more comprehensive than others. This research is focused on genetic programming and machine
learning algorithms to reliably identify benign and malignant breast cancer. This study aimed to optimise the testing algorithm. We used genetic programming methods to choose classification machines' best features and parameter values. Data mining is an important step of library discovery where intelligent methods are used to detect patterns. We are analysing data accessible from the U.C.I. deep-learning data
set in Wisconsin. In this experiment, we equate four Weka clustering strategies with genetic clustering. A comparison of results reveals that sequential minimal optimization (S.M.O.) is better than I.B.K. and B.F. Tree processes, i.e. 97.71%.
Glioblastomas brain tumour segmentation based on convolutional neural network...IJECEIAES
Brain tumour segmentation can improve diagnostics efficiency, rise the prediction rate and treatment planning. This will help the doctors and experts in their work. Where many types of brain tumour may be classified easily, the gliomas tumour is challenging to be segmented because of the diffusion between the tumour and the surrounding edema. Another important challenge with this type of brain tumour is that the tumour may grow anywhere in the brain with different shape and size. Brain cancer presents one of the most famous diseases over the world, which encourage the researchers to find a high-throughput system for tumour detection and classification. Several approaches have been proposed to design automatic detection and classification systems. This paper presents an integrated framework to segment the gliomas brain tumour automatically using pixel clustering for the MRI images foreground and background and classify its type based on deep learning mechanism, which is the convolutional neural network. In this work, a novel segmentation and classification system is proposed to detect the tumour cells and classify the brain image if it is healthy or not. After collecting data for healthy and non-healthy brain images, satisfactory results are found and registered using computer vision approaches. This approach can be used as a part of a bigger diagnosis system for breast tumour detection and manipulation.
Twin support vector machine using kernel function for colorectal cancer detec...journalBEEI
Nowadays, machine learning technology is needed in the medical field. therefore, this research is useful for solving problems in the medical field by using machine learning. Many cases of colorectal cancer are diagnosed late. When colorectal cancer is detected, the cancer is usually well developed. Machine learning is an approach that is part of artificial intelligence and can detect colorectal cancer early. This study discusses colorectal cancer detection using twin support vector machine (SVM) method and kernel function i.e. linear kernels, polynomial kernels, RBF kernels, and gaussian kernels. By comparing the accuracy and running time, then we will know which method is better in classifying the colorectal cancer dataset that we get from Al-Islam Hospital, Bandung, Indonesia. The results showed that polynomial kernels has better accuracy and running time. It can be seen with a maximum accuracy of twin SVM using polynomial kernels 86% and 0.502 seconds running time.
Breast Cancer Detection from Mammography Images Using Machine Learning Algorithms (U-Net Segmentation and Dense Net Classifier implementation are in progress)
Breast cancer histological images nuclei segmentation and optimized classifi...IJECEIAES
Breast cancer incidences have grown worldwide during the previous few years. The histological images obtained from a biopsy of breast tissues are regarded as being the highest accurate approach to determine whether any cells exhibit symptoms of cancer. The visible position of nuclei inside the image is achieved through the use of instance segmentation, nevertheless, this work involves nucleus segmentation and features classification of the predicted nucleus for the achievement of best accuracy. The extracted features map using the feature pyramid network has been modified using segmenting objects by locations (SOLO) convolution with grasshopper optimization for multiclass classification. A breast cancer multiclassification technique based on a suggested deep learning algorithm was examined to achieve the accuracy of 99.2% using a huge database of ICIAR 2018, demonstrating the method’s efficacy in offering an important weapon for breast cancer multi-classification in a medical setting. The segmentation accuracy achieved is 88.46%.
Breast cancer detection using ensemble of convolutional neural networksIJECEIAES
Early detection leading to timely treatment in the initial stages of cancer may decrease the breast cancer death rate. We propose deep learning techniques along with image processing for the detection of tumors. The availability of online datasets and advances in graphical processing units (GPU) have promoted the application of deep learning models for the detection of breast cancer. In this paper, deep learning models using convolutional neural network (CNN) have been built to automatically classify mammograms into benign and malignant. Issues like overfitting and dataset imbalance are overcome. Experimentation has been done on two publicly available datasets, namely mammographic image analysis society (MIAS) database and digital database for screening mammography (DDSM). Robustness of the models is accomplished by merging the datasets. In our experimentation, MatConvNet has achieved an accuracy of 94.2% on the merged dataset, performing the best amongst all the CNN models used individually. Hungarian optimization algorithm is employed for selection of individual CNN models to form an ensemble. Ensemble of CNN models led to an improved performance, resulting in an accuracy of 95.7%.
The current big challenge facing radiologists in healthcare is the automatic detection and classification of masses in breast mammogram images. In the last few years, many researchers have proposed various solutions to this problem. These solutions are effectively dependent and work on annotated breast image data. But these solutions fail when applied to unlabeled and non-annotated breast image data. Therefore, this paper provides the solution to this problem with the help of a neural network that considers any kind of unlabeled data for its procedure. In this solution, the algorithm automatically extracts tumors in images using a segmentation approach, and after that, the features of the tumor are extracted for further processing. This approach used a double thresholding-based segmentation technique to obtain a perfect location of the tumor region, which was not possible in existing techniques in the literature. The experimental results also show that the proposed algorithm provides better accuracy compared to the accuracy of existing algorithms in the literature.
Breast cancer classification with histopathological image based on machine le...IJECEIAES
Breast cancer represents one of the most common reasons for death in the worldwide. It has a substantially higher death rate than other types of cancer. Early detection can enhance the chances of receiving proper treatment and survival. In order to address this problem, this work has provided a convolutional neural network (CNN) deep learning (DL) based model on the classification that may be used to differentiate breast cancer histopathology images as benign or malignant. Besides that, five different types of pre-trained CNN architectures have been used to investigate the performance of the model to solve this problem which are the residual neural network-50 (ResNet-50), visual geometry group-19 (VGG-19), Inception-V3, and AlexNet while the ResNet-50 is also functions as a feature extractor to retrieve information from images and passed them to machine learning algorithms, in this case, a random forest (RF) and k-nearest neighbors (KNN) are employed for classification. In this paper, experiments are done using the BreakHis public dataset. As a result, the ResNet-50 network has the highest test accuracy of 97% to classify breast cancer images.
Modified fuzzy rough set technique with stacked autoencoder model for magneti...IJECEIAES
Breast cancer is the common cancer in women, where early detection reduces the mortality rate. The magnetic resonance imaging (MRI) images are efficient in analyzing breast cancer, but it is hard to identify the abnormalities. The manual breast cancer detection in MRI images is inefficient; therefore, a deep learning-based system is implemented in this manuscript. Initially, the visual quality improvement is done using region growing and adaptive histogram equalization (AHE), and then, the breast lesion is segmented by Otsu thresholding with morphological transform. Next, the features are extracted from the segmented lesion, and a modified fuzzy rough set technique is proposed to reduce the dimensions of the extracted features that decreases the system complexity and computational time. The active features are fed to the stacked autoencoder for classifying the benign and malignant classes. The results demonstrated that the proposed model attained 99% and 99.22% of classification accuracy on the benchmark datasets, which are higher related to the comparative classifiers: decision tree, naïve Bayes, random forest and k-nearest neighbor (KNN). The obtained results state that the proposed model superiorly screens and detects the breast lesions that assists clinicians in effective therapeutic intervention and timely treatment.
An approach of cervical cancer diagnosis using class weighting and oversampli...TELKOMNIKA JOURNAL
Globally, cervical cancer caused 604,127 new cases and 341,831 deaths in 2020, according to the global cancer observatory. In addition, the number of cervical cancer patients who have no symptoms has grown recently. Therefore, giving patients early notice of the possibility of cervical cancer is a useful task since it would enable them to have a clear understanding of their health state. The use of artificial intelligence (AI), particularly in machine learning, in this work is continually uncovering cervical cancer. With the help of a logit model and a new deep learning technique, we hope to identify cervical cancer using patient-provided data. For better outcomes, we employ Keras deep learning and its technique, which includes class weighting and oversampling. In comparison to the actual diagnostic result, the experimental result with model accuracy is 94.18%, and it also demonstrates a successful logit model cervical cancer prediction.
A deep learning framework for accurate diagnosis of colorectal cancer using h...IJECEIAES
Colorectal cancer (CRC) is one of the most prevalent malignancies worldwide, with high mortality and incidence rates. Early detection of the disease may increase the probability of survival, making it critical to develop effective procedures for precise treatment. In the past few years, there has been an increased use of deep learning techniques in image classification that aid in the detection of various types of cancer. In this study, convolutional neural network (CNN) models were used to classify colorectal cancer into benign and malignant. After applying various data preprocessing techniques to the image dataset, we evaluated our prototypes using three distinct subsets of testing data, representing 20%, 30%, and 40% of the total dataset. Additionally, four pre-trained CNN models (ResNet-18, ResNet-50, GoogLeNet, and MobileNetV2) were trained, and the network architectural techniques were compared by applying the Adam optimizer. Finally, we assessed the performance of algorithms in terms of accuracy, sensitivity, specificity, precision, F1-score, and area under the receiver operating characteristic (ROC) curve (AUC). In this research, deep learning approaches demonstrated high efficacy in accurately diagnosing colorectal cancer. This indicates that these techniques have an important and significant value for advancing medical research.
Social media marketing (SMM) is a form of digital marketing that utilizes social media platforms to promote products, services, or brands. The goal of social media marketing is to connect with the target audience, build brand awareness, increase website traffic, and drive engagement and conversions. Here are some key aspects of social media marketing:
Strategy Development:
Identify your target audience: Understand the demographics, interests, and online behavior of your target audience.
Set clear goals: Define specific, measurable, achievable, relevant, and time-bound (SMART) goals for your social media campaigns.
Choose the right platforms: Select social media platforms that align with your target audience and business objectives.
Content Creation:
Create engaging content: Develop content that resonates with your audience, such as images, videos, infographics, and text posts.
Maintain consistency: Establish a consistent posting schedule to keep your audience engaged and informed.
Use a variety of content types: Experiment with different content formats to keep your social media presence diverse and interesting.
Audience Engagement:
Respond to comments and messages: Engage with your audience by responding to comments, messages, and mentions in a timely manner.
Encourage user-generated content: Encourage your followers to create and share content related to your brand.
Run contests and giveaways: Organize contests or giveaways to boost engagement and attract new followers.
Paid Advertising:
Utilize paid social media advertising: Platforms like Facebook, Instagram, Twitter, and LinkedIn offer advertising options to reach a larger audience.
Targeted advertising: Use advanced targeting options to reach specific demographics, interests, and behaviors.
Analytics and Monitoring:
Use analytics tools: Monitor the performance of your social media campaigns using analytics tools provided by the platforms or third-party tools.
Adjust strategies based on data: Analyze the data and adjust your strategies to optimize performance and achieve better results.
Influencer Marketing:
Collaborate with influencers: Partner with influencers who align with your brand to reach a wider audience and build credibility.
Leverage user trust: Influencers can help establish trust with their followers, leading to increased brand credibility.
Social Media Trends:
Stay updated: Keep track of emerging trends in social media marketing and adapt your strategies accordingly.
Experiment with new features: Platforms regularly introduce new features; experiment with these features to stay ahead of the curve.
Remember that effective social media marketing requires a consistent and strategic approach. Regularly assess your performance, listen to your audience, and adjust your strategies to meet your goals.
Multiclass skin lesion classification with CNN and Transfer LearningYashIyengar
A multiclass classifier has been developed using transfer learning approach to detect seven different types of skin lesions. A modified VGG19 architecture is used for this purpose. The proposed model has achieved 78 % validation accuracy and 74 % test accuracy.
A novel approach to jointly address localization and classification of breast...IJECEIAES
Localization of the cancerous region as well as classification of the type of the cancer is highly inter-linked with each other. However, investigation towards existing approaches depicts that these problems are always iindividually solved where there is still a big research gap for a generalized solution towards addressing both the problems. Therefore, the proposed manuscript presents a simple, novel, and less-iterative computational model that jointly address the localization-classification problems taking the case study of early diagnosis of breast cancer. The proposed study harnesses the potential of simple bio-inspired optimization technique in order to obtained better local and global best outcome to confirm the accuracy of the outcome. The study outcome of the proposed system exhibits that proposed system offers higher accuracy and lower response time in contrast with other existing classifiers that are freqently witnessed in existing approaches of classification in medical image process.
Similar to M sc research_project_report_x18134599 (20)
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
1. Breast Cancer Detection from
Histopathological Images using Deep
Learning and Transfer Learning
MSc Research Project
Data Analytics
Mansi Chowkkar
Student ID: 18134599
School of Computing
National College of Ireland
Supervisor: Cristina Muntean
2. National College of Ireland
Project Submission Sheet
School of Computing
Student Name: Mansi Chowkkar
Student ID: x18134599
Programme: Data Analytics
Year: 2019-2020
Module: Research Project
Supervisor: Cristina Muntean
Submission Due Date: 12/12/2019
Project Title:
Word Count: 6300
Page Count: 23
I hereby certify that the information contained in this (my submission) is information
pertaining to research I conducted for this project. All information other than my own
contribution will be fully referenced and listed in the relevant bibliography section at the
rear of the project.
ALL internet material must be referenced in the bibliography section. Students are
required to use the Referencing Standard specified in the report template. To use other
author’s written or electronic work is illegal (plagiarism) and may result in disciplinary
action.
Signature:
Date: 12th December 2019
PLEASE READ THE FOLLOWING INSTRUCTIONS AND CHECKLIST:
Attach a completed copy of this sheet to each project (including multiple copies).
Attach a Moodle submission receipt of the online project submission, to
each project (including multiple copies).
You must ensure that you retain a HARD COPY of the project, both for
your own reference and in case a project is lost or mislaid. It is not sufficient to keep
a copy on computer.
Assignments that are submitted to the Programme Coordinator office must be placed
into the assignment box located outside the office.
Office Use Only
Signature:
Date:
Penalty Applied (if applicable):
3. Breast Cancer Detection from Histopathological
images using Deep Learning and Transfer Learning
Mansi Chowkkar
x18134599
Abstract
Breast Cancer is the most common cancer in women and it’s harming women’s
mental and physical health. Due to complexities present in Breast Cancer images,
image processing technique is required in the detection of cancer. Early detection
of Breast cancer required new deep learning and transfer learning techniques. In
this paper, histopathological images are used as a dataset from Kaggle. Images are
processed using histogram normalization techniques. This research project imple-
ments the Convolutional Neural Network(CNN) model based on deep learning and
DenseNet-121 based on transfer-learning. Transfer learning uses the Imagenet pre-
trained model for training. Hyper-parameter tuning is done for increasing accuracy
and precision value. Research achieved 90.9 % test accuracy using the CNN model
and 88.03 % accuracy by the transfer learning model.
1 Introduction
Breast cancer is the most common cancer among women in the world. As compared
to other cancer types, Breast Cancer has the highest mortality rate. 8.8 million deaths
happened in 2015 due to breast cancer. This death rate will increase up to 2.1 million
per year over the existing rate , according to the world health organization Elelimy and
Mohamed (2018). For early detection, there are different imaging modalities such as
mammography, ultrasound, MRI, tomography, and biopsy. Mammography is a simple
and common technique for early detection of cancer but it cannot detect breast cancer
in a dense breast. CT scan method uses radiation which generates genetic mutation.
Ultrasound method produces false-positive results. Breast Biopsy is the most precise
solution where a sample piece of tissue is taken for the microscopic examination. This
study of biological tissues under the microscope is called histopathology as mentioned
in Kaushal, Bhat, Koundal and Singla (2019). An expert doctor is required to examine
the histopathological images and it is a time-consuming process. The computer-assisted
diagnosis(CAD) is used for classifying malignant and benign cancer from images. This
study focuses on classifying histopathological images using deep learning and transfer
learning techniques. The use of Deep learning will reduce the load from the radiologist
and they can more concentrate on only suspicious cases.
As deep learning-based CNN uses more layers, new problems are arising. The gradient
is transferred to many layers and information can get lost by the time it reaches to the last
layer. This problem can be addressed by using VGG16, ResNet, CNN and DenseNet.
Although these models have different network structure and training procedures, they
1
4. serve the same feature of providing a connecting path from the previous layer to the next
layer. DenseNet reduces gradient problem and provides feature reusability. DenseNet
reduces the number of parameters used in the network for training the model as mentioned
by Huang et al. (2017).
This study will use Histogram normalisation as an image processing technique. Pro-
cessed images are augmented to avoid biasing and overfitting problems. This research
uses CNN with a parameter tuning process to improve the accuracy. This approach
increases the true negative number for cancer detection. Transfer learning will be imple-
mented for avoiding the over-fitting problem and training the model on less number of
images for specific magnification Sabeena Beevi et al. (2019). Transfer learning is used
with the Imagenet pre-trained weight and Sigmoid classification function. Deep learning
and fine-tuning method will improve the diagnostic efficiency and provide accurate cancer
detection results to the doctors. This behavior will add diagnostic value to the medical
domain by saving time.
1.1 Research Question
This research tries to find a solution to the following research question:
How efficiently, accurately and quickly Deep learning and Transfer learning
algorithm detect Breast Cancer from Histopathological images?
2 Related Work
Many Deep learning researches on breast cancer detection are done on mammography and
ultrasound images. Recent advancement in transfer learning and fine-tuning technology
is useful in the medical imaging field for efficient detection of diseases. To detect cancer
from image data, various image processing, deep learning and transfer learning methods
are discuss and summarised below.
2.1 Deep Learning in Health care
In Machine learning, Deep learning algorithms are rapidly becoming a choice for the
medical image analysing technique. Research done by Litjens et al. (2017) summarises
image classification techniques using deep learning models. This research study explains
various packages in python for CNN models. It compares all the CNN models and pre-
possessing techniques for medical image data. This study has shown that Deep learning
has a good impact on medical image analysis. In the research study Li, Xie and Shen
(2019), Atrous Dense Net(ADN) and deep-reverse active learning(DRAL) is applied to
pathological image data. DRAL is used to remove mislabelling from the training of can-
cer image data. To overcome the over-fitting problem RefineNet with six layers of CNN
is adopted for the DRAL model in this research. Statistical analysis proved that VGG-16
and ADN models both achieved the same performance. ADN achieved better accuracy
as compared to DenseNet-121, ResNet and VGG16. DenseNet-121 has a large model size
hence it can perform better on different datasets.
The deep learning approach is used in Vo et al. (2019) for image classification. The
deep convolution neural network learning is used in this research for feature extraction
from multi-train training image data. This has improved multiple labeled cancer detection
accuracy. The use of gradient boosting is also proposed for enhancing the performance
2
5. of classification. In this study, the Inception-ResNet-v2 model is compared with the
Cruz-Roa-CNN model and the Inception-ResNet-v2 model outperformed. To improve
the performance of this model, in future use of large image data is suggested to overcome
the data limitation problem. Hence deep learning models will provide better accuracy on
large histopathological image data. A recent study Azli et al. (2018) classifies cancer from
mammography images through a backpropagation neural networks. Research shows that
CNN with 100 hidden nodes improves the performance of the classifier than CNN with 3
nodes. CNN with 100 nodes achieved an accuracy of 70%. The architecture of CNN can
be modified to improve accuracy with hyper parameter tuning and adding Dense layers.
In a study performed by Xiang et al. (2019), CNN is implemented for the classification of
histopathological images. Dataset consists of 4 different types of magnifications images.
All images with all magnifications are used in the training process of the CNN model
and accuracy is calculated. Data augmentation and cross-validation methods are used
for improving model accuracy. The softmax classifier is used in the output layer of
the CNN. This study achieved 93.2% accuracy. Stochastic gradient descent (SGD) and
backpropagation are used here for the parameter updating process. The similar data is
being used in this study hence this research can be extended here. To extend Xiang et al.
(2019) study different architectures for the CNN model can be used in this research with
different parameter tuning.
2.2 Image Processing techniques
Detection of cancerous cells from histopathological images helps in the early detection pro-
cess in medical diagnosis. In Medical Image processing, Normalisation plays an important
role in reducing contrast and intensity. In the Anghel et al. (2019) study, stain normalisa-
tion is performed on the image data using the unsupervised stain-vector method. In this
study of classification using the neural network, the algorithm is compared before stain
normalisation and after stain normalisation. The study proved that Stain normalisation
improved the accuracy of the CNN model from 79% to 87%. Hence normalisation has a
major impact on the accuracy of CNN model. Colour normalisation and stain separation
from the images are difficult processing. In Hidalgo-Gavira et al. (2018) this paper, blind
colour deconvolution method based on the Bayesian technique is proposed. This method
finds stain concentration from the histopathological images based on the color vector mat-
rix. This method is proved to be faster, easier, and robust to different parameter values.
Kaushal, Koundal and Singla (2019) study compares different segmentation techniques to
detect cancerous cells from histopathological images. These techniques are evaluated us-
ing accuracy, true positive and false positive rate. Segmentation can be performed before
applying a classification model to image data for better performance of classification.
Jiang et al. (2019) proposed a deep learning method symmetric residual convolutional
neural network (SR-CNN) method for enhancing the images by sharpening the edges
of the images. Results are evaluated using real-time patient data by augmenting images
under-sample cone-beam computed tomography (CBCT). In Rundo et al. (2019) research
study, the MedGa image enhancement process is introduced which uses a bimodal intens-
ity histogram. This method improves image quality and appearance for the region of in-
terest. This method crops the image in the rectangular box to increase the region interest
of the image. Enhancing the contrast of the actual region of interest is more important
than enhancing the whole contrast of the image in medical image processing. Zero-mean
normalisation method is used for pre-processing mammography images for breast cancer
3
6. classification inLi, Zhuang, Li, Zhao and Ma (2019). This processing enhanced the model
robustness and increased the training speed of the model. Image normalisation can also
have an impact on accuracy, true negative values which can be evaluated in this research
study.
2.3 DenseNet
Deep learning is used by most of the researchers for classifying medical images. In the
research Li, Zhuang, Li, Zhao and Ma (2019), mammography images are classified as
benign and malignant cancer using the DenseNet-2 model. This model replaces the first
convolution layer with Inception-Net which increases the performance of the model. Aug-
mentation is used to overcome the over-fitting problem and data insufficiency problem
in this research. Ten fold cross-validation is performed to check which model performed
best. The result has shown that DenseNet-2 performed better with 94.55% accuracy than
VGGNet, DenseNet, and AlexNet. DenseNet can be evaluated on different datasets to
check its performance using some parameter tuning in the network without any modi-
fications. Recent research on CNN shown that Dense Net can provide accurate results
with less number of parameters Gottapu and Dagli (2018). In this research DenseNet,
architecture is used for brain MRI image segmentation. This method is used for finding
abnormalities in the brain. Results are compared with and without compression. This
model achieved 95.5 % accuracy. Different dataset (brain segmentation) is tested to make
DenseNet model as a benchmark. For benchmarking this model can even be tested on
images without segmentation as a classifier.
For classifying pancreatic cancer from CT scan images Li, Reichert, Lin, Tselousov,
Braren, Fu, Schmid, Li, Menze and Shi (2019) research study applied Dense Net feature
learning model. This method is based on the CAD approach for the early detection and
segmentation of pancreatic cancer. For the visualising results for the radiologists, the
study implemented the saliency maps analysis tool. DenseNet model achieved 72.8 %
accuracy which is higher than the diagnostic accuracy. As compared to other researches
this accuracy is lower and hence more research can be done on DenseNet to increase the
accuracy. Research Oyama and Yamanaka (2017) implemented the DenseNetSal model
for predicting saliency maps which uses VGG16 with 150 convolution layers. DenseNet-
161 is used by this research which uses 161 convolution layers and four dense blocks.
DenseNet-161 architecture is modified here and it is observed that it performed better
than VGG16. Here accuracy of DenseNet-161 increased by using two scales as 0.5 and
1, in future parameters can also be changed to see the impact on accuracy. Huang
et al. (2017) compared DenseNet-121, DenseNet-161, DenseNet-201, and DenseNet-BC on
different image data and imagenet. A study proved that DenseNet requires less parameter
change and less computation time for achieving an accuracy similar to ResNets. As the
parameters grow in DenseNet, it shows consistent performance growth.
2.4 Transfer Learning
In the medical image classification and diagnosis, CNN plays an important role and its
a growing technique. For training the CNN model we require a large amount of image
data but in the medical domain, there is an insufficient amount of image data due to
privacy-preserving policies. Hence transfer learning is growing in the medical image
diagnosis field rapidly to overcome this problem Vogado et al. (2018). In Talo (2019)
4
7. study, transfer learning is used for classifying histopathological images. ResNet-50 is
used for classification of colored images and DenseNet-161 is used for grayscale image
classification and it achieved %95.7 accuracy. These models use pre-trained weights
which are trained on ImageNet with 1000 classes of image data. DenseNet-161 uses
fewer parameters than CNN and it reduces over-fitting problems. The transfer learning
method increases the rate of the training process and accelerates CNN model construction.
DenseNet-121 uses even less parameters than DenseNet-161. In the future, this model
can be used as a transfer learning method using imagenet as a pretrained weight.
Sabeena Beevi et al. (2019) research study used the transfer learning approach by
pre-trained convolutional neural networks for feature extraction. In this study, the out-
put layer of softmax activation is replaced by random forest. In this research automated
mitosis detection from the histopathological images is carried out using VGGNet and
Caffe model with ImageNet as a pre-trained weight. The use of the transfer learning ap-
proach overcome the issue of an insufficient large amount of labeled data. In the research
study Vogado et al. (2018), leukaemia diagnosis is made using transferred learning where
pre-trained CNN algorithms( VGG, AlexNet)are used for feature extraction. The SVM
classifier then classifies these extracted features. In this study robustness of the model is
proved using three different data-sets in training. Results are evaluated using confusion
matrix, accuracy and kappa index. The transferred learning-based model’s performance
is calculated in d. N´obrega et al. (2018) to classifying lung cancer from CT lung images.
Transfer learning models are used here to extract the features and then it is provided to
the classifier. It is seen that ResNet50 and SVM RBF provided the best results with f1
score 78.83% which is more than the related work on the same dataset. In future models
based on transfer, learning can be used to improve accuracy.
2.5 Hyper parameter Tuning
Deep learning has shown progress in image classification in the medical field. Deep
learning requires a large number of images for training, validation, and testing. To resolve
this problem Oyama and Yamanaka (2017) used a pre-trained CNN model and performed
fine-tuning layer by layer using a transfer learning strategy. Accuracy is improved to 94%
on contrast-enhanced magnetic resonance images(CE-MRI) data using a cross-validation
method. Tao Tan (2018) study implements transfer learning using DenseNet-121 model
with fine tuning. Experiment uses numbers of convolutional layers and the number of
epochs as a part of the fine tuning. Study obtained 82.0% accuracy overall after fine
tuning. Research shown the benefits of sequential fine-tuning in cancer detection. Due
to the insufficient amount of image data study could not able to tune a large number of
parameters in the model. This can be extended in this research study with comparatively
large image data set.
2.6 Summery
This study implements a histopathological image processing technique as we have seen
image normalisation increases the accuracy. As mentioned in Xiang et al. (2019), this
research will use a basic CNN model built from scratch for classification with parameter
tuning. DenseNet has shown great value in medical image classification. A little research
is done on DenseNet for histopathological images. Therefore this research paper extending
the use of DenseNet based on transfer learning for histopathological data. In DenseNet
5
8. models, very few research is available on DenseNet-121, this research will extend the
transfer learning-based DenseNet-121 model used in Tao Tan (2018). As we have seen
parameter tuning increases the accuracy without much modifications and cost to the
model. Therefore this research implements parameter-tuning on DenseNet and DenseNet-
121 parameters to get the best suitable parameters for histopathological image data.
3 Methodology
This research work applied the CNN model to histopathological images to classify benign
and malignant cancer accurately as discussed in 2.6. For implementing the transfer learn-
ing method, research uses the DenseNet-121 model as a novelty on the histopathological
image data set. The Cross-Industry Standard Process for Data Mining (CRISP-DM)
method is used for this research as shown in figure 3. It will be explained in Business
understanding, data understanding, data preparation, and, modeling sections below as
mentioned inTie et al. (2011).
3.1 Business Understanding
Detection of cancer from mammography images is a very common method but a breast
biopsy is a very precise method and its prediction requires more time and expertise.
Pathologists and doctors evaluate histopathological images after preprocessing of images.
As the preparation of images involves sectioning of images, collection, staining, and fix-
ation step Elelimy and Mohamed (2018). Hence evaluation of biopsy images under the
microscope is very time consuming for pathologists and it is highly impacted on the ex-
perience of the doctor in this field. As discussed in the related work summary, a deep
Learning-based CNN model is used in this research. CNN is used to reduce the time re-
quired for evaluating histopathological images. In the medical field, images are insufficient
for training the CNN model with a large number of layers. Hence transfer learning-based
DenseNet-121 model and parameter-tuning approach are used to overcome this problem.
3.2 Data Understanding
Image data is taken from Kaggle with version 1 1
. This data is having only two kernels
available. Total 7909 images are present with four different magnification as 40X, 100X,
200X, and, 400X. Table 1 explains the image distribution concerning image magnifications
and the two classes. Figure 1 shows malignant and benign images of all the magnifications
from the dataset. 40X B indicates a benign image of 40X magnification and 40X M
indicates a 40X magnification image of malignant type. These images are then divided
into train, test and valid folders with 80:10:10 proportion as explained in table 1. 5439
images are of type malignant and 2480 images of type benign. All images are three
dimensional with each image size 460*700. Considering all these factors about image
data, all data preparation steps are discussed in the following section.
1
https://www.kaggle.com/kritika397/breast-cancer-dataset-from-breakhis
6
9. Figure 1: Different Magnification Images
Magnification Benign Malignant Total
40X train: 500
valid: 62
test: 63
train: 1176
valid: 92
test: 97
1990
100X train: 515
valid: 64
test: 65
train: 1149
valid: 144
test: 144
2081
200X train: 498
valid: 62
test: 63
train: 1112
valid: 139
test: 139
2013
400X train: 470
valid: 59
test: 59
train: 985
valid: 123
test: 124
1820
Total 2480 5429 7909
Table 1: Image distribution as per the class and magnification
3.3 Data Preparation
For providing images to the CNN and DenseNet-121, images are normalised for more
readability. Images then augmented and saved using rotation and batch size parameters.
At the time of model execution images are augmented run time using different parameters
such as rotation, zoom, scaling, batch size and image size.
3.3.1 Image Normalisation
Medical images require image enhancement to help pathologists for detecting disease
and its diagnosis. Images go under an automated deep learning model needs to im-
prove its quality for better results. In Zhuang and Guan (2017), Histogram Equalization
method for image enhancement based on mean and variance is performed to increase the
contrast and brightness of the image without compromising details of the image data.
As mentioned above, in this research the Histogram Equalization method is applied to
histopathological image data for increasing contrast and brightness of the image. Histo-
pathological images need to undergo microscopic examination hence they contain some
staining process and difficult to read. Image normalization will increase the contrast
of the image. Hence it increases the accuracy and performance of the deep learning
7
10. model. Figure 2 shows the effect of histogram normalisation on one of the benign types
of histopathological images.
Figure 2: Histogram Normalisation effect
3.3.2 Image Augmentation and Upsampling
After normalising images, images are upsampled before providing it to CNN based
DenseNet model and the transfer learning. In the study Tellez et al. (2019), basic aug-
mentation is used with scaling, gaussian blurring, and gaussian noise. Morphological aug-
mentation consists of horizontal and vertical mirroring of images, 90 degree, 180-degree
image rotations or scaling of images by 0.5 to 1.0 factor. In this research, benign im-
ages are upsampled to match with malignant image count. For augmentation, 90-degrees
rotation and batch size of 32 is used. While providing as an input to the classification
model run time augmentation is performed by tuning different parameters as explained in
detail in the evaluation section. Table 2 explains the number of benign histopathological
images with all magnification before and after upsampling. Table 3 explains all the hy-
perparameters used and their corresponding values tuned for selecting the best suitable
parameters for the classifiers. Effect of all values from table 3 is explained in detail in
the section 6.
Magnification Benign im-
ages before
upsampling
Benign images
after upsampling
40X 500 986
100X 515 1017
200X 498 982
400X 470 932
All magnifications 1983 3795
Table 2: Histopathological image data distribution
8
11. Parameters in Augmentation Values used
Image Rotation 60, 90, 180, 270
Width shift range 0.2
Height shift range 0.1
zoom range 0.4, 0.5
target size 224*224, 128*128, 64*64, 32*32
batch size 64, 32,16
shuffle true
colour mode RGB
horizontal flip true
Table 3: Parameters used in Image Augmentation
3.4 Modelling
This section explains the architecture used for the CNN and DenseNet-121 model.
Figure 3: Flow Diagram
3.4.1 CNN model
CNN consists of Transition layers, dense blocks, and convolutional layers. CNN consists
of a sequence of neural networks for extracting features. To circulate features at different
9
12. level it concatenates all features from previous layers. There are mainly five layers present
in the architecture of the CNN as mentioned in Sabeena Beevi et al. (2019). To capture
the local information from the input, the convolutional layer is used. The filter is used to
compute the dot product of its weights and the input image region. The batch normalisa-
tion layer controls the output from the previous layer and sends it to the next activation
layer. Since normalisation is used in the gradient process, it balances the weights in the
network. Rectified Linear Unit (ReLU) is used as an activation function in the CNN.
ReLU accelerates the execution of CNN. The complexity of the features increases as CNN
becomes deeper. MaxPooling layer is used for shrinking the feature map size. Each fully
connected layer neurons connect with the previous layer for calculating class count. The
output layer consists of a softmax or sigmoid as a classification function.
3.4.2 DenseNet121 model based on Transfer Learning
Transfer learning requires a small number of images for training the model. The transfer
learning approach is used in this research study by using DenseNet-121. The model uses
121 layers and Imagenet as a pre-trained weight as shown in the figure 3. Dense layers
are very narrow and classification is based on knowledge of already trained weights as
shown in. Therefore transfer learning learns complicated training of images from fewer
parameters and less computation time Oyama and Yamanaka (2017). Table 4 explains
the architecture of the DenseNet-121 model as explained in Allaouzi et al. (2019). It
gives details about output layers and convolutional layers used for each dense block and
transition block.
Layers Output DenseNet-121
k=12
Convolution 112*112 7 * 7 Conv, stride2
Pooling 56 *56 3*3 max pooling, stride2
Dense Block 1 56*56
1 ∗ 1conv
3 ∗ 3conv
*6
Transition 1 56*56
28*28
1*1 conv
2*2 Avg pool, stride 2
Dense Block 2 28*28
1 ∗ 1conv
3 ∗ 3conv
*12
Transition 2 28*28
14*14
1*1 conv
2*2 Avg pool, stride 2
Dense Block 3 7*7
1 ∗ 1conv
3 ∗ 3conv
*24
Transition 3 14*14
7*7
1*1 conv
2*2 avaerage pool, stride 2
Dense Block 4 7*7
1 ∗ 1conv
3 ∗ 3conv
*16
Classification
layer
1*1(Sigmoid) 7*7 global avg
pool
Table 4: DenseNet-121 Architecture
10
13. 4 Design Specification
The following are the specifications used for both the models.
Convolution layer: In this layer the number of filters is present is the same as the
growth rate for extracting features Khened et al. (2019). In this experiment growth rate
is 12 hence there are 12 filters present in the DenseNet model.
BatchNormalisation: When the output of the previous convolution layer provided
to the next layer, BatchNormalisation is used for normalising output which reduces the
overfitting problem.
MaxPooling2D: One max-pooling layer is used in this project to reduce the dimen-
sionality of the feature map.
Dropout layer: 0.2 Dropout is set to skip the connection for matching the input
channels. Dropout layer helps in reducing overfitting of the model
Dense (Fully Connected): The dense block is placed before the output classifier for
aggregating the information which is extracted from previous layers. In this study, three
dense blocks are used in the DenseNet model.
Softmax layer: The final output activation function in the dense layer used is a
softmax layer which generates the final classification label map.
Loss Function: In medical images, class imbalance occurs between background and
region of interest. To overcome this issue here the cross-entropy loss function is used.
Weight: In the transfer learning ’imagenet’ pre-trained weight is used for training
DenseNet-121.
Optimizer: Adam optimizer is used here with a learning rate of 0.0001.
Sigmoid layer: The final output activation function in the dense layer of DenseNet-
121 used is a Sigmoid layer which generates the final classification label map.
5 Implementation
This section provides information on how the two models are implemented to classify
breast cancer accurately from histopathological images.
5.1 Setup
CNN and transfer learning models require more time to process the images while training
the model. The experiment is carried out on the Google colaboratory with 100 GB drive
storage, 12.72 GB RAM and 48.97 GB run-time GPU. Both the models require more
time for execution on large image data as they consist of more layers. Runtime usage of
GPU accelerates the execution time of CNN and DenseNet-121. For implementing CNN
and DenseNet121, python libraries Keras and TensorFlow are used. Python version 3 is
used on google colaboratory notebook. Data is uploaded on google drive and accessed by
mounting google drive using python API. Image normalisation and up-sampling are done
using numpy, PIL and Keras libraries from python. Normalised and Upsampled images
are provided to the CNN and DenseNet-121 as shown in the figure 3.
5.2 Data Handling
All the histopathological images are read from google drive and loaded in the data frame.
Different data frames according to benign and malignant classes concerning different
11
14. magnifications are red into different files. Function called get histogram() is written for
histogram normalisation. This function is called from hist0() by providing flatten images
and size 256. hist0() function takes the filename as an input with all images which need
to be normalised. New normalised images are written into a new folder in a for loop for
further use in the deep learning models.
Normalised images are red into data frames and checked for its biasing. It is then con-
verted into .csv file. For getting better performance benign images are upsampled. While
training the model on mixed magnification images, first all mixed magnification images
are mixed and then upsampled as shown in table 2. ImagedataGenerator API from the
Keras library is used for upsampling. Images are rotated by 90 degrees while augment-
ing and saved in the directory using datagen.flow() function. Run time augmentation is
also carried out to provide train images to the model. Parameters in augmentation are
changed accordingly in different case studies to study its impact on the accuracy of the
model.
5.3 Architecture of the CNN model
For implementing the CNN model, functions are written to create a convolution layers,
Dense blocks, transition layers, and Dense network.
1. Convolution layers function is written using ’relu’ activation function and 0.2
dropout value. Dropout is changed to 0.3 to check the impact on accuracy but it is seen
that 0.2 dropout provides better accuracy than 0.3. Convolution layer takes input as a
dense block and filter size. The filter size is the same as that of the growth rate which is
set to 12. This layer returns the convolution layer as a function of dropout, normalization,
and activation.
2. The dense Block function is written which returns dense block concatenated with
the previous input layer. This layer also returns the number of filters in the block.
3. Transition block consists of the Batch normalization layer, 2D convolution layer,
’relu’ activation layer and average pooling layer.
4. Finally, a Dense net is created which creates an input layer with input image shape.
Input image shape is changed in different case studies to check its impact on accuracy.
The output layer is created with the ’softmax’ activation function.
5. CNN is created with three dense blocks and five layers. The model has then
compiled with ’Adam’ optimizer and ’cross-entropy’ loss function.
6. model.fit generator() API is used to run the model with train data and valid data.
Epochs are varied in different case studies.
The model is compiled using the model.compile() function with Adam optimizer and
0.0001 learning rate.
5.4 Transfer Learning: DenseNet-121
Transfer learning is used to transfer the knowledge of already trained models to an-
other model for the classification of image data. The DenseNet-121 model is used
with the ’Imagenet’ pre-trained weight as shown in figure 3. In the output layer, the
Sigmoid activation function is used. Model is executed on the number of epochs us-
ing model.fit generator() function. This function takes training images, validation im-
ages, steps count and number of epochs as an input. class weight parameter is used
with value ’auto’ to improve the true negative values( cancer count). figure 4 shows
12
15. the code used in DenseNet-121 model. Test accuracy of the model is calculated using
model.evaluate generatot() function.
Figure 4: DenseNet-121
6 Evaluation
This section discusses the comprehensive analysis of two models and all parameters which
are tuned to get the best performance. For this research study, two models are used on
the histopathological image dataset from Kaggle. DenseNet-121 is used as a novelty on
this data set. For evaluation process loss and accuracy of training and validation data is
calculated per epoch for both the models. Accuracy and Loss plot for training and valid-
ation is plotted. The test accuracy is calculated as a function of prediction and test data.
Confusion metrics for each model is calculated to get a true positive and true negative
value. The experiments explained below, discuss the increased accuracy and increased
value of true negative as they are important factors for medical diagnosis. Precision
(Specificity) is also calculated as a function of true negative and false positive value. As
mentioned in Tao Tan (2018), fine-tuning is performed by changing different parameters
in the model and the effect on accuracy is observed in the following experiments.
13
16. 6.1 Experiment 1: Effect of Different magnifications on test
Accuracy of the CNN model
All magnification images are provided as training and validation data in the CNN model.
In this experiment, test accuracy is calculated for different magnification images and
results are listed in table 5. The experiment evaluates accuracy on different image sizes as
shown in table 5. Accuracy is also compared against different image rotation (60,90,180)
as mentioned by Li, Zhuang, Li, Zhao and Ma (2019). Table 5 shows image size, rotation
of the image at the time of augmentation and magnification impact on test accuracy.
Image size Rotation Accuracy
for 40X
Accuracy
for 100X
Accuracy
for 200X
Accuracy
for 400X
Accuracy
for Mix
128*128 60 81.25 76.07 70.79 67.75 71.08
128*128 90 70.62 83.7 85.6 82.4 84.15
128*128 180 77.5 84.6 84.15 81.42 75.59
64*64 60 63.75 83.7 85.14 78.14 72.01
64*64 90 65 86 82.67 80.87 73.2
64*64 180 71.25 90.9 87.12 81.42 76.65
32*32 60 58.75 81.3 78.12 74.8 68.96
32*32 90 61.25 87.08 85.04 81.96 72.14
32*32 180 68.12 85.16 84.65 79.23 71.08
Table 5: Test Accuracy of CNN for different Magnifications
In a research study, Kumar et al. (2020) accuracy are compared for different mag-
nifications according to different models. Research shows that accuracy is more when
magnification is 100X and 200X since these magnifications capture more area of the re-
gion and extract more details. Results in Table 5 proves this fact that 100X and 200X
magnification images have more accuracy. The highest accuracy of 90.9% is achieved
with 64*64 image size, 180-degree rotation, and 100X test image magnification. As seen
from the results, it can be concluded that even 400X magnification gives better accuracy
as compared to 40X. When low magnification images are used, to obtain more accuracy
large image size is advisable as explained by Table 5. 40X images provide more accuracy
when image size is 128*128 as compared to 64*64 and 32*32. Hence it is concluded that
higher magnification images provide better results in the image prediction with any size
of the image. It is also observed that while augmenting images, 180 and 90 degrees of
rotation must be used for better results in CNN as compared to 60 degrees.
6.2 Experiment 2: Effect of class weight parameter on accuracy
and true negative value in CNN
The fit.generator() function takes an input parameter as train data, valid data, number of
epochs, steps for train and validation and class weight. hen data consists of imbalanced
classes, class weight is used to balance a particular class value. In this research, data is
already balanced in upsampling. Therefor ’auto’ value is used for the class weight. Effect
of class weight = ’auto’ is observed on accuracy and true negative value. The experiment
is carried out on 64*64 image size data with all magnification images in training and
validation. Rotation is set to 60 degrees for training and validation images. As 60 degree
14
17. provided less accuracy in experiment 1, in this experiment trying to increase the accuracy
of CNN for 60-degree rotation. The test accuracy is calculated on different magnifications
with and without class weight as shown in table 6.
Test Mag-
nification
Accuracy with
class weight =
auto
Accuracy
without
class weight
True Neg-
ative with
class weight
True negat-
ive without
class weight
Total ma-
lignant
images
40X 67.5 63.75 103 64 139
100X 89.9 83.7 104 92 144
200X 86.63 85.14 99 100 139
Table 6: Effect of class weight on accuracy and the true negative value
As seen from experiment results, weight class increases the true negative values and
accuracy. It is beneficial for the healthcare domain to identify breast cancer cases from
images. The precision value is calculated using the following formula. It is observed that
for 40X image precision is increased from 45.32% to 74.10%. Accuracy not increased by
major difference as compared to precision.
Precision(Specificity) =
TN
TN + FP
6.3 Experiment 3: Comparison of Accuracy of CNN Vs
Transfer learning with magnification 100X
As we have seen in experiment one, highest accuracy is achieved when magnification
is 100X hence here transfer learning is applied on 100X images. Accuracy from CNN
model and DenseNet-121 is compared on 64*64 image size with 100X magnification on
different image rotations. For DenseNet-121, Imagenet pre-trained weight is used. Both
the models are trained on 12 epochs.
Test Mag-
nification
Rotation Accuracy of
CNN
Accuracy of
DenseNet-121
100X 60 65.62 75.5
100X 90 66.8 68.89
100X 180 68.12 83.25
100X 270 75 79.42
Table 7: Effect of Transfer Learning on Accuracy
Table 7 provides the accuracy of CNN and DenseNet-121 model on 2094 training
images. Transfer learning increased the accuracy of 100X magnification image data and
it is observed that the highest accuracy is obtained when rotation is set to 180 degrees.
Transfer learning performed better on fewer images as compared to the CNN model.
15
18. 6.4 Experiment 4: Effect of different activation functions and
weights in transfer learning
To select the best activation function and pre-trained weight in transfer learning this ex-
periment is carried out. 128*128 image size and 100X magnification images are used for
training, validation, and testing. Sigmoid and Softmax both activation functions are used
and it is observed from results that sigmoid activation gives more accuracy in DenseNet-
121. As mentioned in the literature review imagenet is widely used in all research studies
as a pre-trained weight, this study compares imagenet and
densenet121 weights tf dim ordering tf kernels notop weights for best results. From res-
ults explained in table 8 shows that imagenet weight gives more accuracy as it is trained
on 1 million different classes.
Rotation Weight used Activation
function used
Batch size Accuracy
60 densenet121 softmax 64 41.6
90 densenet121 softmax 64 76.06
180 densenet121 softmax 64 81.33
60 imagenet softmax 64 82.2
90 imagenet softmax 64 82.29
180 imagenet softmax 64 85.64
60 imagenet sigmoid 64 85.64
90 imagenet sigmoid 64 88.03
180 imagenet sigmoid 64 86.12
60 imagenet sigmoid 32 82.2
90 imagenet sigmoid 32 82.77
180 imagenet sigmoid 32 80.38
Table 8: Effect of activation function, batch size and weight on Accuracy
The highest accuracy obtained is 88.03% with 90-degree rotation, imagenet as trained
weight, sigmoid output function, and 64 batch size. Table 8 shows that imagenet gives
better accuracy than dense121 weight. It is observed from Table 8 that the batch size
of 64 is more efficient in DenseNet-121. Softmax and sigmoid function have not much
difference in accuracy but sigmoid performed better in DenseNet-121.
6.5 Experiment 5: Parameter tuning in transfer learning to
achieve the highest accuracy
As evaluated from experiment 4, here Imagenet as a pretrained weight and Sigmoid as
an activation function is used. Effect of different image size and rotations are tested on
accuracy in DenseNet-121.
16
19. image size Rotation Test Accur-
acy of 100X
Test Accur-
acy of 200X
224*224 60 64.11 67.32
224*224 90 85.16 75
224*224 180 72.48 84.65
128*128 60 85.64 74
128*128 90 88.03 71.03
128*128 180 86.12 56.43
64*64 60 76.07 76.73
64*64 90 66.98 72.77
64*64 180 62.9 67.81
Table 9: DenseNet-121 Accuracy for different rotation, magnification and image size
It is observed from Table 9 that the highest accuracy for 100X image is achieved when
image size is 128*128 and rotation is 90 degrees. the highest accuracy for 200X images is
achieved when image size is 224*224 and rotation is 180 degrees. The true positive value
achieved is 117 out of 139 images that is 84.17% precise. A true negative value indicates
the accurate detection of cancer from images.
6.6 Experiment 6: Effect of Magnification and normalisation
on Computation time
To evaluate the performance of CNN and DenseNet-121, computation time for both
models is calculated on normalised and non-normalised images. Effect of different mag-
nifications on computation time is also shown in figure 6.
Figure 5: Effect of Normalisation on computation time
Figure 5 shows that Normalised images have less computation time for both models.
CNN has less computation time as compared to DenseNet-121. All though DenseNet-121
have 121 layers in the model, its computation time is relatively less. This concludes that
transfer learning takes less time for classification as it uses already trained knowledge.
Figure 6 shows that magnification has no major impact on computation time.
17
20. Figure 6: DenseNet-121
6.7 Experiment 7: Effect of image Normalisation on accuracy
and computation time
In this experiment CNN and DenseNet-121 models are executed on normalised and
without normalised images and accuracy is compared. To run the models, 128*128 image
size is used with 180 degrees of rotation and 64 batch size. 10 epochs are executed for
both models with a 0.5 zooming factor.
Test Mag-
nification
Classifier Accuracy
without Nor-
malisation
Accuracy
with Norm-
alisation
40X CNN 66.87 71.25
100X CNN 69.85 70.81
200X CNN 77.72 68.8
400X CNN 58.46 68.3
Mix CNN 65.9 66.9
40X DenseNet-121 49.3 74.3
100X DenseNet-121 48 75.11
200X DenseNet-121 50.99 64.6
400X DenseNet-121 43.16 62.2
Mix DenseNet-121 45.88 65.11
Table 10: Effect of image normalisation on accuracy
For 100X magnification images, the accuracy of DenseNet-121 increased from 48 to
75.11. For 400X magnification images, CNN accuracy increased from 58.56 to 68.3. After
histogram image normalisation. The true negative value achieved for 200X magnification
is 138 out of 139. It gives 99.28% precision for the model.
6.8 Discussion
The main aim of this research study is to increase the accuracy and precision value
for cancer detection. Image normalisation preprocessing increased the average accuracy
of both the models. Its effect majorly seen in the DenseNet-121. Hyper-parameter
18
21. tuning increased the accuracy of the CNN model as well as a transfer learning model.
However, transfer learning model, DenseNet-121 provided the highest accuracy of 88.03,
its accuracy is lower when large image data with mixed magnifications are trained. Figure
7 shows the loss and accuracy of DenseNet-121 on training and validation data on 15
epochs. As seen from the figure, as loss decreases by train data, valid loss also decreases.
Similarly, as accuracy goes on increasing by train data, valid data accuracy also increases.
Figure 8 shows the configuration metrics for 100X image data with 128*128 size using
the CNN model. In this figure 0 indicates benign cases and 1 indicates malignant cases.
A true negative value is 119 from 144 malignant cases and it shows 20 true positive value
from 65 benign cases. It shows that 30% sensitivity and 82.63% specificity(precision). If
the model is trained on a large number of epochs, the model will train more accurately
and it may increase the accuracy and sensitivity. Due to limitations on hardware, this
experiment is performed on a fewer number of epochs.
Figure 7: Loss and Accuracy of DenseNet-121 model while training
As compared to state-of-art accuracy of 94.55% reported in Xiang et al. (2019), this
CNN model achieved less accuracy of 90.9%. The reason behind this may be that the
state-of-art uses Inception-V3 model architecture. CNN architecture of state-of-art is very
complex and hence they could not try fine-tuning on more parameters. This research tried
fine-tuning on many parameters and it is concluded from results that CNN built here best
work with 64*64 image size with dropout value=0.2 and number of transition layers=5.
Transfer learning with fine-tuning achieved 88.03% overall accuracy as seen in experiment
5. DenseNet-121 achieved more accuracy as compared to state-of-art Tao Tan (2018).
Study compared image size, rotations, weights, activation function and several epochs in
the DenseNet-121. Since images were 2165 for 100X magnifications and 2044 for 200X
magnification respectively more accuracy is achieved as more parameters are available to
fine-tune. DenseNet-121 execution is limited to only two weights for pre-trained weights.
CNN trained model could be used as a pre-trained weight in DenseNet-121.
19
22. Figure 8: Confusion Metrics
7 Conclusion and Future Work
This research implemented a deep learning model CNN to find breast cancer from histo-
pathological images. The research found cancerous cases 90.9% accurately. The research
concludes that 100X and 200X magnification images give more accuracy in the diagnostic
study. Hence biopsy results can be improved using 100x and 200X magnification in the
clinical field. Histogram normalization method increased the precision value to 99.28%
and it reduced the execution time. Hence CNN model can be used on histopathological
image data for a quick and accurate finding of the cancer. Histogram normalization,
image augmentation and fine-tuning are utilized during the training process to solve the
overfitting problem. In the future, Stain-normalization can be applied to histopatho-
logical images before classifying. Also, cross-validation can be used in the future for
improving the performance of the CNN network Xiang et al. (2019).
Research implemented transfer learning-based DenseNet-121 model for working on
individual magnification image data. DenseNet-121 achieved 88.3% accuracy after per-
forming fine-tuning which is greater than state-of-art. It is concluded that Imagenet
pretrained weight and Sigmoid output activation function is best suitable for this image
data. In transfer learning, 128*128 images with 90-degree rotation outperformed in terms
of accuracy and 224*224 image outperformed in terms of cancer detection precision. The
use of class weight parameter in the model execution increased the true negative value in
confusion metrics as seen in figure 8. However this experiment did not achieve good true
positive value. Research finds out cancerous cases correctly with 99.28% precision but it
failed to detect non-cancerous cases correctly. This concludes that our model is best suit-
able for detecting cancerous cases than detecting non-cancerous cases. In the future to
20
23. increase the sensitivity of the model more parameters could be tuned. In future, transfer
learning-based various models can be tried on this image data. The number of epochs
and different image enhancement process can try out for the better performance of the
model. To extend this study feature extraction can be performed using transfer-learning.
These extracted features can be further classified using the CNN model.
8 Acknowledgment
I would like to express my sincere gratitude to my supervisor Cristina Muntean for ment-
oring and supporting me in this journey. She guided me in every stage of this tough
project which makes it easy and achievable. I would also like to thank my husband and
parents for believing and allowing me to achieve my goals.
References
Allaouzi, I., Ahmed, M. B. and Benamrou, B. (2019). An Encoder-Decoder model
for visual question answering in the medical domain, CEUR Workshop Proceedings
2380(September 2019): 9–12.
Anghel, A., Stanisavljevic, M., Andani, S., Papandreou, N., Ruschoff, J. H., Wild, P.
and Gabrani, M. (2019). A high-performance system for robust stain normalization of
whole-slide images in histopathology., Frontiers in Medicine .
Azli, E. A., Huddin, A. B., Ibrahim, M. F. and Samad, S. A. (2018). Mass classification
in mammograms using neural network, Proceedings of the 2017 6th International Con-
ference on Electrical Engineering and Informatics: Sustainable Society Through Digital
Innovation, ICEEI 2017 2017-November: 1–5.
d. N´obrega, R. V. M., Peixoto, S. A., d. Silva, S. P. P. and Filho, P. P. R. (2018).
Lung nodule classification via deep transfer learning in ct lung images, 2018 IEEE 31st
International Symposium on Computer-Based Medical Systems (CBMS), pp. 244–249.
Elelimy, E. and Mohamed, A. A. (2018). Towards automatic classification of breast cancer
histopathological image, 2018 13th International Conference on Computer Engineering
and Systems (ICCES), pp. 299–306.
Gottapu, R. D. and Dagli, C. H. (2018). Densenet for anatomical brain segmentation.,
Procedia Computer Science 140(Cyber Physical Systems and Deep Learning Chicago,
Illinois November 5-7, 2018): 179 – 185.
Hidalgo-Gavira, N., Mateos, J., Vega, M., Molina, R. and Katsaggelos, A. K. (2018).
Blind color deconvolution of histopathological images using a variational bayesian
approach, 2018 25th IEEE International Conference on Image Processing (ICIP),
pp. 983–987.
Huang, G., Liu, Z., Van Der Maaten, L. and Weinberger, K. Q. (2017). Densely connected
convolutional networks, Proceedings - 30th IEEE Conference on Computer Vision and
Pattern Recognition, CVPR 2017 2017-January: 2261–2269.
21
24. Jiang, Z., Chen, Y., Zhang, Y., Ge, Y., Yin, F. F. and Ren, L. (2019). Augmentation of
CBCT Reconstructed From Under-Sampled Projections Using Deep Learning, IEEE
transactions on medical imaging 38(11): 2705–2715.
Kaushal, C., Bhat, S., Koundal, D. and Singla, A. (2019). Recent trends in computer
assisted diagnosis (cad) system for breast cancer diagnosis using histopathological im-
ages., IRBM 40(4): 211.
Kaushal, C., Koundal, D. and Singla, A. (2019). Comparative analysis of segmenta-
tion techniques using histopathological images of breast cancer, 2019 3rd International
Conference on Computing Methodologies and Communication (ICCMC), pp. 261–266.
Khened, M., Kollerathu, V. A. and Krishnamurthi, G. (2019). Fully convolutional multi-
scale residual DenseNets for cardiac segmentation and automated cardiac diagnosis
using ensemble of classifiers, Medical Image Analysis 51: 21–45.
Kumar, A., Singh, S. K., Saxena, S., Lakshmanan, K., Sangaiah, A. K., Chauhan, H.,
Shrivastava, S. and Singh, R. K. (2020). Deep feature learning for histopathological
image classification of canine mammary tumors and human breast cancer, Information
Sciences 508: 405–421.
Li, H., Reichert, M., Lin, K., Tselousov, N., Braren, R., Fu, D., Schmid, R., Li, J., Menze,
B. and Shi, K. (2019). Differential diagnosis for pancreatic cysts in ct scans using
densely-connected convolutional networks, 2019 41st Annual International Conference
of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 2095–2098.
Li, H., Zhuang, S., Li, D.-a., Zhao, J. and Ma, Y. (2019). Benign and malignant classi-
fication of mammogram images based on deep learning., Biomedical Signal Processing
and Control 51: 347 – 354.
Li, Y., Xie, X. and Shen, L. (2019). Reverse active learning based atrous DenseNet for
pathological image classification, pp. 1–15.
Litjens, G., Kooi, T., Bejnordi, B. E., Setio, A. A. A., Ciompi, F., Ghafoorian, M.,
van der Laak, J. A., van Ginneken, B. and S´anchez, C. I. (2017). A survey on deep
learning in medical image analysis., Medical Image Analysis 42: 60 – 88.
Oyama, T. and Yamanaka, T. (2017). Fully convolutional densenet for saliency-map
prediction, 2017 4th IAPR Asian Conference on Pattern Recognition (ACPR), pp. 334–
339.
Rundo, L., Tangherloni, A., Nobile, M. S., Militello, C., Besozzi, D., Mauri, G. and
Cazzaniga, P. (2019). Medga: A novel evolutionary method for image enhancement in
medical imaging systems., Expert Systems With Applications 119: 387 – 399.
Sabeena Beevi, K., Nair, M. S. and Bindu, G. (2019). Automatic mitosis detection in
breast histopathology images using convolutional neural network based deep transfer
learning., Biocybernetics and Biomedical Engineering 39(1): 214 – 223.
Talo, M. (2019). Automated classification of histopathology images using transfer learn-
ing, Artificial Intelligence in Medicine 101: 101743.
22
25. Tao Tan, Z. L. (2018). Optimize transfer learning for lung diseases in bronchoscopy using
a new concept: Sequential fine-tuning., IEEE Journal of Translational Engineering in
Health and Medicine, Translational Engineering in Health and Medicine, IEEE Journal
of, IEEE J. Transl. Eng. Health Med p. 1.
Tellez, D., Litjens, G., B´andi, P., Bulten, W., Bokhorst, J.-M., Ciompi, F. and van der
Laak, J. (2019). Quantifying the effects of data augmentation and stain color normal-
ization in convolutional neural networks for computational pathology., Medical Image
Analysis 58.
Tie, J., Jin, J. and Wang, X. (2011). Study on application model of three-tiered ar-
chitecture, 2011 2nd International Conference on Mechanic Automation and Control
Engineering, MACE 2011 - Proceedings pp. 7715–7718.
Vo, D. M., Nguyen, N.-Q. and Lee, S.-W. (2019). Classification of breast cancer his-
tology images using incremental boosting convolution networks, Information Sciences
482: 123 – 138.
Vogado, L. H., Veras, R. M., Araujo, F. H., Silva, R. R. and Aires, K. R. (2018). Leukemia
diagnosis in blood slides using transfer learning in cnns and svm for classification.,
Engineering Applications of Artificial Intelligence 72: 415 – 422.
Xiang, Z., Ting, Z., Weiyan, F. and Cong, L. (2019). Breast cancer diagnosis from
histopathological image based on deep learning, 2019 Chinese Control And Decision
Conference (CCDC), pp. 4616–4619.
Zhuang, L. and Guan, Y. (2017). Image enhancement via subimage histogram equaliza-
tion based on mean and variance., Computational Intelligence and Neuroscience 2017.
23