SlideShare a Scribd company logo
1 of 8
Download to read offline
IAES International Journal of Artificial Intelligence (IJ-AI)
Vol. 14, No. 1, March 2024, pp. 1179~1186
ISSN: 2252-8938, DOI: 10.11591/ijai.v13.i1.pp1179-1186  1179
Journal homepage: http://ijai.iaescore.com
Advanced mask region-based convolutional neural network
based deep-learning model for lung cancer detection
Bhavani Krishna1
, Gopalakrishna Madigondanahalli Thimmaiah2
1
Department of ISE, Dayananda Sagar College of Engineering, Autonomous College Affiliated to Visvesvaraya Technological
University, Belagavi, India
2
Department of AI&ML, SJB Institute of Technology, Affiliated to Visvesvaraya Technological University, Belagavi, India
Article Info ABSTRACT
Article history:
Received Sep 15, 2023
Revised Oct 25, 2023
Accepted Nov 16, 2023
Millions of individuals are affected each year by lung cancer, a serious
global health concern. It may also cause numerous potentially fatal
pulmonary problems, including infections, hemorrhage, or collapse. Finding
a consistent and an effective way to ascertain lung cancer using medical
imaging techniques is one of the primary issues in medical image
processing. The difficulty of this task stems from the fact that the regions of
the lungs that are affected by cancer might differ greatly in expressions of
their size, location, shape, and aesthetics. Identifying whether the identified
area is benign (non-cancerous) or malignant (cancerous) is another difficult
task. Finding the appropriate course of treatment for the patient will depend
on this. A critical stage in the identification of lung malignancy is
identifying the knobs that are expected to be malevolent. To solve these
issues, in this study work we employ a deep learning methodology based on
Mask region-based convolutional neural network (Mask-RCNN). For the
purpose of identifying and locating infected lung regions on computed
tomography (CT) scan images, model is built utilizing the customized Mask-
RCNN. In accordance with the evaluation's findings, the model scored
99.32% for accuracy and 99.45% for mean DICE, respectively.
Keywords:
Average precision
Computed tomography
Lung cancer
Malignant
Mask-RCNN
This is an open access article under the CC BY-SA license.
Corresponding Author:
Bhavani Krishna
Department of ISE, Dayananda Sagar College of Engineering, Autonomous College Affiliated to
Visvesvaraya Technological University
Belagavi, India
Email: bhavanik7@gmail.com
1. INTRODUCTION
Lung cancer is a foremost origin of bereavement amid people with cancer worldwide. The current
status of medical research is challenged by the statistic that cancer is one of the supreme feared diseases in
the world. The amount of cancer cases worldwide has been increasing, and lifestyle factors greatly contribute
to this pattern. Pneumonodular growths, which are tiny cell masses that develop inside the lungs, are
particular of the symptoms of lung cancer. It might be challenging to discriminate amongst pulmonary knobs
that are malignant (cancerous) and benign (non-cancerous). Because they have the prospective to travel to
other regions of the body and intensify, malignant lung nodules must be found and treated as soon as feasible.
Doctors must consider numerous aspects, including the patient's medical history and symptoms, the form,
size, location, and texture of the nodules, in mandate to discriminate amongst malignant and benign lung
nodules. Doctors use a variety of diagnostic techniques to assess pulmonary nodules, including computed
tomography (CT) scans, which produce detailed imageries of the lungs and show the morphology of the
nodules, positron emission tomography (PET) scans, which quantify the metabolic motion of the nodules and
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 1, March 2024: 1179-1186
1180
show in what way they are developing, and needle biopsy, which entails taking a sample of tissue from the
nodule and examining it under a microscope. With the usage of these techniques, it is possible to determine
the most effective course of action by estimating the likelihood of malignancy for each lung nodule
identified.
The enlargement of nodules, which are unnatural cell growths that happen in the lungs, is one
amongst the means that lung tumors can begin. CT scans of the lungs show these nodules. A person can have
single or many nodules in one or both lungs, conditioning on the cause and type of the nodules. Some of the
common causes of lung nodules are respiratory diseases, such as pneumonia, tuberculosis or cancer. The
features of the lung knobs can aid to ascertain the likelihood of them being malignant or cancerous. Some of
these characteristics are the quantity of nodules (single or multiple), shape, and location of the nodules; and
the modifications in the nodules over time. In imperative to distinguish between benign and malignant
nodules, several researchers have been striving to create ways to recognize and categorize lung nodules based
on these characteristics.
For the purpose of identifying lung cancer nodules from CT scans, Venkatesh et al. [1] offer a
unique method that combines Otsu thresholding and the cuckoo search algorithm. The identified nodules are
then classified as either benevolent or malicious using a CNN. They proclaim that their technique
outperforms existing approaches based on element swarm optimization and genetic algorithms by achieving
an accurateness of 96.97%. Feng et al. [2] use a deep learning strategy to distinguish peripheral and non-
peripheral lung cancer imageries using the ResNet50 model and feature pyramid network algorithm. To make
the characteristics less dimensional and boost the model's performance, they apply Fisher coefficients. They
claim accuracy of 97.2% and a dice coefficient of 0.986, which are comparable to state-of-the-art techniques.
Using Mask R-CNN, a well-liked framework for example segmentation, Cai et al. [3] propose a methodology
for lung 3D image identification and segmentation. To aid in investigation and diagnosis, they also employ a
ray-casting volume rendering approach to visualize the segmented lung cancer knobs in 3D space. They
assess their methodology expending two datasets: the Ali TianChi challenge datasets, which are separate
datasets from a lung cancer prediction competition and the LUNA16 dataset, that is an openly accessible
dataset for lung knob research. On these datasets, they attain a sensitivity of 88.1% and 88.7%, respectively.
Additionally, they offer specialized 3D models of the lungs for pulmonary nodule detection that can be
needed to analyze the nodules' spatial distribution and form. Another publicly accessible dataset for
pulmonary image interpretation is the LIDC-IDRI dataset, which Liu et al. [4] utilize to segment lung
nodules. As a support job for nodule segmentation, they use block-based classification learned on the COCO
dataset. By giving the model more data and direction, they oppose that this methodology can upsurge the
segmentation's robustness and accuracy. On the LUNA16 dataset, Kopelowitz et al. [5] use Mask RCNN to
conduct lung nodule detection and segmentation. They assess the nodule size by comparing the expected
volumes to the actual volumes, which is a crucial step in the diagnosis and planning of lung cancer treatment.
They demonstrate that the technique can measure nodule size reliably and offer helpful data for clinical
decision-making. Mask R-CNN is used by Ter-Sarkisov et al. [6] to detect COVID-19 in chest X-ray images.
In mandate to accomplish the segmentation without balancing the dataset, which is a prevalent problem in
medical image analysis owing to statistics scarcity and class imbalance, they adopt COVID-CT-MaskNet's
approach, which is a modified version of Mask R-CNN. They outstrip the currently used approaches for
COVID-19 detection, achieving accuracy of 91.66% and sensitivity of 90.80%. The YOLO-v5 technique is
applied by Liu et al. [7] to identify lung nodules from CT images. The YOLO-v5 object recognition
technique can process photos in real time and is quick and effective. They achieve multi-scale feature fusion
using the BiFPN structure and stochastic-pooling to highlight key characteristics in the feature records, which
can advance recognition performance and accuracy. A dual head network (DHN) with two branches—one for
nodule recognition and one for nodule classification—has been proposed by Tsai et al. [8]. They demonstrate
that, in terms of classification accuracy and recall, DHN can outperform single head networks.
Worldwide, non-communicable diseases like cancer pose serious threats to public health. CT scans
can identify the presence of malignant tissue, which can help save thousands of lives each year only if caught
early and treated. Depending on the type of malignancy identified, the quality of the images used, and the
specific image processing techniques being employed, the efficacy of the many methods for diagnosing
cancer via image processing can vary. Highly consistent prototypes are essential for the speedy, accurate, and
spontaneous diagnosis of lung cancer. The methodology adopted here in this endeavor is based on accurate
recognition and cataloging of lung cancer images which indicates that the nodule and categorization can be
reliably identified. A sophisticated Mask RCNN architecture that displays promising results has been
established to lever these problems using image processing and deep learning approaches. In brief, the ability
to learn defect structures, hyper parameter tuning, and loss function customization are the reasons why the
suggested architecture performed better on the considered dataset. The multi-scale feature fusion further
improved the models performance.
Int J Artif Intell ISSN: 2252-8938 
Advanced mask region-based convolutional neural network based deep … (Bhavani Krishna)
1181
2. MATERIALS AND METHOD
2.1. Materials
An assortment of digital visuals is mentioned as a dataset, and it is utilized by the developers to
develop, test, and evaluate their models and algorithms. In directive to gain insight from the instances in the
dataset, the model is created and trained. In this learning, we utilize the LUNA16 challenge dataset, which
focuses on a thorough analysis of automated nodule detection methods on the LIDC-IDRI dataset, a different
dataset which comprises lung pictures with annotated nodules [9]-[12].
2.2. Method
Object detection is a procedure that comprises computer vision and image processing which aims to
recognize and locate the objects of interest in images and videos [13]-[18]. For example, in medical images,
object detection can be exploited for identifying abnormal cells, fractures, and other issues automatically.
Cancer detection has two steps. First is the the target feature extraction, and second is Classifying and
positioning the objects. Assigning a group of class labels to the items of interest in the input image is the
main phase in the technique of image classification [19]-[20]. Mask R-CNN has numerous aspects that
influence its performance [21]-[25], like the magnitude of the input image, the threshold for intersection over
union (IoU) between predicted and ground truth boxes, and the definite quantity of regions of interest (ROIs)
per image that are proposed by the model. To use Mask R-CNN, the model architecture is built and loaded
with the pre-trained weights from a dataset like COCO or ImageNet. Then, customized images are loaded to
the model for training or inference. The model will then output info approximately on each target that it
detects in the image as exposed in Figure 1 and Figure 2 recapitulates the algorithm.
Figure 1. Advanced Mask RCNN architecture for lung cancer detection
1. Begin
2. Preprocess the Input images
3. Define the backbone network and the Region Proposal Network (RPN)
4. Define the Mask Head, boundary headand the class head
5. Initialize the model using the backbone and the heads
6. Learn the features through training
7. Compute the losses and Predict the mask for each Input image
8. Perform post processing and refine detection if necessary
9. End
Figure 2. Pseudo code of advanced Mask RCNN algorithm
A common choice of backbone linkage is a residual network (ResNet) with a feature pyramid
network (FPN), which creates a hierarchy of feature maps at diverse rules and levels of abstraction. This
allows Mask R-CNN to handle objects of varying sizes and shapes. The overall organization of the Proposed
Architecture used in this line of work can be illustrated by Figure 3. The backbone network is the base
convolution system that excerpts characteristics from the input image. Mask R-CNN uses backbone
networks, such as ResNet101 or ResNet by feature pyramid network (FPN). FPN is a technique that creates a
feature pyramid with multiple levels of features with different resolutions and scales. This helps to detect
objects of different sizes more accurately. The region proposal network (RPN) is a fully convolution network
that takes the structures from the backbone network and generates a group of RoIs that are likely to contain
objects. The RPN uses a sliding window approach over the feature map and applies a 33 convolution
followed by two sibling 11 convolutions: one for forecasting the abjectness score (how likely the region
contains an object) and one for forecasting the bounding rectangle coordinates (the size and locality of the
region). The RPN uses anchor rectangles with diverse rules and characteristic ratios to generate multiple
proposals for each sliding window. In a layer called RoI Align, the extracted features are matched up by the
input RoIs.
Data
preprocessing
Advanced
Mask RCNN
Lung CT
images
Detect lung cancer
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 1, March 2024: 1179-1186
1182
When using RoI Pooling (an earlier method utilized in Faster R-CNN), which quantizes RoIs into
discrete grid cells, it addresses a misalignment issue that arises. The mask branch generates K m x m masks
for each RoI, one for every K classes, where m is the mask resolution (for instance, 2828). The ground
actuality class-specific mask is the only one considered during training because the mask branch only creates
masks for positive RoIs (those that have an IoU overlap with a ground truth box above a threshold).
Figure 3. Overall organization of the proposed advanced Mask RCNN architecture
After applying a softmax function to each pixel of the K masks, one mask is chosen centered on the
projected class to obtain the final result. The proposed architecture components are described in Table 1. The
output mask generation can be depicted as shown in Figure 4.
Table 1. Proposed architecture components
Component Description
Backbone network A basic convolutional network that analyses the input data and extracts features.
Region proposal network A full convolutional communication system that generates regions of interest (RoIs) that are
likely to contain objects.
RoI align A layer that aligns the extracted features with the input RoIs.
Mask representation A branch that works in tandem with the grouping and regression with bounding boxes branches to
forecast binary masks for every RoI.
Figure 4. Ouptut mask generation
Mask R-CNN is for lung cancer recognition using CT scans by adapting the algorithm to the
specific features of the lung images as: i) Use Mask R-CNN to segment the lung region from the CT images
Int J Artif Intell ISSN: 2252-8938 
Advanced mask region-based convolutional neural network based deep … (Bhavani Krishna)
1183
and remove the non-tumor regions. This will reduce the effect of unused regions and improve the accuracy of
tumor segmentation; ii) Apply diverse feature abstraction linkages for Mask R-CNN, such as ResNet101.
This will affect the convergence of the loss values and the extraction of features from the CT images; iii)
Smudge optimizer for Mask R-CNN, such as SGD. This will affect the segmentation accuracy and training
time of the model; iv) Use a customized loss utility that combines both pixel-wise and region-wise losses.
This will balance the trade-off between precision and recall and optimize the detection performance; and v)
Applying image amplification procedures like rotation, scaling, flipping, and cropping aids to upsurge the
diversity and sturdiness of the training data. This will help overcome the challenges of limited and
imbalanced data in the dataset.
3. RESULTS AND DISCUSSION
Mask R-CNN is a state-of-the-art model for target spotting and segmentation that builds on the
Faster R-CNN framework. It adds a parallel branch to the extant ramification for predicting bounding boxes
and class tags that is accountable for spawning a binary concealment for every perceived object. In other
words, Mask R-CNN = Faster R-CNN + FCN, where FCN stands for fully convolutional network, a kind of
neural structure that can process images of any size and output spatial maps of features and predictions. For
this work, Mask R-CNN is further disciplined along a custom dataset with an 80:20 split, meaning that 80%
of the procurable images are exploited in training and the remaining 20% for testing. To determine whether
an anchor box (a predefined box with a certain shape and size) contains an object or not, an intersection over
union (IoU) threshold is used. IoU is a metric that measures the overlap between two regions, such as an
anchor box and a ground truth box. If the IoU is above the threshold, the anchor box is considered as
positive; otherwise, it is negative. Figure 5 shows the AP achieved on the considered dataset. Figure 6 depicts
the DICE value scored on the considered dataset.
Figure 5. AP over images
Figure 6. DICE over 180 images
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 1, March 2024: 1179-1186
1184
The loss function of Mask R-CNN is an aggregation of three losses: grouping loss, bounding
rectangle regression loss and mask loss as given in (1). The classification loss and boundary rectangle
regression loss are the same as in Faster R-CNN. The concealment failure is defined as the intermediate dual
cross-entropy forfeiture over the elements of the predicted mask and the ground actuality concealment for
each region of concern and every class. The whole loss is given by the summation of these three losses, they
are: i) Lcls: The classification loss, same as Faster R-CNN. This is computed using a categorical cross-
entropy loss between the anticipated and true class labels of the regions of interest (RoIs); ii) Lbox: The
boundary rectangle loss, same as Faster R-CNN. This is computed as a smooth L1 loss between the
anticipated and true matches of the boundary boxes; and iii) Lmask: One binary mask with resolution m*m
for every K classes is encoded by the binary mask loss. This is estimated as a binary cross-entropy loss
among the expected and actual pixels in the masks.
L = Lcls + Lbbox + Lmask. (1)
Figure 7 displays the sample detection of nodule with the probability of detection being a nodule.
Figure 7. Nodule detection with probability score
To measure how well an object detection model performs, we need a metric task that takes into
contemplation both the accuracy of the classification and the correctness of the localization. One such metric
is the mean average precision (mAP), that is sustained by the average precision (AP) of each object class. AP
is computed exploiting the intersection over union (IoU) among the expected boundary box and the actual
boundary box. IoU is a magnitude relation that computes how much the two boxes overlap, ranging from 0
(no overlap) to 1 (perfect overlap). A higher IoU means a better localization. To determine whether a
prediction is correct or not, we need to set a threshold for the IoU. For example, if we set the threshold to 0.5,
then any prediction with IoU greater than or equivalent to 0.5 is considered a true positive (TP), meaning that
the model correctly detected and classified the object. Any prediction with IoU less than 0.5 is considered a
false positive (FP), signifying that the model either detected a target that did not exist or wrongly classified it.
Similarly, any ground truth object that the model failed to detect is considered a false negative (FN), meaning
that the model missed an existing object. A true negative (TN) is any portion of the representation where
there is no object and the model did not predict any.
Recall and Precision are cardinal metrics which are derivatives from TP, FP, FN, and TN. Precision
processes as to how many of the predicted objects are correct as in (2), while recall processes to compute
how many of the actual objects are detected as given by (3). They are defined by the pursuing formulas.
Precision = TP / (TP + FP) (2)
Recall = TP / (TP + FN) (3)
To compute AP for a bestowed target class, we plot a precision-recall curve for different IoU
thresholds and calculate the region below the curve (AUC). The higher the AUC, the better the AP.
The formula for F1 score is:
F1=2*((precision*recall)/ precision+recall) (4)
F1 score is a measure that combines precision and recall, which are ratios of true positives, false positives,
and false negatives as specified by (4). It is calculated as the harmonic mean of precision and recall, not the
Int J Artif Intell ISSN: 2252-8938 
Advanced mask region-based convolutional neural network based deep … (Bhavani Krishna)
1185
arithmetic mean2. F1 score is good when both precision and recall are high, and it is bad when either of them
is low. F1 score can be weighted to give different importance to precision and recall. The results achieved by
the implementation of general Mask RCNN and the proposed advanced Mask RCNN specifically for lung
cancer detection is depicted in Table 2.
Table 2. Comparison of general Mask RCNN and proposed advanced Mask RCNN
Metrics General Mask RCNN Proposed advanced Mask RCNN
Precision 63.04 98.77
Recall 54.48 96.62
Accuracy 60.32 99.32
F1 Score 78.73 97.23
Mean Dice Value 57.11 99.55
In conclusion, proposed architecture's higher performance on the Luna 16 dataset can be attributable
to its capacity to learn abnormality features, hyper parameter adjustment, and customized loss function. Its
exceptional performance was further enhanced by the multi-scale feature fusion using skip connections. The
outcomes undeniably demonstrate that the suggested approach outperforms the current cutting-edge
methodologies.
4. CONCLUSION
Mask R-CNN has applications in various fields, such as computer vision, medical imaging,
autonomous driving, and more. Researchers and practitioners keep extending the model for different tasks,
including panoptic segmentation, where instances are segmented and labeled by category. The exponential
proliferation of malignant tissues in the lungs characterizes lung malignancy, a dangerous condition. The
ability to save thousands of lives depends heavily on the primal spotting of lung malignancy. The
effectiveness of the various methods for detecting cancer using image processing can vary relying on the
form of malignancy being found, the standard of images that are utilized, and the particular image processing
procedures being used. The quick, precise, and spontaneous identification of lung cancer needs highly
consistent prototypes. The framework utilized in this effort demonstrates the reliable recognition of the
nodule and classification. The outcomes in this effort are shown using Mask-RCNN by backbone Resnet101.
The values obtained are 99.32% Accuracy and 99.45% for Mean DICE respectively. Collaboration and data
exchange between research organizations and healthcare professionals might speed up the progress of lung
cancer detection. Further future systems for diagnosing lung cancer will be reliant on the integration of
numerous data sources, advancements in deep learning techniques, collaboration, and extensive clinical
validation. While Mask R-CNN performed remarkably well, it wasn't exempt from challenges such as
computational complexity, training data requirements, and the trade-off between speed and accuracy.
ACKNOWLEDGEMENTS
We would like to express thanks to the Lung Image Database Consortium for making the lung
cancer dataset public.
REFERENCES
[1] C. Venaktesh, D. Ramana, S. Y. Lakkisetty, S. S. Band, S. Agarwal, and A. Mosavi, “A neural network and optimization based
lung cancer detection system in CT images,” Frontiers in Public Health, vol. 10, Juni 2022, doi: 10.3389/fpubh.2022.769692.
[2] J. Feng and J. Jiang, “Deep learning-based chest CT image features in diagnosis of lung cancer,” Computational and
Mathematical Methods in Medicine, vol. 2022, Januari 2022, doi: 10.1155/2022/4153211.
[3] L. Cai, T. Long, Y. Dai and Y. Huang, "Mask R-CNN-based detection and segmentation for pulmonary nodule 3D visualization
diagnosis," in IEEE Access, vol. 8, pp. 44400-44409, 2020, doi: 10.1109/ACCESS.2020.2976432.
[4] M. Liu, J. Dong, X. Dong, H. Yu and L. Qi, "Segmentation of lung nodule in ct images based on mask R-CNN," 2018 9th
International Conference on Awareness Science and Technology (iCAST), Fukuoka, Japan, 2018, pp. 1-6, doi:
10.1109/ICAwST.2018.8517248.
[5] E. Kopelowitz and G. Engelhard, “Lung nodules detection and segmentation using 3D mask-RCNN,” Electrical Engineering and
Systems Science > Image and Video Processing, arXiv, preprint arXiv:1907.07676, 2018.
[6] A. Ter-Sarkisov, “Covid-ct-mask-net: prediction of covid-19 from ct scans using regional features,” Applied Intelligence, vol. 52,
no. 9, pp. 9664-9675, Jan 2022, doi: 10.1007%2Fs10489-021-02731-6.
[7] K. Liu, "STBi-YOLO: A Real-Time Object Detection Method for Lung Nodule Recognition," in IEEE Access, vol. 10, pp.
75385-75394, 2022, doi: 10.1109/ACCESS.2022.3192034.
[8] C.-H.Tsai and Y.-S. Peng, “Multi-task lung nodule detection in chest radiographs with a dual head network,” Electrical
Engineering and Systems Science > Image and Video Processing, arXiv, preprint arXiv: 2207.03050, 2022.
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 1, March 2024: 1179-1186
1186
[9] Kh. Aghajani, “Automatic detection of lung nodules on computer tomography scans with a deep direct regression method,”
Journal of AI and Data Mining, vol. 10, no. 2, pp. 207-215, April 2022, doi: 10.22044/jadm.2022.11431.2303.
[10] R. Zhang, C. Cheng, X. Zhao, and X. Li, “Multiscale mask R-CNN-based lung tumor detection using PET imaging,” Molecular
Imaging, July 2019, doi: 10.1177/1536012119863531.
[11] B. K. Hatuwal and H. C. Thapa, "Lung cancer detection using convolutional neural network on histopathological images,"
Internasional Journal Computer Trends and Technology, vol. 68, no. 10, pp. 21-24, 2020, doi: 10.14445/22312803/IJCTT-
V68I10P104.
[12] F. Taher, N. Prakash, A. Shaffie, A. Soliman, and A. El-Baz, “An overview of lung cancer classification algorithms and their
performances,” IAENG International Journal of Computer Science, vol. 48, no. 4, December 2021.
[13] H. Yan, H. Lu, M. Ye, K. Yan, Y. Xu and Q. Jin, "Improved Mask R-CNN for Lung Nodule Segmentation," 2019 10th
International Conference on Information Technology in Medicine and Education (ITME), Qingdao, China, 2019, pp. 137-141,
doi: 10.1109/ITME.2019.00041.
[14] W. Rahane, H. Dalvi, Y. Magar, A. Kalane and S. Jondhale, "Lung Cancer Detection Using Image Processing and Machine
Learning HealthCare," 2018 International Conference on Current Trends towards Converging Technologies (ICCTCT),
Coimbatore, India, 2018, pp. 1-5, doi: 10.1109/ICCTCT.2018.8551008.
[15] K. He, G. Gkioxari, P. Dollár and R. Girshick, "Mask R-CNN," 2017 IEEE International Conference on Computer Vision
(ICCV), Venice, Italy, 2017, pp. 2980-2988, doi: 10.1109/ICCV.2017.322.
[16] K. Bhavani and M. T. Gopalakrishna, “Comparative analysis of traditional classification and deep learning in lung cancer
prediction,” Biomedical Engineering Applications, Basis and Communications, vol. 35, no. 2, December 2022, doi:
10.4015/S101623722250048X.
[17] K. Bhavani and M. T. Gopalakrishna, “Nodule shadow based lung cancer detection using deep learning,” European Chemical
Bulletin, vol. 12, no. 9, pp. 133-152, 2023, doi: 10.48047/ecb/2023.12.9.16.
[18] K. Bhavani and M. T. Gopalakrishna, “Analysis of pre-processing techniques on CT DICOM images,” International Journal of
Computer Sciences and Engineering, vol. 7, no. 6, June 2019, doi: 10.26438/ijcse/v7i6.9598.
[19] H. Huang, Y. Li, R. Wu, Z. Li, Z., and J. Zhang, “Benign-malignant classification of pulmonary nodule with deep feature
optimization framework,” Biomedical Signal Processing and Control, vol. 76, July 2022.
[20] M. A. Mezher, A. Altamimi, and R. Altamimi, “A genetic folding strategy based support vector machine to optimize lung cancer
classification,” Frontiers in Artificial Intelligence, vol. 5, Juni 2022, doi: 10.3389/frai.2022.826374.
[21] M. Humayun, R. Sujatha, S. N. Almuayqil, and N. Z. Jhanjhi, "A transfer learning approach with a convolutional neural network
for the classification of lung carcinoma," Healthcare, vol. 10, no. 6, p. 1058, June 2022, doi: 10.3390/healthcare10061058.
[22] C. Shankara and S. A. Hariprasad, “Artificial neural network for lung cancer detection using CT images,” International Journal of
Health Sciences, vol. 6, no. S2, pp. 2708-2724, 2022, doi: 10.53730/ijhs.v6nS2.5639.
[23] R. Anand, N. A. Rao, and D. S. Sumukh, “Lung cancer detection and prediction using deep learning,” International Journal of
Engineering Applied Sciences and Technology, vol. 7, no. 1, pp. 313-320, May 2022.
[24] G. V. Saji, T. Vazim and S. Sundar, "Deep learning methods for lung cancer detection, classification and prediction-a review,"
2021 Fourth International Conference on Microelectronics, Signals & Systems (ICMSS), Kollam, India, 2021, pp. 1-5, doi:
10.1109/ICMSS53060.2021.9673598.
[25] I. Ali, M. Muzammil, I. U. Haq, A. A. Khaliq, and S. Abdullah, "Efficient lung nodule classification using transferable texture
convolutional neural network," in IEEE Access, vol. 8, pp. 175859-175870, 2020, doi: 10.1109/ACCESS.2020.3026080.
BIOGRAPHIES OF AUTHORS
Bhavani Krishna received Bachelor of Engineering in Computer Science and
Engineering in 2006, Master Degree M.Tech. in Computer Science and Engineering in 2009.
She is working as an Assistant Professor in the Department of Information science and
Engineering, Dayananda Sagar College of Engineering, Bangalore, India since 2009. She is
also currently pursuing Ph.D. from Visvesvaraya Technological University, Karnataka. She
has published many research papers in reputed international journals and conferences. Her
area of research include image processing, Computer Vision, AI. She can be contacted at
email: bhavani-ise@dayanandasagar.edu.
Gopalakrishna Madigondanahalli Thimmaiah received Bachelor of
Engineering in Computer Science & Engineering in 1999 from Bangalore University of
Karnataka, India, the Master Degree M.Tech. in Computer Science and Engineering from
Visvesvaraya Technological University, Karnataka, India in 2004 and PhD in Computer
Science and Engineering from Visvesvaraya Technological University, Karnataka, India in
2014. He is currently working as Professor & Head in the Department of Artificial Inteliigence
& Machine learning, SJB Institute of Technology, Bangalore, India. He has published many
research papers in reputed international journals and conferences. His areas of research
include Computer Vision, Pattern Recognition, Image Processing, Visual Surveillance-Object
Detection/ Classifications/ Tracking/ Identifications. He can be contacted at email:
aimlhod@sjbit.edu.in.

More Related Content

Similar to Advanced mask region-based convolutional neural network based deep-learning model for lung cancer detection

Cancer recurrence prediction using
Cancer recurrence prediction usingCancer recurrence prediction using
Cancer recurrence prediction usingijcsity
 
IRJET- Breast Cancer Detection from Histopathology Images: A Review
IRJET-  	  Breast Cancer Detection from Histopathology Images: A ReviewIRJET-  	  Breast Cancer Detection from Histopathology Images: A Review
IRJET- Breast Cancer Detection from Histopathology Images: A ReviewIRJET Journal
 
X-Ray Disease Identifier
X-Ray Disease IdentifierX-Ray Disease Identifier
X-Ray Disease IdentifierIRJET Journal
 
Paper id 25201472
Paper id 25201472Paper id 25201472
Paper id 25201472IJRAT
 
PNEUMONIA DIAGNOSIS USING CHEST X-RAY IMAGES AND CNN
PNEUMONIA DIAGNOSIS USING CHEST X-RAY IMAGES AND CNNPNEUMONIA DIAGNOSIS USING CHEST X-RAY IMAGES AND CNN
PNEUMONIA DIAGNOSIS USING CHEST X-RAY IMAGES AND CNNIRJET Journal
 
Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...
Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...
Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...IRJET Journal
 
Evaluation of SVM performance in the detection of lung cancer in marked CT s...
Evaluation of SVM performance in the detection of lung cancer  in marked CT s...Evaluation of SVM performance in the detection of lung cancer  in marked CT s...
Evaluation of SVM performance in the detection of lung cancer in marked CT s...nooriasukmaningtyas
 
Deep learning method for lung cancer identification and classification
Deep learning method for lung cancer identification and classificationDeep learning method for lung cancer identification and classification
Deep learning method for lung cancer identification and classificationIAESIJAI
 
Segmentation and Classification of Lung Nodule in Chest Radiograph Image
Segmentation and Classification of Lung Nodule in Chest Radiograph ImageSegmentation and Classification of Lung Nodule in Chest Radiograph Image
Segmentation and Classification of Lung Nodule in Chest Radiograph ImageIJTET Journal
 
Pneumonia Detection System using AI
Pneumonia Detection System using AIPneumonia Detection System using AI
Pneumonia Detection System using AIIRJET Journal
 
YOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model Proposal
YOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model ProposalYOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model Proposal
YOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model ProposalIRJET Journal
 
The Evolution and Impact of Medical Science Journals in Advancing Healthcare
The Evolution and Impact of Medical Science Journals in Advancing HealthcareThe Evolution and Impact of Medical Science Journals in Advancing Healthcare
The Evolution and Impact of Medical Science Journals in Advancing Healthcaresana473753
 
journals on medical
journals on medicaljournals on medical
journals on medicalsana473753
 
covid 19 detection using x ray based on neural network
covid 19 detection using x ray based on neural networkcovid 19 detection using x ray based on neural network
covid 19 detection using x ray based on neural networkArifuzzamanFaisal2
 
Oral Cancer Detection Using Image Processing and Deep Neural Networks.
Oral Cancer Detection Using Image Processing and Deep Neural Networks.Oral Cancer Detection Using Image Processing and Deep Neural Networks.
Oral Cancer Detection Using Image Processing and Deep Neural Networks.IRJET Journal
 
Computer-aided diagnostic system kinds and pulmonary nodule detection efficacy
Computer-aided diagnostic system kinds and pulmonary nodule  detection efficacyComputer-aided diagnostic system kinds and pulmonary nodule  detection efficacy
Computer-aided diagnostic system kinds and pulmonary nodule detection efficacyIJECEIAES
 
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...semualkaira
 
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...semualkaira
 
ARTIFICIAL INTELLIGENCE BASED COVID-19 DETECTION USING COMPUTED TOMOGRAPHY IM...
ARTIFICIAL INTELLIGENCE BASED COVID-19 DETECTION USING COMPUTED TOMOGRAPHY IM...ARTIFICIAL INTELLIGENCE BASED COVID-19 DETECTION USING COMPUTED TOMOGRAPHY IM...
ARTIFICIAL INTELLIGENCE BASED COVID-19 DETECTION USING COMPUTED TOMOGRAPHY IM...IRJET Journal
 
Cancer recognition from dna microarray gene expression data using averaged on...
Cancer recognition from dna microarray gene expression data using averaged on...Cancer recognition from dna microarray gene expression data using averaged on...
Cancer recognition from dna microarray gene expression data using averaged on...IJCI JOURNAL
 

Similar to Advanced mask region-based convolutional neural network based deep-learning model for lung cancer detection (20)

Cancer recurrence prediction using
Cancer recurrence prediction usingCancer recurrence prediction using
Cancer recurrence prediction using
 
IRJET- Breast Cancer Detection from Histopathology Images: A Review
IRJET-  	  Breast Cancer Detection from Histopathology Images: A ReviewIRJET-  	  Breast Cancer Detection from Histopathology Images: A Review
IRJET- Breast Cancer Detection from Histopathology Images: A Review
 
X-Ray Disease Identifier
X-Ray Disease IdentifierX-Ray Disease Identifier
X-Ray Disease Identifier
 
Paper id 25201472
Paper id 25201472Paper id 25201472
Paper id 25201472
 
PNEUMONIA DIAGNOSIS USING CHEST X-RAY IMAGES AND CNN
PNEUMONIA DIAGNOSIS USING CHEST X-RAY IMAGES AND CNNPNEUMONIA DIAGNOSIS USING CHEST X-RAY IMAGES AND CNN
PNEUMONIA DIAGNOSIS USING CHEST X-RAY IMAGES AND CNN
 
Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...
Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...
Lung Nodule Feature Extraction and Classification using Improved Neural Netwo...
 
Evaluation of SVM performance in the detection of lung cancer in marked CT s...
Evaluation of SVM performance in the detection of lung cancer  in marked CT s...Evaluation of SVM performance in the detection of lung cancer  in marked CT s...
Evaluation of SVM performance in the detection of lung cancer in marked CT s...
 
Deep learning method for lung cancer identification and classification
Deep learning method for lung cancer identification and classificationDeep learning method for lung cancer identification and classification
Deep learning method for lung cancer identification and classification
 
Segmentation and Classification of Lung Nodule in Chest Radiograph Image
Segmentation and Classification of Lung Nodule in Chest Radiograph ImageSegmentation and Classification of Lung Nodule in Chest Radiograph Image
Segmentation and Classification of Lung Nodule in Chest Radiograph Image
 
Pneumonia Detection System using AI
Pneumonia Detection System using AIPneumonia Detection System using AI
Pneumonia Detection System using AI
 
YOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model Proposal
YOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model ProposalYOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model Proposal
YOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model Proposal
 
The Evolution and Impact of Medical Science Journals in Advancing Healthcare
The Evolution and Impact of Medical Science Journals in Advancing HealthcareThe Evolution and Impact of Medical Science Journals in Advancing Healthcare
The Evolution and Impact of Medical Science Journals in Advancing Healthcare
 
journals on medical
journals on medicaljournals on medical
journals on medical
 
covid 19 detection using x ray based on neural network
covid 19 detection using x ray based on neural networkcovid 19 detection using x ray based on neural network
covid 19 detection using x ray based on neural network
 
Oral Cancer Detection Using Image Processing and Deep Neural Networks.
Oral Cancer Detection Using Image Processing and Deep Neural Networks.Oral Cancer Detection Using Image Processing and Deep Neural Networks.
Oral Cancer Detection Using Image Processing and Deep Neural Networks.
 
Computer-aided diagnostic system kinds and pulmonary nodule detection efficacy
Computer-aided diagnostic system kinds and pulmonary nodule  detection efficacyComputer-aided diagnostic system kinds and pulmonary nodule  detection efficacy
Computer-aided diagnostic system kinds and pulmonary nodule detection efficacy
 
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
 
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
The Clinical use of Artificial Intelligence in the Analysis of Chest Radiogra...
 
ARTIFICIAL INTELLIGENCE BASED COVID-19 DETECTION USING COMPUTED TOMOGRAPHY IM...
ARTIFICIAL INTELLIGENCE BASED COVID-19 DETECTION USING COMPUTED TOMOGRAPHY IM...ARTIFICIAL INTELLIGENCE BASED COVID-19 DETECTION USING COMPUTED TOMOGRAPHY IM...
ARTIFICIAL INTELLIGENCE BASED COVID-19 DETECTION USING COMPUTED TOMOGRAPHY IM...
 
Cancer recognition from dna microarray gene expression data using averaged on...
Cancer recognition from dna microarray gene expression data using averaged on...Cancer recognition from dna microarray gene expression data using averaged on...
Cancer recognition from dna microarray gene expression data using averaged on...
 

More from IAESIJAI

Convolutional neural network with binary moth flame optimization for emotion ...
Convolutional neural network with binary moth flame optimization for emotion ...Convolutional neural network with binary moth flame optimization for emotion ...
Convolutional neural network with binary moth flame optimization for emotion ...IAESIJAI
 
A novel ensemble model for detecting fake news
A novel ensemble model for detecting fake newsA novel ensemble model for detecting fake news
A novel ensemble model for detecting fake newsIAESIJAI
 
K-centroid convergence clustering identification in one-label per type for di...
K-centroid convergence clustering identification in one-label per type for di...K-centroid convergence clustering identification in one-label per type for di...
K-centroid convergence clustering identification in one-label per type for di...IAESIJAI
 
Plant leaf detection through machine learning based image classification appr...
Plant leaf detection through machine learning based image classification appr...Plant leaf detection through machine learning based image classification appr...
Plant leaf detection through machine learning based image classification appr...IAESIJAI
 
Backbone search for object detection for applications in intrusion warning sy...
Backbone search for object detection for applications in intrusion warning sy...Backbone search for object detection for applications in intrusion warning sy...
Backbone search for object detection for applications in intrusion warning sy...IAESIJAI
 
Optically processed Kannada script realization with Siamese neural network model
Optically processed Kannada script realization with Siamese neural network modelOptically processed Kannada script realization with Siamese neural network model
Optically processed Kannada script realization with Siamese neural network modelIAESIJAI
 
Embedded artificial intelligence system using deep learning and raspberrypi f...
Embedded artificial intelligence system using deep learning and raspberrypi f...Embedded artificial intelligence system using deep learning and raspberrypi f...
Embedded artificial intelligence system using deep learning and raspberrypi f...IAESIJAI
 
Deep learning based biometric authentication using electrocardiogram and iris
Deep learning based biometric authentication using electrocardiogram and irisDeep learning based biometric authentication using electrocardiogram and iris
Deep learning based biometric authentication using electrocardiogram and irisIAESIJAI
 
Hybrid channel and spatial attention-UNet for skin lesion segmentation
Hybrid channel and spatial attention-UNet for skin lesion segmentationHybrid channel and spatial attention-UNet for skin lesion segmentation
Hybrid channel and spatial attention-UNet for skin lesion segmentationIAESIJAI
 
Photoplethysmogram signal reconstruction through integrated compression sensi...
Photoplethysmogram signal reconstruction through integrated compression sensi...Photoplethysmogram signal reconstruction through integrated compression sensi...
Photoplethysmogram signal reconstruction through integrated compression sensi...IAESIJAI
 
Speaker identification under noisy conditions using hybrid convolutional neur...
Speaker identification under noisy conditions using hybrid convolutional neur...Speaker identification under noisy conditions using hybrid convolutional neur...
Speaker identification under noisy conditions using hybrid convolutional neur...IAESIJAI
 
Multi-channel microseismic signals classification with convolutional neural n...
Multi-channel microseismic signals classification with convolutional neural n...Multi-channel microseismic signals classification with convolutional neural n...
Multi-channel microseismic signals classification with convolutional neural n...IAESIJAI
 
Sophisticated face mask dataset: a novel dataset for effective coronavirus di...
Sophisticated face mask dataset: a novel dataset for effective coronavirus di...Sophisticated face mask dataset: a novel dataset for effective coronavirus di...
Sophisticated face mask dataset: a novel dataset for effective coronavirus di...IAESIJAI
 
Transfer learning for epilepsy detection using spectrogram images
Transfer learning for epilepsy detection using spectrogram imagesTransfer learning for epilepsy detection using spectrogram images
Transfer learning for epilepsy detection using spectrogram imagesIAESIJAI
 
Deep neural network for lateral control of self-driving cars in urban environ...
Deep neural network for lateral control of self-driving cars in urban environ...Deep neural network for lateral control of self-driving cars in urban environ...
Deep neural network for lateral control of self-driving cars in urban environ...IAESIJAI
 
Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...
Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...
Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...IAESIJAI
 
Efficient commodity price forecasting using long short-term memory model
Efficient commodity price forecasting using long short-term memory modelEfficient commodity price forecasting using long short-term memory model
Efficient commodity price forecasting using long short-term memory modelIAESIJAI
 
1-dimensional convolutional neural networks for predicting sudden cardiac
1-dimensional convolutional neural networks for predicting sudden cardiac1-dimensional convolutional neural networks for predicting sudden cardiac
1-dimensional convolutional neural networks for predicting sudden cardiacIAESIJAI
 
A deep learning-based approach for early detection of disease in sugarcane pl...
A deep learning-based approach for early detection of disease in sugarcane pl...A deep learning-based approach for early detection of disease in sugarcane pl...
A deep learning-based approach for early detection of disease in sugarcane pl...IAESIJAI
 
Signature verification based on proposed fast hyper deep neural network
Signature verification based on proposed fast hyper deep neural networkSignature verification based on proposed fast hyper deep neural network
Signature verification based on proposed fast hyper deep neural networkIAESIJAI
 

More from IAESIJAI (20)

Convolutional neural network with binary moth flame optimization for emotion ...
Convolutional neural network with binary moth flame optimization for emotion ...Convolutional neural network with binary moth flame optimization for emotion ...
Convolutional neural network with binary moth flame optimization for emotion ...
 
A novel ensemble model for detecting fake news
A novel ensemble model for detecting fake newsA novel ensemble model for detecting fake news
A novel ensemble model for detecting fake news
 
K-centroid convergence clustering identification in one-label per type for di...
K-centroid convergence clustering identification in one-label per type for di...K-centroid convergence clustering identification in one-label per type for di...
K-centroid convergence clustering identification in one-label per type for di...
 
Plant leaf detection through machine learning based image classification appr...
Plant leaf detection through machine learning based image classification appr...Plant leaf detection through machine learning based image classification appr...
Plant leaf detection through machine learning based image classification appr...
 
Backbone search for object detection for applications in intrusion warning sy...
Backbone search for object detection for applications in intrusion warning sy...Backbone search for object detection for applications in intrusion warning sy...
Backbone search for object detection for applications in intrusion warning sy...
 
Optically processed Kannada script realization with Siamese neural network model
Optically processed Kannada script realization with Siamese neural network modelOptically processed Kannada script realization with Siamese neural network model
Optically processed Kannada script realization with Siamese neural network model
 
Embedded artificial intelligence system using deep learning and raspberrypi f...
Embedded artificial intelligence system using deep learning and raspberrypi f...Embedded artificial intelligence system using deep learning and raspberrypi f...
Embedded artificial intelligence system using deep learning and raspberrypi f...
 
Deep learning based biometric authentication using electrocardiogram and iris
Deep learning based biometric authentication using electrocardiogram and irisDeep learning based biometric authentication using electrocardiogram and iris
Deep learning based biometric authentication using electrocardiogram and iris
 
Hybrid channel and spatial attention-UNet for skin lesion segmentation
Hybrid channel and spatial attention-UNet for skin lesion segmentationHybrid channel and spatial attention-UNet for skin lesion segmentation
Hybrid channel and spatial attention-UNet for skin lesion segmentation
 
Photoplethysmogram signal reconstruction through integrated compression sensi...
Photoplethysmogram signal reconstruction through integrated compression sensi...Photoplethysmogram signal reconstruction through integrated compression sensi...
Photoplethysmogram signal reconstruction through integrated compression sensi...
 
Speaker identification under noisy conditions using hybrid convolutional neur...
Speaker identification under noisy conditions using hybrid convolutional neur...Speaker identification under noisy conditions using hybrid convolutional neur...
Speaker identification under noisy conditions using hybrid convolutional neur...
 
Multi-channel microseismic signals classification with convolutional neural n...
Multi-channel microseismic signals classification with convolutional neural n...Multi-channel microseismic signals classification with convolutional neural n...
Multi-channel microseismic signals classification with convolutional neural n...
 
Sophisticated face mask dataset: a novel dataset for effective coronavirus di...
Sophisticated face mask dataset: a novel dataset for effective coronavirus di...Sophisticated face mask dataset: a novel dataset for effective coronavirus di...
Sophisticated face mask dataset: a novel dataset for effective coronavirus di...
 
Transfer learning for epilepsy detection using spectrogram images
Transfer learning for epilepsy detection using spectrogram imagesTransfer learning for epilepsy detection using spectrogram images
Transfer learning for epilepsy detection using spectrogram images
 
Deep neural network for lateral control of self-driving cars in urban environ...
Deep neural network for lateral control of self-driving cars in urban environ...Deep neural network for lateral control of self-driving cars in urban environ...
Deep neural network for lateral control of self-driving cars in urban environ...
 
Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...
Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...
Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...
 
Efficient commodity price forecasting using long short-term memory model
Efficient commodity price forecasting using long short-term memory modelEfficient commodity price forecasting using long short-term memory model
Efficient commodity price forecasting using long short-term memory model
 
1-dimensional convolutional neural networks for predicting sudden cardiac
1-dimensional convolutional neural networks for predicting sudden cardiac1-dimensional convolutional neural networks for predicting sudden cardiac
1-dimensional convolutional neural networks for predicting sudden cardiac
 
A deep learning-based approach for early detection of disease in sugarcane pl...
A deep learning-based approach for early detection of disease in sugarcane pl...A deep learning-based approach for early detection of disease in sugarcane pl...
A deep learning-based approach for early detection of disease in sugarcane pl...
 
Signature verification based on proposed fast hyper deep neural network
Signature verification based on proposed fast hyper deep neural networkSignature verification based on proposed fast hyper deep neural network
Signature verification based on proposed fast hyper deep neural network
 

Recently uploaded

Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDGMarianaLemus7
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
costume and set research powerpoint presentation
costume and set research powerpoint presentationcostume and set research powerpoint presentation
costume and set research powerpoint presentationphoebematthew05
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Wonjun Hwang
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsHyundai Motor Group
 
Pigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
Unlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsUnlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsPrecisely
 
Bluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfBluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfngoud9212
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):comworks
 
Artificial intelligence in the post-deep learning era
Artificial intelligence in the post-deep learning eraArtificial intelligence in the post-deep learning era
Artificial intelligence in the post-deep learning eraDeakin University
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 

Recently uploaded (20)

Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDG
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
costume and set research powerpoint presentation
costume and set research powerpoint presentationcostume and set research powerpoint presentation
costume and set research powerpoint presentation
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
 
Vulnerability_Management_GRC_by Sohang Sengupta.pptx
Vulnerability_Management_GRC_by Sohang Sengupta.pptxVulnerability_Management_GRC_by Sohang Sengupta.pptx
Vulnerability_Management_GRC_by Sohang Sengupta.pptx
 
Pigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping Elbows
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
Unlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsUnlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power Systems
 
Bluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfBluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdf
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):
 
Artificial intelligence in the post-deep learning era
Artificial intelligence in the post-deep learning eraArtificial intelligence in the post-deep learning era
Artificial intelligence in the post-deep learning era
 
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort ServiceHot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 

Advanced mask region-based convolutional neural network based deep-learning model for lung cancer detection

  • 1. IAES International Journal of Artificial Intelligence (IJ-AI) Vol. 14, No. 1, March 2024, pp. 1179~1186 ISSN: 2252-8938, DOI: 10.11591/ijai.v13.i1.pp1179-1186  1179 Journal homepage: http://ijai.iaescore.com Advanced mask region-based convolutional neural network based deep-learning model for lung cancer detection Bhavani Krishna1 , Gopalakrishna Madigondanahalli Thimmaiah2 1 Department of ISE, Dayananda Sagar College of Engineering, Autonomous College Affiliated to Visvesvaraya Technological University, Belagavi, India 2 Department of AI&ML, SJB Institute of Technology, Affiliated to Visvesvaraya Technological University, Belagavi, India Article Info ABSTRACT Article history: Received Sep 15, 2023 Revised Oct 25, 2023 Accepted Nov 16, 2023 Millions of individuals are affected each year by lung cancer, a serious global health concern. It may also cause numerous potentially fatal pulmonary problems, including infections, hemorrhage, or collapse. Finding a consistent and an effective way to ascertain lung cancer using medical imaging techniques is one of the primary issues in medical image processing. The difficulty of this task stems from the fact that the regions of the lungs that are affected by cancer might differ greatly in expressions of their size, location, shape, and aesthetics. Identifying whether the identified area is benign (non-cancerous) or malignant (cancerous) is another difficult task. Finding the appropriate course of treatment for the patient will depend on this. A critical stage in the identification of lung malignancy is identifying the knobs that are expected to be malevolent. To solve these issues, in this study work we employ a deep learning methodology based on Mask region-based convolutional neural network (Mask-RCNN). For the purpose of identifying and locating infected lung regions on computed tomography (CT) scan images, model is built utilizing the customized Mask- RCNN. In accordance with the evaluation's findings, the model scored 99.32% for accuracy and 99.45% for mean DICE, respectively. Keywords: Average precision Computed tomography Lung cancer Malignant Mask-RCNN This is an open access article under the CC BY-SA license. Corresponding Author: Bhavani Krishna Department of ISE, Dayananda Sagar College of Engineering, Autonomous College Affiliated to Visvesvaraya Technological University Belagavi, India Email: bhavanik7@gmail.com 1. INTRODUCTION Lung cancer is a foremost origin of bereavement amid people with cancer worldwide. The current status of medical research is challenged by the statistic that cancer is one of the supreme feared diseases in the world. The amount of cancer cases worldwide has been increasing, and lifestyle factors greatly contribute to this pattern. Pneumonodular growths, which are tiny cell masses that develop inside the lungs, are particular of the symptoms of lung cancer. It might be challenging to discriminate amongst pulmonary knobs that are malignant (cancerous) and benign (non-cancerous). Because they have the prospective to travel to other regions of the body and intensify, malignant lung nodules must be found and treated as soon as feasible. Doctors must consider numerous aspects, including the patient's medical history and symptoms, the form, size, location, and texture of the nodules, in mandate to discriminate amongst malignant and benign lung nodules. Doctors use a variety of diagnostic techniques to assess pulmonary nodules, including computed tomography (CT) scans, which produce detailed imageries of the lungs and show the morphology of the nodules, positron emission tomography (PET) scans, which quantify the metabolic motion of the nodules and
  • 2.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 1, March 2024: 1179-1186 1180 show in what way they are developing, and needle biopsy, which entails taking a sample of tissue from the nodule and examining it under a microscope. With the usage of these techniques, it is possible to determine the most effective course of action by estimating the likelihood of malignancy for each lung nodule identified. The enlargement of nodules, which are unnatural cell growths that happen in the lungs, is one amongst the means that lung tumors can begin. CT scans of the lungs show these nodules. A person can have single or many nodules in one or both lungs, conditioning on the cause and type of the nodules. Some of the common causes of lung nodules are respiratory diseases, such as pneumonia, tuberculosis or cancer. The features of the lung knobs can aid to ascertain the likelihood of them being malignant or cancerous. Some of these characteristics are the quantity of nodules (single or multiple), shape, and location of the nodules; and the modifications in the nodules over time. In imperative to distinguish between benign and malignant nodules, several researchers have been striving to create ways to recognize and categorize lung nodules based on these characteristics. For the purpose of identifying lung cancer nodules from CT scans, Venkatesh et al. [1] offer a unique method that combines Otsu thresholding and the cuckoo search algorithm. The identified nodules are then classified as either benevolent or malicious using a CNN. They proclaim that their technique outperforms existing approaches based on element swarm optimization and genetic algorithms by achieving an accurateness of 96.97%. Feng et al. [2] use a deep learning strategy to distinguish peripheral and non- peripheral lung cancer imageries using the ResNet50 model and feature pyramid network algorithm. To make the characteristics less dimensional and boost the model's performance, they apply Fisher coefficients. They claim accuracy of 97.2% and a dice coefficient of 0.986, which are comparable to state-of-the-art techniques. Using Mask R-CNN, a well-liked framework for example segmentation, Cai et al. [3] propose a methodology for lung 3D image identification and segmentation. To aid in investigation and diagnosis, they also employ a ray-casting volume rendering approach to visualize the segmented lung cancer knobs in 3D space. They assess their methodology expending two datasets: the Ali TianChi challenge datasets, which are separate datasets from a lung cancer prediction competition and the LUNA16 dataset, that is an openly accessible dataset for lung knob research. On these datasets, they attain a sensitivity of 88.1% and 88.7%, respectively. Additionally, they offer specialized 3D models of the lungs for pulmonary nodule detection that can be needed to analyze the nodules' spatial distribution and form. Another publicly accessible dataset for pulmonary image interpretation is the LIDC-IDRI dataset, which Liu et al. [4] utilize to segment lung nodules. As a support job for nodule segmentation, they use block-based classification learned on the COCO dataset. By giving the model more data and direction, they oppose that this methodology can upsurge the segmentation's robustness and accuracy. On the LUNA16 dataset, Kopelowitz et al. [5] use Mask RCNN to conduct lung nodule detection and segmentation. They assess the nodule size by comparing the expected volumes to the actual volumes, which is a crucial step in the diagnosis and planning of lung cancer treatment. They demonstrate that the technique can measure nodule size reliably and offer helpful data for clinical decision-making. Mask R-CNN is used by Ter-Sarkisov et al. [6] to detect COVID-19 in chest X-ray images. In mandate to accomplish the segmentation without balancing the dataset, which is a prevalent problem in medical image analysis owing to statistics scarcity and class imbalance, they adopt COVID-CT-MaskNet's approach, which is a modified version of Mask R-CNN. They outstrip the currently used approaches for COVID-19 detection, achieving accuracy of 91.66% and sensitivity of 90.80%. The YOLO-v5 technique is applied by Liu et al. [7] to identify lung nodules from CT images. The YOLO-v5 object recognition technique can process photos in real time and is quick and effective. They achieve multi-scale feature fusion using the BiFPN structure and stochastic-pooling to highlight key characteristics in the feature records, which can advance recognition performance and accuracy. A dual head network (DHN) with two branches—one for nodule recognition and one for nodule classification—has been proposed by Tsai et al. [8]. They demonstrate that, in terms of classification accuracy and recall, DHN can outperform single head networks. Worldwide, non-communicable diseases like cancer pose serious threats to public health. CT scans can identify the presence of malignant tissue, which can help save thousands of lives each year only if caught early and treated. Depending on the type of malignancy identified, the quality of the images used, and the specific image processing techniques being employed, the efficacy of the many methods for diagnosing cancer via image processing can vary. Highly consistent prototypes are essential for the speedy, accurate, and spontaneous diagnosis of lung cancer. The methodology adopted here in this endeavor is based on accurate recognition and cataloging of lung cancer images which indicates that the nodule and categorization can be reliably identified. A sophisticated Mask RCNN architecture that displays promising results has been established to lever these problems using image processing and deep learning approaches. In brief, the ability to learn defect structures, hyper parameter tuning, and loss function customization are the reasons why the suggested architecture performed better on the considered dataset. The multi-scale feature fusion further improved the models performance.
  • 3. Int J Artif Intell ISSN: 2252-8938  Advanced mask region-based convolutional neural network based deep … (Bhavani Krishna) 1181 2. MATERIALS AND METHOD 2.1. Materials An assortment of digital visuals is mentioned as a dataset, and it is utilized by the developers to develop, test, and evaluate their models and algorithms. In directive to gain insight from the instances in the dataset, the model is created and trained. In this learning, we utilize the LUNA16 challenge dataset, which focuses on a thorough analysis of automated nodule detection methods on the LIDC-IDRI dataset, a different dataset which comprises lung pictures with annotated nodules [9]-[12]. 2.2. Method Object detection is a procedure that comprises computer vision and image processing which aims to recognize and locate the objects of interest in images and videos [13]-[18]. For example, in medical images, object detection can be exploited for identifying abnormal cells, fractures, and other issues automatically. Cancer detection has two steps. First is the the target feature extraction, and second is Classifying and positioning the objects. Assigning a group of class labels to the items of interest in the input image is the main phase in the technique of image classification [19]-[20]. Mask R-CNN has numerous aspects that influence its performance [21]-[25], like the magnitude of the input image, the threshold for intersection over union (IoU) between predicted and ground truth boxes, and the definite quantity of regions of interest (ROIs) per image that are proposed by the model. To use Mask R-CNN, the model architecture is built and loaded with the pre-trained weights from a dataset like COCO or ImageNet. Then, customized images are loaded to the model for training or inference. The model will then output info approximately on each target that it detects in the image as exposed in Figure 1 and Figure 2 recapitulates the algorithm. Figure 1. Advanced Mask RCNN architecture for lung cancer detection 1. Begin 2. Preprocess the Input images 3. Define the backbone network and the Region Proposal Network (RPN) 4. Define the Mask Head, boundary headand the class head 5. Initialize the model using the backbone and the heads 6. Learn the features through training 7. Compute the losses and Predict the mask for each Input image 8. Perform post processing and refine detection if necessary 9. End Figure 2. Pseudo code of advanced Mask RCNN algorithm A common choice of backbone linkage is a residual network (ResNet) with a feature pyramid network (FPN), which creates a hierarchy of feature maps at diverse rules and levels of abstraction. This allows Mask R-CNN to handle objects of varying sizes and shapes. The overall organization of the Proposed Architecture used in this line of work can be illustrated by Figure 3. The backbone network is the base convolution system that excerpts characteristics from the input image. Mask R-CNN uses backbone networks, such as ResNet101 or ResNet by feature pyramid network (FPN). FPN is a technique that creates a feature pyramid with multiple levels of features with different resolutions and scales. This helps to detect objects of different sizes more accurately. The region proposal network (RPN) is a fully convolution network that takes the structures from the backbone network and generates a group of RoIs that are likely to contain objects. The RPN uses a sliding window approach over the feature map and applies a 33 convolution followed by two sibling 11 convolutions: one for forecasting the abjectness score (how likely the region contains an object) and one for forecasting the bounding rectangle coordinates (the size and locality of the region). The RPN uses anchor rectangles with diverse rules and characteristic ratios to generate multiple proposals for each sliding window. In a layer called RoI Align, the extracted features are matched up by the input RoIs. Data preprocessing Advanced Mask RCNN Lung CT images Detect lung cancer
  • 4.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 1, March 2024: 1179-1186 1182 When using RoI Pooling (an earlier method utilized in Faster R-CNN), which quantizes RoIs into discrete grid cells, it addresses a misalignment issue that arises. The mask branch generates K m x m masks for each RoI, one for every K classes, where m is the mask resolution (for instance, 2828). The ground actuality class-specific mask is the only one considered during training because the mask branch only creates masks for positive RoIs (those that have an IoU overlap with a ground truth box above a threshold). Figure 3. Overall organization of the proposed advanced Mask RCNN architecture After applying a softmax function to each pixel of the K masks, one mask is chosen centered on the projected class to obtain the final result. The proposed architecture components are described in Table 1. The output mask generation can be depicted as shown in Figure 4. Table 1. Proposed architecture components Component Description Backbone network A basic convolutional network that analyses the input data and extracts features. Region proposal network A full convolutional communication system that generates regions of interest (RoIs) that are likely to contain objects. RoI align A layer that aligns the extracted features with the input RoIs. Mask representation A branch that works in tandem with the grouping and regression with bounding boxes branches to forecast binary masks for every RoI. Figure 4. Ouptut mask generation Mask R-CNN is for lung cancer recognition using CT scans by adapting the algorithm to the specific features of the lung images as: i) Use Mask R-CNN to segment the lung region from the CT images
  • 5. Int J Artif Intell ISSN: 2252-8938  Advanced mask region-based convolutional neural network based deep … (Bhavani Krishna) 1183 and remove the non-tumor regions. This will reduce the effect of unused regions and improve the accuracy of tumor segmentation; ii) Apply diverse feature abstraction linkages for Mask R-CNN, such as ResNet101. This will affect the convergence of the loss values and the extraction of features from the CT images; iii) Smudge optimizer for Mask R-CNN, such as SGD. This will affect the segmentation accuracy and training time of the model; iv) Use a customized loss utility that combines both pixel-wise and region-wise losses. This will balance the trade-off between precision and recall and optimize the detection performance; and v) Applying image amplification procedures like rotation, scaling, flipping, and cropping aids to upsurge the diversity and sturdiness of the training data. This will help overcome the challenges of limited and imbalanced data in the dataset. 3. RESULTS AND DISCUSSION Mask R-CNN is a state-of-the-art model for target spotting and segmentation that builds on the Faster R-CNN framework. It adds a parallel branch to the extant ramification for predicting bounding boxes and class tags that is accountable for spawning a binary concealment for every perceived object. In other words, Mask R-CNN = Faster R-CNN + FCN, where FCN stands for fully convolutional network, a kind of neural structure that can process images of any size and output spatial maps of features and predictions. For this work, Mask R-CNN is further disciplined along a custom dataset with an 80:20 split, meaning that 80% of the procurable images are exploited in training and the remaining 20% for testing. To determine whether an anchor box (a predefined box with a certain shape and size) contains an object or not, an intersection over union (IoU) threshold is used. IoU is a metric that measures the overlap between two regions, such as an anchor box and a ground truth box. If the IoU is above the threshold, the anchor box is considered as positive; otherwise, it is negative. Figure 5 shows the AP achieved on the considered dataset. Figure 6 depicts the DICE value scored on the considered dataset. Figure 5. AP over images Figure 6. DICE over 180 images
  • 6.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 1, March 2024: 1179-1186 1184 The loss function of Mask R-CNN is an aggregation of three losses: grouping loss, bounding rectangle regression loss and mask loss as given in (1). The classification loss and boundary rectangle regression loss are the same as in Faster R-CNN. The concealment failure is defined as the intermediate dual cross-entropy forfeiture over the elements of the predicted mask and the ground actuality concealment for each region of concern and every class. The whole loss is given by the summation of these three losses, they are: i) Lcls: The classification loss, same as Faster R-CNN. This is computed using a categorical cross- entropy loss between the anticipated and true class labels of the regions of interest (RoIs); ii) Lbox: The boundary rectangle loss, same as Faster R-CNN. This is computed as a smooth L1 loss between the anticipated and true matches of the boundary boxes; and iii) Lmask: One binary mask with resolution m*m for every K classes is encoded by the binary mask loss. This is estimated as a binary cross-entropy loss among the expected and actual pixels in the masks. L = Lcls + Lbbox + Lmask. (1) Figure 7 displays the sample detection of nodule with the probability of detection being a nodule. Figure 7. Nodule detection with probability score To measure how well an object detection model performs, we need a metric task that takes into contemplation both the accuracy of the classification and the correctness of the localization. One such metric is the mean average precision (mAP), that is sustained by the average precision (AP) of each object class. AP is computed exploiting the intersection over union (IoU) among the expected boundary box and the actual boundary box. IoU is a magnitude relation that computes how much the two boxes overlap, ranging from 0 (no overlap) to 1 (perfect overlap). A higher IoU means a better localization. To determine whether a prediction is correct or not, we need to set a threshold for the IoU. For example, if we set the threshold to 0.5, then any prediction with IoU greater than or equivalent to 0.5 is considered a true positive (TP), meaning that the model correctly detected and classified the object. Any prediction with IoU less than 0.5 is considered a false positive (FP), signifying that the model either detected a target that did not exist or wrongly classified it. Similarly, any ground truth object that the model failed to detect is considered a false negative (FN), meaning that the model missed an existing object. A true negative (TN) is any portion of the representation where there is no object and the model did not predict any. Recall and Precision are cardinal metrics which are derivatives from TP, FP, FN, and TN. Precision processes as to how many of the predicted objects are correct as in (2), while recall processes to compute how many of the actual objects are detected as given by (3). They are defined by the pursuing formulas. Precision = TP / (TP + FP) (2) Recall = TP / (TP + FN) (3) To compute AP for a bestowed target class, we plot a precision-recall curve for different IoU thresholds and calculate the region below the curve (AUC). The higher the AUC, the better the AP. The formula for F1 score is: F1=2*((precision*recall)/ precision+recall) (4) F1 score is a measure that combines precision and recall, which are ratios of true positives, false positives, and false negatives as specified by (4). It is calculated as the harmonic mean of precision and recall, not the
  • 7. Int J Artif Intell ISSN: 2252-8938  Advanced mask region-based convolutional neural network based deep … (Bhavani Krishna) 1185 arithmetic mean2. F1 score is good when both precision and recall are high, and it is bad when either of them is low. F1 score can be weighted to give different importance to precision and recall. The results achieved by the implementation of general Mask RCNN and the proposed advanced Mask RCNN specifically for lung cancer detection is depicted in Table 2. Table 2. Comparison of general Mask RCNN and proposed advanced Mask RCNN Metrics General Mask RCNN Proposed advanced Mask RCNN Precision 63.04 98.77 Recall 54.48 96.62 Accuracy 60.32 99.32 F1 Score 78.73 97.23 Mean Dice Value 57.11 99.55 In conclusion, proposed architecture's higher performance on the Luna 16 dataset can be attributable to its capacity to learn abnormality features, hyper parameter adjustment, and customized loss function. Its exceptional performance was further enhanced by the multi-scale feature fusion using skip connections. The outcomes undeniably demonstrate that the suggested approach outperforms the current cutting-edge methodologies. 4. CONCLUSION Mask R-CNN has applications in various fields, such as computer vision, medical imaging, autonomous driving, and more. Researchers and practitioners keep extending the model for different tasks, including panoptic segmentation, where instances are segmented and labeled by category. The exponential proliferation of malignant tissues in the lungs characterizes lung malignancy, a dangerous condition. The ability to save thousands of lives depends heavily on the primal spotting of lung malignancy. The effectiveness of the various methods for detecting cancer using image processing can vary relying on the form of malignancy being found, the standard of images that are utilized, and the particular image processing procedures being used. The quick, precise, and spontaneous identification of lung cancer needs highly consistent prototypes. The framework utilized in this effort demonstrates the reliable recognition of the nodule and classification. The outcomes in this effort are shown using Mask-RCNN by backbone Resnet101. The values obtained are 99.32% Accuracy and 99.45% for Mean DICE respectively. Collaboration and data exchange between research organizations and healthcare professionals might speed up the progress of lung cancer detection. Further future systems for diagnosing lung cancer will be reliant on the integration of numerous data sources, advancements in deep learning techniques, collaboration, and extensive clinical validation. While Mask R-CNN performed remarkably well, it wasn't exempt from challenges such as computational complexity, training data requirements, and the trade-off between speed and accuracy. ACKNOWLEDGEMENTS We would like to express thanks to the Lung Image Database Consortium for making the lung cancer dataset public. REFERENCES [1] C. Venaktesh, D. Ramana, S. Y. Lakkisetty, S. S. Band, S. Agarwal, and A. Mosavi, “A neural network and optimization based lung cancer detection system in CT images,” Frontiers in Public Health, vol. 10, Juni 2022, doi: 10.3389/fpubh.2022.769692. [2] J. Feng and J. Jiang, “Deep learning-based chest CT image features in diagnosis of lung cancer,” Computational and Mathematical Methods in Medicine, vol. 2022, Januari 2022, doi: 10.1155/2022/4153211. [3] L. Cai, T. Long, Y. Dai and Y. Huang, "Mask R-CNN-based detection and segmentation for pulmonary nodule 3D visualization diagnosis," in IEEE Access, vol. 8, pp. 44400-44409, 2020, doi: 10.1109/ACCESS.2020.2976432. [4] M. Liu, J. Dong, X. Dong, H. Yu and L. Qi, "Segmentation of lung nodule in ct images based on mask R-CNN," 2018 9th International Conference on Awareness Science and Technology (iCAST), Fukuoka, Japan, 2018, pp. 1-6, doi: 10.1109/ICAwST.2018.8517248. [5] E. Kopelowitz and G. Engelhard, “Lung nodules detection and segmentation using 3D mask-RCNN,” Electrical Engineering and Systems Science > Image and Video Processing, arXiv, preprint arXiv:1907.07676, 2018. [6] A. Ter-Sarkisov, “Covid-ct-mask-net: prediction of covid-19 from ct scans using regional features,” Applied Intelligence, vol. 52, no. 9, pp. 9664-9675, Jan 2022, doi: 10.1007%2Fs10489-021-02731-6. [7] K. Liu, "STBi-YOLO: A Real-Time Object Detection Method for Lung Nodule Recognition," in IEEE Access, vol. 10, pp. 75385-75394, 2022, doi: 10.1109/ACCESS.2022.3192034. [8] C.-H.Tsai and Y.-S. Peng, “Multi-task lung nodule detection in chest radiographs with a dual head network,” Electrical Engineering and Systems Science > Image and Video Processing, arXiv, preprint arXiv: 2207.03050, 2022.
  • 8.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 1, March 2024: 1179-1186 1186 [9] Kh. Aghajani, “Automatic detection of lung nodules on computer tomography scans with a deep direct regression method,” Journal of AI and Data Mining, vol. 10, no. 2, pp. 207-215, April 2022, doi: 10.22044/jadm.2022.11431.2303. [10] R. Zhang, C. Cheng, X. Zhao, and X. Li, “Multiscale mask R-CNN-based lung tumor detection using PET imaging,” Molecular Imaging, July 2019, doi: 10.1177/1536012119863531. [11] B. K. Hatuwal and H. C. Thapa, "Lung cancer detection using convolutional neural network on histopathological images," Internasional Journal Computer Trends and Technology, vol. 68, no. 10, pp. 21-24, 2020, doi: 10.14445/22312803/IJCTT- V68I10P104. [12] F. Taher, N. Prakash, A. Shaffie, A. Soliman, and A. El-Baz, “An overview of lung cancer classification algorithms and their performances,” IAENG International Journal of Computer Science, vol. 48, no. 4, December 2021. [13] H. Yan, H. Lu, M. Ye, K. Yan, Y. Xu and Q. Jin, "Improved Mask R-CNN for Lung Nodule Segmentation," 2019 10th International Conference on Information Technology in Medicine and Education (ITME), Qingdao, China, 2019, pp. 137-141, doi: 10.1109/ITME.2019.00041. [14] W. Rahane, H. Dalvi, Y. Magar, A. Kalane and S. Jondhale, "Lung Cancer Detection Using Image Processing and Machine Learning HealthCare," 2018 International Conference on Current Trends towards Converging Technologies (ICCTCT), Coimbatore, India, 2018, pp. 1-5, doi: 10.1109/ICCTCT.2018.8551008. [15] K. He, G. Gkioxari, P. Dollár and R. Girshick, "Mask R-CNN," 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 2017, pp. 2980-2988, doi: 10.1109/ICCV.2017.322. [16] K. Bhavani and M. T. Gopalakrishna, “Comparative analysis of traditional classification and deep learning in lung cancer prediction,” Biomedical Engineering Applications, Basis and Communications, vol. 35, no. 2, December 2022, doi: 10.4015/S101623722250048X. [17] K. Bhavani and M. T. Gopalakrishna, “Nodule shadow based lung cancer detection using deep learning,” European Chemical Bulletin, vol. 12, no. 9, pp. 133-152, 2023, doi: 10.48047/ecb/2023.12.9.16. [18] K. Bhavani and M. T. Gopalakrishna, “Analysis of pre-processing techniques on CT DICOM images,” International Journal of Computer Sciences and Engineering, vol. 7, no. 6, June 2019, doi: 10.26438/ijcse/v7i6.9598. [19] H. Huang, Y. Li, R. Wu, Z. Li, Z., and J. Zhang, “Benign-malignant classification of pulmonary nodule with deep feature optimization framework,” Biomedical Signal Processing and Control, vol. 76, July 2022. [20] M. A. Mezher, A. Altamimi, and R. Altamimi, “A genetic folding strategy based support vector machine to optimize lung cancer classification,” Frontiers in Artificial Intelligence, vol. 5, Juni 2022, doi: 10.3389/frai.2022.826374. [21] M. Humayun, R. Sujatha, S. N. Almuayqil, and N. Z. Jhanjhi, "A transfer learning approach with a convolutional neural network for the classification of lung carcinoma," Healthcare, vol. 10, no. 6, p. 1058, June 2022, doi: 10.3390/healthcare10061058. [22] C. Shankara and S. A. Hariprasad, “Artificial neural network for lung cancer detection using CT images,” International Journal of Health Sciences, vol. 6, no. S2, pp. 2708-2724, 2022, doi: 10.53730/ijhs.v6nS2.5639. [23] R. Anand, N. A. Rao, and D. S. Sumukh, “Lung cancer detection and prediction using deep learning,” International Journal of Engineering Applied Sciences and Technology, vol. 7, no. 1, pp. 313-320, May 2022. [24] G. V. Saji, T. Vazim and S. Sundar, "Deep learning methods for lung cancer detection, classification and prediction-a review," 2021 Fourth International Conference on Microelectronics, Signals & Systems (ICMSS), Kollam, India, 2021, pp. 1-5, doi: 10.1109/ICMSS53060.2021.9673598. [25] I. Ali, M. Muzammil, I. U. Haq, A. A. Khaliq, and S. Abdullah, "Efficient lung nodule classification using transferable texture convolutional neural network," in IEEE Access, vol. 8, pp. 175859-175870, 2020, doi: 10.1109/ACCESS.2020.3026080. BIOGRAPHIES OF AUTHORS Bhavani Krishna received Bachelor of Engineering in Computer Science and Engineering in 2006, Master Degree M.Tech. in Computer Science and Engineering in 2009. She is working as an Assistant Professor in the Department of Information science and Engineering, Dayananda Sagar College of Engineering, Bangalore, India since 2009. She is also currently pursuing Ph.D. from Visvesvaraya Technological University, Karnataka. She has published many research papers in reputed international journals and conferences. Her area of research include image processing, Computer Vision, AI. She can be contacted at email: bhavani-ise@dayanandasagar.edu. Gopalakrishna Madigondanahalli Thimmaiah received Bachelor of Engineering in Computer Science & Engineering in 1999 from Bangalore University of Karnataka, India, the Master Degree M.Tech. in Computer Science and Engineering from Visvesvaraya Technological University, Karnataka, India in 2004 and PhD in Computer Science and Engineering from Visvesvaraya Technological University, Karnataka, India in 2014. He is currently working as Professor & Head in the Department of Artificial Inteliigence & Machine learning, SJB Institute of Technology, Bangalore, India. He has published many research papers in reputed international journals and conferences. His areas of research include Computer Vision, Pattern Recognition, Image Processing, Visual Surveillance-Object Detection/ Classifications/ Tracking/ Identifications. He can be contacted at email: aimlhod@sjbit.edu.in.