Lung Cancer Detection Model Using Deep Learning Te
Lung Cancer Detection Model Using Deep Learning Te
sciences
Article
Lung Cancer Detection Model Using Deep Learning Technique
Abdul Rahaman Wahab Sait
Abstract: Globally, lung cancer (LC) is the primary factor for the highest cancer-related mortality
rate. Deep learning (DL)-based medical image analysis plays a crucial role in LC detection and
diagnosis. It can identify early signs of LC using positron emission tomography (PET) and computed
tomography (CT) images. However, the existing DL-based LC detection models demand substantial
computational resources. Healthcare centers face challenges in handling the complexities in the model
implementation. Therefore, the author aimed to build a DL-based LC detection model using PET/CT
images. Effective image preprocessing and augmentation techniques were followed to overcome
the noises and artifacts. A convolutional neural network (CNN) model was constructed using the
DenseNet-121 model for feature extraction. The author applied deep autoencoders to minimize
the feature dimensionality. The MobileNet V3-Small model was used to identify the types of LC
using the features. The author applied quantization-aware training and early stopping strategies
to improve the proposed LC detection accuracy with less computational power. In addition, the
Adam optimization (AO) algorithm was used to fine-tune the hyper-parameters in order to reduce
the training time for detecting the LC type. The Lung-PET-CT-Dx dataset was used for performance
evaluation. The experimental outcome highlighted that the proposed model obtained an accuracy
of 98.6 and a Cohen’s Kappa value of 95.8 with fewer parameters. The proposed model can be
implemented in real-time to support radiologists and physicians in detecting LC in the earlier stages.
In the future, liquid neural networks and ensemble learning techniques will be used to enhance the
performance of the proposed LC detection model.
Keywords: lung cancer; machine learning; deep learning; positron emission tomography; computed
tomography; convolutional neural network
2.1.Image
2.1. ImageAcquisition
Acquisition
The author utilizesthe
The author utilizes the PET/CT(Lung-PET-CT-Dx)
PET/CT (Lung-PET-CT-Dx)dataset datasettotogeneralize
generalizethethe pro-
proposed
posed LC detection model. Lung-PET-CT-Dx is a recently developed
LC detection model. Lung-PET-CT-Dx is a recently developed dataset that comprises dataset that com-
prisesannotated
31,562 31,562 annotated
images.images. It is publicly
It is publicly available
available in the repository
in the repository [36,37].[36,37]. The im-
The images were
ages were retrospectively
retrospectively collected fromcollected fromhealthcare
multiple multiple healthcare centers
centers across acrossThe
China. China. The
computer
computer
center center and
and cancer cancerHarbin
institute, institute, HarbinUniversity
Medical Medical University
in Harbin,in Harbin, China, sup-the
China, supported
ported the dataset owners in obtaining permission to collect the PET/CT
dataset owners in obtaining permission to collect the PET/CT images. In addition, images. In addi-the
tion, the patient’s personal information was not included in the dataset. The dataset con-
patient’s personal information was not included in the dataset. The dataset contains CT
tains CT and PET-CT DICOM images of LC patients. It provides XML annotation files to
and PET-CT DICOM images of LC patients. It provides XML annotation files to indicate
indicate tumor location and bounding boxes. The tissue histopathological diagnosis was
tumor location and bounding boxes. The tissue histopathological diagnosis was used for
used for the patients’ classification. The CT image resolution was 512 × 512 pixels at 1
the patients’ classification. The CT image resolution was 512 × 512 pixels at 1 mm × 1 mm,
mm × 1 mm, and the PET image resolution was 200 × 200 pixels at 4.07 mm × 4.07
and the PET image resolution was 200 × 200 pixels at 4.07 mm× 4.07 mm. Five academic
mm. Five academic thoracic radiologists with expertise in LC detection were employed to
thoracic radiologists with expertise in LC detection were employed to annotate the tumor
annotate the tumor location. The Labellmg 1.4.0 package was used to capture the annota-
location. The Labellmg 1.4.0 package was used to capture the annotation. The images were
tion. The images were broadly classified into adenocarcinoma (A), small cell carcinoma
broadly classified into adenocarcinoma
(B), large cell carcinoma (E), and squamous (A), small cell carcinoma
cell carcinoma (B), largethe
(G). However, celldataset
carcinoma
is
(E), and squamous cell carcinoma (G). However, the dataset is highly imbalanced. Table 1
offers the dataset characteristics. Figure 2 highlights the sample LC images.
Appl. Sci. 2023, 13, x FOR PEER REVIEW 5 of 18
2.2.Image
2.2. ImagePreprocessing
Preprocessing
PET and
PET and CT CT images
images are usually
usually obtained
obtainedas asseparate
separatescans
scansandandmaymayexhibit
exhibitmisalign-
misalign-
mentdue
ment duetotopatient
patient mobility.
mobility. Various
Various artifacts,
artifacts, including
including motion
motion artifacts
artifacts in PETin PET and
and metal
metal artifacts
artifacts in CT may
in CT scans, scans, may influence
influence these images.
these images. The image
The image registration
registration is usedistoused to
guaran-
guarantee
tee spatial alignment
spatial alignment betweenbetween two images.
two images. The symmetric
The symmetric normalization
normalization (SyN)
(SyN) function
isfunction is widely
widely used used for non-linear
for non-linear image registration.
image registration. The author The authorthe
applied applied the SyNto
SyN function
function
handle thetodeformations
handle the deformations
and improve and improve
the image the image alignment.
alignment. The SyN integrates
The SyN function function
integrates
PET PETactivity
metabolic metabolic activity
with with CT anatomical
CT anatomical data. It employs
data. It employs a forward a forward transfor-for
transformation
generating the images. A similarity metric measures the similarity between the sourcethe
mation for generating the images. A similarity metric measures the similarity between and
sourceimage.
target and target
Theimage. The SyNemploys
SyN function function employs cross-correlation
cross-correlation computation
computation to gen-
to generate the
erate the normalized
normalized image. Inimage. In addition,
addition, it computes
it computes velocityvelocity
to applytosmoothness
apply smoothness to the
to the images.
images.
The The and
forward forward andtransformations
inverse inverse transformations
are used are used to the
to improve improve the alignment
alignment accuracy.ac-
The
images were resized to 512 × 512 pixels to improve the classification accuracy. Equation (1)
shows the computation of the SyN function. The function captures the complex defor-
Appl. Sci. 2023, 13, 12510 6 of 17
mation in the PET/CT images by building a Gaussian pyramid. It identifies the spatial
transformation for generating the spatially coincident features.
Ii = SyN ( Ii , ϕ1 , ϕ2 ), i = 1, . . . , n (1)
where s is the image size, scale is the scaling factor, and NI is the nearest neighbor interpolation.
The attenuation correction function enhances the anatomical and functional informa-
tion of the PET/CT images. It reconstructs the images by reducing the artifact impressions.
The author applies attenuation correction to address the artifacts of the PET images. In
addition, the metal artifact correction technique eliminates the artifacts of the CT images.
Equation (3) outlines the attenuation correction and metal artifact correction process. The
attenuation correction function identifies variations in tissue density by measuring the
attenuation of the gamma ray photons. It reduces the impact of scatter artifacts by con-
structing the primary photon attenuation. The metal attenuation correction function applies
an iterative reconstruction technique to reduce the metal artifacts. In addition, it replaces
the metal artifact regions with interpolated values of the surrounding tissue regions of
the images.
where A and M are the attenuation coefficients for reducing the impression of the artifacts.
To improve the image contrast level, the author employs the Retinex filtering technique
to address non-uniform illumination and shading of PET/CT images. Retinex filtering
enhances the color and brightness of the PET/CT images. It preserves the overall structure
and details of the images. It decomposes the images using the illumination and reflectance
components. The logarithmic transformation technique modifies the image by multiplying
the illumination and reflectance components into an additive relationship. As a result, the
images are reconstructed with improved contrast and visibility. Equation (4) highlights
the computational form of the Retinex filtering function. A logarithmic transformation
decomposes the images into reflectance and illumination components. The function applies
convolution to separate low- and high-frequency components. An exponential function
transforms the images into linear space.
Ii = Retinex_Filter ( Ii , I M, RF ), i = 1, . . . , n (4)
Fixed_size_ f eatures
𝐹𝑖𝑥𝑒𝑑_𝑠𝑖𝑧𝑒_𝑓𝑒𝑎𝑡𝑢𝑟𝑒𝑠 = Global_Average_Pooling_2D ( f eatures)
= 𝐺𝑙𝑜𝑏𝑎𝑙_𝐴𝑣𝑒𝑟𝑎𝑔𝑒_𝑃𝑜𝑜𝑙𝑖𝑛𝑔_2𝐷(𝑓𝑒𝑎𝑡𝑢𝑟𝑒𝑠) (7) (7)
Figure The
3. 3.
Figure proposed
The feature
proposed extraction
feature process.
extraction process.
2.5. Dimensionality Reduction
Deep autoencoders are used to minimize PET/CT image dimensionality by developing
a compact latent space representation [38]. This method employs an encoder–decoder
architecture to attempt a reconstruction of the input image from its compressed lower-
dimensional representation. The author uses the multilayer encoder network to reduce the
feature dimensionality. The encoder’s layers facilitate the conversion of input data into an
expression of reduced dimensionality by utilizing learned weights and biases. Lowering
dimensions is accomplished by employing several processes like convolutions, pooling,
Appl. Sci. 2023, 13, 12510 8 of 17
and non-linear activations. The encoder’s final stage compresses the input picture into
latent space. Compared to the original image, the dimensionality of this latent space is
substantially less. In the autoencoder design, the latent space is the limiting factor. The
condensed version of the input image captures its fundamental elements and patterns.
Using the low-dimensional representation from the latent space, the decoder network
attempts to recreate the original feature. The number of dimensions of the representation
expands with decoder layers. The decrease in dimensionality can be attributed to the
constriction in the latent space. The encoder network preserves essential characteristics of
the features while removing irrelevant or identical data.
→ e Ii
σ I = Ij
(8)
∑4j=1 e
→
where σ() is the Softmax function, I is the image, e Ii is the standard exponential factor of
the input vector, and e Ij is the standard exponential factor of the output vector.
Furthermore, the author employs a learning rate and decay schedule for improving
training stability and convergence. The data loading and preprocessing pipeline is opti-
mized to reduce the training time. The author introduces quantization-aware training to
maintain the trade-off between model size and inference speed [43]. Quantization-aware
training optimizes the proposed model using weight and activation redundancy [43]. It
allows the model to be lowered to a quarter of their original size and memory footprint.
In addition, the inference rates are increased by a factor of two to four with minimal or
no loss of accuracy [43]. The training process enables the proposed model accessible for
edge devices and real-time applications. Finally, early stopping is implemented to prevent
overfitting and reduce the training time.
TP + TN
Accuracy = (9)
TP + TN + FP + FN
TP
Precision = (10)
TP + FP
TP
Recall = (11)
TP + FN
2 × Precision × Recall
F1 − Score = (12)
( Precision + Recall )
Cohen’s Kappa measures the model’s consistency in medical image categorization
or classification settings [44]. It is a standardized method for measuring the multi-class
classification ability of the DL models. It is beneficial for imbalanced datasets where preci-
sion may not be reliable for evaluating the model’s effectiveness. In addition, it quantifies
inter-rater or inter-classifier agreement in multi-class classification problems based on
chance agreement. Equation (13) represents the mathematical form of Cohen’s Kappa.
2 × ( TP × TN − FN × FP)
Kappa = (13)
( TP + FP) + ( FP + TN ) + ( TP + FN ) + ( FN + TN )
Standard deviation (SD) and confidence intervals (CI) are used to analyze the data
distribution and population parameter estimation. The author used these metrics to identify
the variability and uncertainty in the outcome. The mean absolute deviation (MAD) and
the root mean square error (RMSE) are employed to evaluate the precision and reliability of
predictions. The MAD is used to compare the predicted and actual outcomes. It determines
the average of the absolute differences between each anticipated value and its associated
real value. The RMSE measures the average squared deviation from the projected value. It
is sensitive to outliers compared to the MAD and emphasizes the significant errors.
3. Results
The author conducted the experiment in Python 3.8.3 using the Keras library. The
proposed model is implemented in Windows 10, Intel i7, 16 GB RAM, and GeForce GTX
950 GPU. The weights of the DenseNet-121 and MobileNet V3-Small models were ex-
tracted from the repository [40,41]. Table 2 presents the hyper-parameters of the proposed
LC model.
Appl. Sci. 2023, 13, 12510 10 of 17
Parameters Value
Image dimension 512 × 512
Convolutional layers 4
Size of strides 2
Activation function Softmax
Loss function Cross-Entropy
Batch Size 256
Fully connected layer 1
Table 5 shows the performance of the proposed LC detection model with different
batch sizes. Figure 4 presents the proposed model’s performance in the multiple epochs.
The proposed model showed a significant improvement in the 180th batch. However, the
author extended the batch size to 256. The proposed feature extraction supported the
proposed LC detection model to produce an exceptional outcome.
Findings of
Figure 4. Findings
Figure of the
the performance
performance analysis
analysis in the multiple epochs.
Table 66 reveals
Table revealsthe
theperformance
performance of of
thethe
proposed model.
proposed TheThe
model. proposed model
proposed obtained
model ob-
an exceptional outcome for each class. The proposed image augmentation
tained an exceptional outcome for each class. The proposed image augmentation tech- technique
effectively
nique supported
effectively the proposed
supported modelmodel
the proposed in identifying each each
in identifying class’s key patterns.
class’s The
key patterns.
findings of the performance analysis are presented in Table 6. Figure 5 highlights
The findings of the performance analysis are presented in Table 6. Figure 5 highlights the the
outcome of the performance evaluation.
outcome of the performance evaluation.
Table 6. Findings of the performance evaluation.
Table 6. Findings of the performance evaluation.
Types of LC Accuracy Precision Recall F1-Score Kappa
Types of LC Accuracy Precision Recall F1-Score Kappa
A 98.6 97.8 97.5 97.6 96.1
BA 98.6
98.6 97.8
97.6 97.5
97.7 97.6
97.6 96.1
95.8
EB 98.6
98.8 97.6
98.1 97.7
98.5 97.6
98.3 95.8
96.1
Appl. Sci. 2023, 13, x FOR PEER REVIEW 12 of 18
GE 98.5
98.8 98.3
98.1 98.7
98.5 98.5
98.3 95.4
96.1
Average
G 98.6
98.5 97.9
98.3 98.1
98.7 98.0
98.5 95.8
95.4
Average 98.6 97.9 98.1 98.0 95.8
Findings of
Figure 5. Findings
Figure of the
the multi-class
multi-class classification.
classification.
Table 77outlines
Table outlinesthe
thegeneralization
generalizationofof
thethe
LCLC detection
detection models
models on Lung-PET-CT-
on the the Lung-PET-
CT-Dx dataset. The proposed model outperformed the existing LC detection
Dx dataset. The proposed model outperformed the existing LC detection models models by
by
achieving an
achieving an exceptional
exceptional outcome.
outcome. The
The MobileNet
MobileNet V3-Small
V3-Small architecture
architecture supported
supported the
the
proposed LC detection model to achieve a better outcome. In addition, the suggested
proposed LC detection model to achieve a better outcome. In addition, the suggested im-image
age preprocessing technique assisted the proposed model in detecting the crucial LC pat-
terns in the PET/CT images. Figure 6 shows the outcome of the comparison analysis.
preprocessing technique assisted the proposed model in detecting the crucial LC patterns
in the PET/CT images. Figure 6 shows the outcome of the comparison analysis.
Table 8. Findings of the AU-ROC, AU-PRC, and loss analysis with SD and CI.
Table 8. Findings of the AU-ROC, AU-PRC, and loss analysis with SD and CI.
4. Discussions
The author proposed an LC detection model using this study’s PET/CT images. An
extensive image preprocessing technique was developed to overcome the challenges of
enhancing the image quality. The SyN function was used to normalize the CT and PET
images. Additionally, the author applied resampling, Retinex filtering, and attenuation
techniques to improve the image quality. The image augmentation was performed to
overcome the data imbalance. A CNN model was developed using the DenseNet-121
model’s weights for the feature extraction. Deep autoencoders reduced the dimensionality
of the images. Finally, a CNN model with a MobileNet V3 model weights was employed
for classifying the type of LC. The author improved the performance of the proposed
model by applying a quantization-aware training approach. In addition, the AO-based
hyperparameter optimization supported the proposed LC model to detect the type of LC
with a minimum hardware and software requirement. The experimental results revealed
the significance of the proposed LC detection model in identifying the LC types.
Table 3 outlined the hyper-parameter optimization’s significance in optimizing the
proposed LC detection parameters. AO-based hyper-parameter optimization supported the
model to yield an exceptional outcome compared to Adagrad and RMSProp optimizations.
Table 4 revealed that the proposed model demands minimum computational power to
generate an effective outcome. It required parameters of 2.1 m and FLOPs of 176 m with a
training time of 128 s for identifying the types of LC. Table 5 and Figure 4 highlighted the
number of epochs of the proposed model to deliver a superior outcome. The proposed LC
detection model achieved an accuracy of 98.6 with a batch size of 256. However, it obtained
an accuracy of 98.4 in the 180th batch. Table 6 outlined the performance evaluation findings
of the suggested LC detection model. The proposed model produced an outstanding
result for each class using the PET/CT images. It is evident that the suggested image
preprocessing and augmentation techniques assisted the proposed LC detection model in
identifying the LC types. Table 7 emphasized the key findings of the comparison analysis.
The suggested model surpassed the existing LC detection approaches by reaching an
optimal accuracy and F1-score using minimal resource utilization. Finally, Table 8 showed
that the proposed model obtained a higher AU-ROC and AU-PRC with a minimum loss
of 0.89.
Barbouchi et al. [34] built a transformer-based LC detection model. They developed
a histologic subtype classification to detect the tumor location from the PET/CT images.
However, the vision transformers demand enormous computational resources compared
to the proposed LC detection model. The proposed LC detection model outperformed
Barbouchi et al.’s model with minimum hardware resources.
El-Hamdi et al. [35] proposed an LC detection model and generalized their model
using the Lung-PET-CT-Dx dataset. They built a multi-class CNN model for classifying the
PET/CT images. The VGG-16 network was used to extract the features. Subsequently, a
classifier was employed to classify the extracted features. Similarly, the proposed model
used the DenseNet-121-based feature extractor. In addition, the author constructed an im-
age classifier using the weights of the MobileNet V3-Small model. The findings highlighted
that the proposed LC detection model outperformed the El-Hamdi et al. model.
Appl. Sci. 2023, 13, 12510 14 of 17
Goswami and Singh [37] developed an image classification model for detecting LC
from PET/CT images. They obtained an optimal outcome using the Lung-PET-CT-Dx
dataset. Likewise, the proposed model achieved superior results by employing effective
image processing, feature extraction, and image classification models.
The DenseNet-121-based LC detection model achieved an accuracy of 96.4 with param-
eters of 8.3 and FLOPs of 679. However, the proposed model obtained exceptional accuracy
with limited computational resources. The EfficientNet-based LC detection model obtained
a higher accuracy and F1-score. However, the model may not perform optimally on the
real-time images. It required a considerable computational resource for classifying the
PET/CT images. In addition, it demands additional training time to capture the fine details
of the images. The ShuffleNet-based LC detection model offered an exceptional outcome.
However, the shallower architecture of the ShuffleNet model may limit the LC detection
performance. In contrast, the proposed LC detection model presented an outstanding
result with limited computational resources. The combination of the DenseNet-121 and
MobileNet V3 models yielded better results.
PET/CT imaging offers a complete assessment of lung tissue, effectively highlighting
regions exhibiting elevated metabolic activity while offering valuable anatomical context.
Healthcare practitioners can choose the surgery, chemotherapy, or radiation therapy for
LC by precisely identifying the types of LC. The proposed model was trained to analyze
and extract features from complex images. It can be implemented in edge devices to render
effective healthcare services to individuals. It can facilitate the detection and classification
of various forms of LC. In addition, it can locate tiny lung nodules that human radiologists
may ignore due to the early stage of development. The timely identification of medical
conditions can significantly enhance patients’ overall prognosis and treatment outcomes.
The proposed LC detection model can assist physicians in customizing LC therapies for
each patient. It can minimize the number of FPs and FNs associated with cancer detection.
This intervention enhances the precision of diagnostic assessments.
Furthermore, the proposed model can substantially accelerate the diagnostic proce-
dure. Radiologists can use the proposed model to select cases and focus on complicated
tasks. It can contribute to advancing research and facilitating clinical trials. The proposed
LC detection model generates outcomes in a limited time. It enables healthcare centers to
identify LC in the earlier stages. Healthcare centers can employ the model to detect LC
without professional radiologists. The reduced number of FPs and FNs guaranteed the
optimal accuracy of the proposed model. Additionally, the proposed model is cost-effective
and can be implemented with low computational resources.
However, the author encountered a few limitations while implementing the proposed
model. The PET/CT images contained motion artifacts and metallic artifacts. The low
spatial resolution caused challenges in identifying the crucial patterns. Tissue misalignment
during the scan reduced the image quality. In addition, PET and CT images were obtained
separately. Thus, there was a misalignment due to the functional information of PET
and the anatomical information of CT images. In addition, the dataset samples were
highly imbalanced. Nonetheless, the author employed compelling image preprocessing
techniques, including the SyN function, artifact correction technique, and attenuation
approaches, to improve image quality. The proposed image augmentation technique
supported the proposed LC detection model to overcome the data imbalance limitation.
While implementing the proposed model, the author faced challenges with the graphical
processing unit configuration. The quantization-aware training approach has supported
the author deploying the model while using low computational resources. The lack of
accessible, high-quality labeled data to train DL models is a significant barrier. Generating
massive datasets with precise annotations is essential for LC diagnosis. The research
requires more diversified, publicly available, and well-annotated datasets. The strong
generalization capabilities of DL models can support a physician in treating a wide range
of populations. Maintaining model robustness and performance in clinical contexts is
Appl. Sci. 2023, 13, 12510 15 of 17
challenging. A balance between FP and FN is crucial. PET/CT images contain noises and
artifacts. Extensive image preprocessing was required to improve image clarity.
To enhance the proposed LC detection model, it is essential to integrate technology de-
velopments with clinical expertise, information availability, and compliance standards. By
considering these factors, the proposed model can serve as a significant resource in assisting
healthcare professionals in making efficient and prompt decisions concerning the diagnosis
and treatment of LC. The liquid neural network (LNN) extends the recurrent neural net-
work’s capabilities by including dynamic hidden states for analyzing video and images [45].
It computes the expected outcome and its shifting hidden state at each prediction step. In
the future, LNN-based applications can significantly outperform currently used machine
vision approaches. Using the LNN technique, the feature engineering process can be im-
proved. A set of optimal features can optimize the multi-class classification accuracy. The
proposed model can be extended using LNN and ensemble learning techniques to reduce
the computation time and resources in identifying LC from the multi-modality images.
5. Conclusions
The author built an LC detection model using the DL technique in this study. The pro-
posed LC detection model employed image preprocessing and augmentation techniques to
improve the performance of the proposed model. The SyN function and Retinex filters were
used to remove noise and artifacts. The GANs technique was used to increase the dataset
size. The author developed the DenseNet-121-based feature extraction model for extracting
the critical features. A deep autoencoder model was employed to reduce the number of
features without losing the essential details. A MobileNet V3-Small model-based image
classifier was constructed using quantization-aware training and early stopping strategies.
The author generalized the proposed model using the Lung-PET-CT-Dx dataset. The find-
ings highlighted the significant contribution of the proposed model in identifying the LC
types. The proposed LC detection model outperformed the existing models. It required
limited computational resources for detecting LC from the complex images. Healthcare
centers can deploy the proposed model without any implementation complexities. In
addition, the lightweight architecture of the proposed model offers an opportunity to
integrate it into the edge devices. However, a few limitations were encountered during the
development. There is a lack of highly balanced datasets for providing extensive training
for the LC detection model. The low-quality images caused challenges in classifying the
LC types. As a result, extended training is required to implement the proposed model in
the real-time application. The integration of the LNN technique can improve the process of
feature extraction. In addition, the ensemble learning approach can accurately classify the
medical images. The author will employ the LNN and ensemble learning approaches to
extract crucial features and classify low-quality images.
Funding: This work was supported by the Deanship of Scientific Research, Vice Presidency for
Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia (Grant No. 4702).
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Lung-PET-CT-Dx Dataset. The Cancer Imaging Archive. https://doi.
org/10.7937/TCIA.2020.NNC2-0461, accessed on 1 January 2023.
Conflicts of Interest: The author declares no conflict of interest.
References
1. Gu, Y.; Chi, J.; Liu, J.; Yang, L.; Zhang, B.; Yu, D.; Zhao, Y.; Lu, X. A survey of computer-aided diagnosis of lung nodules from CT
scans using deep learning. Comput. Biol. Med. 2021, 137, 104806. [CrossRef]
2. Sharif, M.I.; Li, J.P.; Naz, J.; Rashid, I. A comprehensive review on multi-organs tumor detection based on machine learning.
Pattern Recognit. Lett. 2020, 131, 30–37. [CrossRef]
Appl. Sci. 2023, 13, 12510 16 of 17
3. Adhikari, T.M.; Liska, H.; Sun, Z.; Wu, Y. A Review of Deep Learning Techniques Applied in Lung Cancer Diagnosis. In Signal and
Information Processing, Networking and Computers: Proceedings of the 6th International Conference on Signal and Information Processing,
Networking and Computers (ICSINC), Guiyang, China, 13–16 August 2019; Springer: Singapore, 2020; pp. 800–807.
4. Saba, T. Recent advancement in cancer detection using machine learning: Systematic survey of decades, comparisons and
challenges. J. Infect. Public Health 2020, 13, 1274–1289. [CrossRef]
5. Ozdemir, O.; Russell, R.L.; Berlin, A.A. A 3D probabilistic deep learning system for detection and diagnosis of lung cancer using
low-dose CT scans. IEEE Trans. Med. Imaging 2019, 39, 1419–1429. [CrossRef]
6. Yu, L.; Tao, G.; Zhu, L.; Wang, G.; Li, Z.; Ye, J.; Chen, Q. Prediction of pathologic stage in non-small cell lung cancer using machine
learning algorithm based on CT image feature analysis. BMC Cancer 2019, 19, 464. [CrossRef]
7. Asuntha, A.; Srinivasan, A. Deep learning for lung Cancer detection and classification. Multimed. Tools Appl. 2020, 79, 7731–7762.
[CrossRef]
8. Shakeel, P.M.; Burhanuddin, M.A.; Desa, M.I. Lung cancer detection from CT image using improved profuse clustering and deep
learning instantaneously trained neural networks. Measurement 2019, 145, 702–712. [CrossRef]
9. Xie, Y.; Meng, W.Y.; Li, R.Z.; Wang, Y.W.; Qian, X.; Chan, C.; Yu, Z.F.; Fan, X.X.; Pan, H.D.; Xie, C.; et al. Early lung cancer
diagnostic biomarker discovery by machine learning methods. Transl. Oncol. 2021, 14, 100907. [CrossRef]
10. Sori, W.J.; Feng, J.; Godana, A.W.; Liu, S.; Gelmecha, D.J. DFD-Net: Lung cancer detection from denoised CT scan image using
deep learning. Front. Comput. Sci. 2021, 15, 152701. [CrossRef]
11. Ak, M.F. A comparative analysis of breast cancer detection and diagnosis using data visualization and machine learning
applications. Healthcare 2020, 8, 111. [CrossRef]
12. Pradhan, K.; Chawla, P. Medical Internet of things using machine learning algorithms for lung cancer detection. J. Manag. Anal.
2020, 7, 591–623. [CrossRef]
13. Heuvelmans, M.A.; van Ooijen, P.M.; Ather, S.; Silva, C.F.; Han, D.; Heussel, C.P.; Hickes, W.; Kauczor, H.U.; Novotny, P.; Peschl,
H.; et al. Lung cancer prediction by Deep Learning to identify benign lung nodules. Lung Cancer 2021, 154, 1–4. [CrossRef]
14. Singh, G.A.P.; Gupta, P.K. Performance analysis of various machine learning-based approaches for detection and classification of
lung cancer in humans. Neural Comput. Appl. 2019, 31, 6863–6877. [CrossRef]
15. Sajja, T.; Devarapalli, R.; Kalluri, H. Lung Cancer Detection Based on CT Scan Images by Using Deep Transfer Learning. Trait. Du
Signal 2019, 36, 339–344. [CrossRef]
16. Coccia, M. Deep learning technology for improving cancer care in society: New directions in cancer imaging driven by artificial
intelligence. Technol. Soc. 2020, 60, 101198. [CrossRef]
17. Bhatia, S.; Sinha, Y.; Goel, L. Lung cancer detection: A deep learning approach. In Soft Computing for Problem Solving: SocProS
2017; Springer: Singapore, 2019; Volume 2, pp. 699–705.
18. Nasrullah, N.; Sang, J.; Alam, M.S.; Mateen, M.; Cai, B.; Hu, H. Automated lung nodule detection and classification using deep
learning combined with multiple strategies. Sensors 2019, 19, 3722. [CrossRef]
19. Chaunzwa, T.L.; Hosny, A.; Xu, Y.; Shafer, A.; Diao, N.; Lanuti, M.; Christiani, D.C.; Mak, R.H.; Aerts, H.J. Deep learning
classification of lung cancer histology using CT images. Sci. Rep. 2021, 11, 5471. [CrossRef]
20. Bhandary, A.; Prabhu, G.A.; Rajinikanth, V.; Thanaraj, K.P.; Satapathy, S.C.; Robbins, D.E.; Shasky, C.; Zhang, Y.D.; Tavares, J.M.R.;
Raja, N.S.M. Deep-learning framework to detect lung abnormality—A study with chest X-Ray and lung CT scan images. Pattern
Recognit. Lett. 2020, 129, 271–278. [CrossRef]
21. Kanavati, F.; Toyokawa, G.; Momosaki, S.; Rambeau, M.; Kozuma, Y.; Shoji, F.; Yamazaki, K.; Takeo, S.; Iizuka, O.; Tsuneki, M.
Weakly-supervised learning for lung carcinoma classification using deep learning. Sci. Rep. 2020, 10, 9297. [CrossRef]
22. Hallitschke, V.J.; Schlumberger, T.; Kataliakos, P.; Marinov, Z.; Kim, M.; Heiliger, L.; Seibold, C.; Kleesiek, J.; Stiefelhagen, R.
Multimodal Interactive Lung Lesion Segmentation: A Framework for Annotating PET/CT Images based on Physiological and
Anatomical Cues. arXiv 2023, arXiv:2301.09914.
23. Ardila, D.; Kiraly, A.P.; Bharadwaj, S.; Choi, B.; Reicher, J.J.; Peng, L.; Tse, D.; Etemadi, M.; Ye, W.; Corrado, G.; et al. End-to-
end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nat. Med. 2019,
25, 954–961. [CrossRef] [PubMed]
24. Marentakis, P.; Karaiskos, P.; Kouloulias, V.; Kelekis, N.; Argentos, S.; Oikonomopoulos, N.; Loukas, C. Lung cancer histology
classification from CT images based on radiomics and deep learning models. Med. Biol. Eng. Comput. 2021, 59, 215–226. [CrossRef]
25. Chabon, J.J.; Hamilton, E.G.; Kurtz, D.M.; Esfahani, M.S.; Moding, E.J.; Stehr, H.; Schroers-Martin, J.; Nabet, B.Y.; Chen, B.;
Chaudhuri, A.A.; et al. Integrating genomic features for non-invasive early lung cancer detection. Nature 2020, 580, 245–251.
[CrossRef]
26. Zhang, R.; Cheng, C.; Zhao, X.; Li, X. Multiscale mask R-CNN–based lung tumor detection using PET imaging. Mol. Imaging
2019, 18, 1536012119863531. [CrossRef] [PubMed]
27. Kriegsmann, M.; Haag, C.; Weis, C.A.; Steinbuss, G.; Warth, A.; Zgorzelski, C.; Muley, T.; Winter, H.; Eichhorn, M.E.; Eichhorn, F.;
et al. Deep learning for the classification of small-cell and non-small-cell lung cancer. Cancers 2020, 12, 1604. [CrossRef] [PubMed]
28. Polat, H.; Danaei Mehr, H. Classification of pulmonary CT images by using hybrid 3D-deep convolutional neural network
architecture. Appl. Sci. 2019, 9, 940. [CrossRef]
29. Patel, D.; Shah, Y.; Thakkar, N.; Shah, K.; Shah, M. Implementation of artificial intelligence techniques for cancer detection.
Augment. Hum. Res. 2020, 5, 6. [CrossRef]
Appl. Sci. 2023, 13, 12510 17 of 17
30. Baranwal, N.; Doravari, P.; Kachhoria, R. Classification of histopathology images of lung cancer using convolutional neural
network (CNN). In Disruptive Developments in Biomedical Applications; CRC Press: Boca Raton, FL, USA, 2022; pp. 75–89.
31. Levy-Jurgenson, A.; Tekpli, X.; Kristensen, V.N.; Yakhini, Z. Spatial transcriptomics inferred from pathology whole-slide images
links tumor heterogeneity to survival in breast and lung cancer. Sci. Rep. 2020, 10, 18802. [CrossRef]
32. Mishra, S.; Thakkar, H.K.; Mallick, P.K.; Tiwari, P.; Alamri, A. A sustainable IoHT based computationally intelligent healthcare
monitoring system for lung cancer risk detection. Sustain. Cities Soc. 2021, 72, 103079. [CrossRef]
33. Luna, J.M.; Chao, H.H.; Diffenderfer, E.S.; Valdes, G.; Chinniah, C.; Ma, G.; Cengel, K.A.; Solberg, T.D.; Berman, A.T.; Simone,
C.B., II. Predicting radiation pneumonitis in locally advanced stage II–III non-small cell lung cancer using machine learning.
Radiother. Oncol. 2019, 133, 106–112. [CrossRef]
34. Barbouchi, K.; El Hamdi, D.; Elouedi, I.; Aïcha, T.B.; Echi, A.K.; Slim, I. A transformer-based deep neural network for detection
and classification of lung cancer via PET/CT images. Int. J. Imaging Syst. Technol. 2023, 33, 1383–1395. [CrossRef]
35. El Hamdi, D.; Elouedi, I.; Slim, I. Computer-Aided Classification of Cell Lung Cancer Via PET/CT Images Using Convolutional
Neural Network. Int. J. Image Graph. 2023, 2450040. [CrossRef]
36. Li, P.; Wang, S.; Li, T.; Lu, J.; HuangFu, Y.; Wang, D. A Large-Scale CT and PET/CT Dataset for Lung Cancer Diagnosis
(Lung-PET-CT-Dx) [Dataset]. Cancer Imaging Arch. 2020. [CrossRef]
37. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer
Imaging Archive (TCIA): Maintaining and Operating a Public Information Repository. J. Digit. Imaging 2013, 26, 1045–1057.
[CrossRef]
38. Goswami, J.; Singh, K.K. Pulmonary Lung Cancer Classification Using Deep Neural Networks. In Machine Vision and Augmented
Intelligence: Select Proceedings of MAI 2022; Springer Nature Singapore: Singapore, 2023; pp. 395–407.
39. Liu, Z.; Yao, C.; Yu, H.; Wu, T. Deep reinforcement learning with its application for lung cancer detection in medical Internet of
Things. Future Gener. Comput. Syst. 2019, 97, 1–9. [CrossRef]
40. DenseNet-121 Model. Available online: https://github.com/flyyufelix/DenseNet-Keras/blob/master/densenet121.py (accessed
on 1 January 2023).
41. MobileNet-V3 Small Model. Available online: https://github.com/topics/mobilenetv3-small (accessed on 3 January 2023).
42. Bi, L.; Kim, J.; Kumar, A.; Feng, D.; Fulham, M. Synthesis of positron emission tomography (PET) images via multi-channel
generative adversarial networks (GANs). In Molecular Imaging, Reconstruction and Analysis of Moving Body Organs, and Stroke
Imaging and Treatment, Proceedings of the Fifth International Workshop, CMMI 2017, Second International Workshop, RAMBO 2017, and
First International Workshop, SWITCH 2017, Held in Conjunction with MICCAI 2017, Québec City, QC, Canada, 14 September 2017;
Springer International Publishing: Cham, Switzerland, 2017; pp. 43–51.
43. Abid, A.; Sinha, P.; Harpale, A.; Gichoya, J.; Purkayastha, S. Optimizing medical image classification models for edge devices. In
Distributed Computing and Artificial Intelligence, Volume 1: 18th International Conference; Springer International Publishing: Cham,
Switzerland, 2022; pp. 77–87.
44. Dominik, M.; Soto-Rey, I.; Kramer, F. Towards a guideline for evaluation metrics in medical image segmentation. BMC Res. Notes
2022, 15, 210.
45. Ramin, H.; Lechner, M.; Amini, A.; Rus, D.; Grosu, R. Liquid time-constant networks. In Proceedings of the 2021 AAAI Conference
on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 7657–7666.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.