Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Dynamic Digital Radiography (DDR) in the Diagnosis of a Diaphragm Dysfunction
Previous Article in Journal
The Challenge of Deep Learning for the Prevention and Automatic Diagnosis of Breast Cancer: A Systematic Review
Previous Article in Special Issue
DeSPPNet: A Multiscale Deep Learning Model for Cardiac Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Toward Robust Lung Cancer Diagnosis: Integrating Multiple CT Datasets, Curriculum Learning, and Explainable AI

by
Amira Bouamrane
1,
Makhlouf Derdour
1,
Akram Bennour
2,*,
Taiseer Abdalla Elfadil Eisa
3,
Abdel-Hamid M. Emara
4,
Mohammed Al-Sarem
5 and
Neesrin Ali Kurdi
6
1
LIAOA Laboratory, University of Oum El-Bouaghi-Larbi Benmhidi, Oum El-Bouaghi 04000, Algeria
2
LAMIS Laboratory, Echahid Cheikh Larbi Tebessi University, Tebessa 12002, Algeria
3
Applied College, King Khalid University, Mahayil 62529, Saudi Arabia
4
Department of Computers and Systems Engineering, Faculty of Engineering, Al-Azhar University, Cairo 11884, Egypt
5
Department of Information Technology, Aylol University College, Yarim 547, Yemen
6
College of Computer Science and Engineering, Taibah University, Medina 41477, Saudi Arabia
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(1), 1; https://doi.org/10.3390/diagnostics15010001
Submission received: 30 October 2024 / Revised: 6 December 2024 / Accepted: 10 December 2024 / Published: 24 December 2024

Abstract

:
Background and Objectives: Computer-aided diagnostic systems have achieved remarkable success in the medical field, particularly in diagnosing malignant tumors, and have done so at a rapid pace. However, the generalizability of the results remains a challenge for researchers and decreases the credibility of these models, which represents a point of criticism by physicians and specialists, especially given the sensitivity of the field. This study proposes a novel model based on deep learning to enhance lung cancer diagnosis quality, understandability, and generalizability. Methods: The proposed approach uses five computed tomography (CT) datasets to assess diversity and heterogeneity. Moreover, the mixup augmentation technique was adopted to facilitate the reliance on salient characteristics by combining features and CT scan labels from datasets to reduce their biases and subjectivity, thus improving the model’s generalization ability and enhancing its robustness. Curriculum learning was used to train the model, starting with simple sets to learn complicated ones quickly. Results: The proposed approach achieved promising results, with an accuracy of 99.38%; precision, specificity, and area under the curve (AUC) of 100%; sensitivity of 98.76%; and F1-score of 99.37%. Additionally, it scored a 00% false positive rate and only a 1.23% false negative rate. An external dataset was used to further validate the proposed method’s effectiveness. The proposed approach achieved optimal results of 100% in all metrics, with 00% false positive and false negative rates. Finally, explainable artificial intelligence (XAI) using Gradient-weighted Class Activation Mapping (Grad-CAM) was employed to better understand the model. Conclusions: This research proposes a robust and interpretable model for lung cancer diagnostics with improved generalizability and validity. Incorporating mixup and curriculum training supported by several datasets underlines its promise for employment as a diagnostic device in the medical industry.

1. Introduction

Given the success of artificial intelligence in general and deep learning in particular in the health field and for disease diagnosis, researchers have increasingly focused on developing computer-aided diagnosis (CADx) systems [1]. Lung cancer has been a prominent area of focus for these systems due to the severity of this disease [2]. It ranks as the top cancer type and is the leading cause of cancer-related deaths worldwide, with more than 2 million cases estimated in 2024 [3], according to the World Health Organization and the American Cancer Society, mainly due to delayed diagnoses. Research has revealed that timely detection is a key factor contributing to significantly improving five-year survival rates and treatment success [4]. This has led researchers to investigate ways in which CADx could further be developed to improve precision and timeliness, especially considering the scarcity of qualified specialist personnel, human subjectivity, and the exhaustion of radiologists due to extended working hours [5,6]. This disease occurs because of the unregulated growth of cells in the body, mostly seen in the lung parenchyma, and it can start anywhere but most often does so from the epithelial tissues within the bronchi or alveoli [7]. Nevertheless, it is, more often than not, manageable, provided that it is detected early enough [8]. The actual threat of lung cancer becomes substantial when metastasis takes place [9]. The growth and dispersion of lung nodules vary with the type and features of lung cancer, and are commonly classified into two main types: small-cell lung carcinoma and non-small-cell lung carcinoma [10]. The computed tomography (CT) scan is a modality that has been around for several years and is mainly used in lung cancer screening [11]. It employs several X-rays, which are later manipulated using computer technology in order to produce a 3-D image of a particular region of interest [12]. It provides a number of cross-sectional images that can identify masses as small as 1 to 2 millimeters in size [13]. However, CT scans have some drawbacks, including reduced contrast for soft tissues [14], particularly in the case of blood-borne tumor infiltration into soft tissue, and the risk of artifacts created from varying tumor shapes and the extents of growth [15]. Additionally, CADx allows for the processing of CT images, recognizes different features in the images, removes artifacts, and analyzes details and characteristics such as the tumor’s shape, size, and location [16]. Moreover, CT images are processed using deep learning algorithms, which help extract high-quality features from CT scans and aid tumor classification [17]. Moreover, the process of lung tumor CADx contains four phases [18], starting with segmenting the lungs from the surrounding tissue and lung wall. Next, the nodules are detected and segmented. We distinguish various segmentation techniques, such as Mask R-CNN, thresholding, U-Net [19], graph-based methods, and DeepLabv3 [20]. After segmentation, the final step was to classify these nodules as either benign or malignant (Figure 1).
Convolutional neural networks (CNNs) enable CT image analysis to extract high-quality and quantitative features through a hierarchical representation of input data [21]. Additionally, transfer learning and deep models trained on large datasets can learn a substantial number of features, making them highly effective for extracting extensive features to distinguish the characteristics of these tumors [22]; such models include EfficientNet [23], MobileNet, DenseNet [24], VGG16 [25], VGG19 [26], and ResNet50 [24]. Moreover, incorporating DL models into hybrid models further augments the feature extraction capability [27,28]. Considering the significance of data, numerous medical centers and specialists have made available lung cancer databases to assist researchers in creating diagnostic systems such as LIDC-IDRI [29,30], NIH dataset [31], IQ-OTH/NCCD [32], and LUNA-16 [33]. Nevertheless, this domain suffers from data heterogeneity and inaccessibility [34]. To solve this problem, some researchers have employed techniques such as data augmentation [35] or dataset fusion [36].
This study presents a hybrid model for lung cancer diagnosis that combines the effectiveness of MobileNetV3Small and ResNet50 for feature extraction and is based on five different and heterogeneous datasets to avoid bias. The proposed approach employs curriculum learning to train on these datasets from the simplest to the most complex data and also uses mixup augmentation techniques to enhance the robustness of the model and improve its generalization capability. Finally, the Gram-CAM is applied to improve model interpretability and explainability by identifying the key features on which the model bases its decisions. The primary contributions of this study are as follows:

Research Contributions

  • A new model that adopts curriculum strategies to analyze training datasets, training on data from the easiest to the most complex, which improves learning efficiency and enables the understanding of harder tasks.
  • High-quality and abundant features were obtained using ResNet50 and MobileNetV3 Small with reduced complexity, as they are used solely for feature extraction.
  • Data augmentation and model robustness were enhanced using mixup augmentation techniques by randomly merging images with their labels. This allows the model to analyze more images with diverse features, thereby increasing its robustness and generalization ability.
  • Bias reduction was achieved by using multiple datasets, and five different databases were used to train the model with diverse demographic samples and varying characteristics, reducing the model’s data dependency and bias, and enhancing its effectiveness.
  • Generalizability Assessment: by evaluating the model using a totally external dataset.
  • The model’s explainability and trustworthiness are enhanced by using Grad- CAM to identify the most relevant regions in the screening that the model relied on for its decisions, thereby increasing the model’s transparency.
The rest of the paper is organized as follows. The Section 2 analyzes related works connected to the paper’s research, focusing on using CT scans and deep learning. The Section 3 presents the materials and methods, summarizing the proposed model, methods used, research problem, desired objectives, and various databases used. The results are presented in the Section 4. The impacts of using curriculum learning and mixup methods are analyzed, and explainable AI is used to rationalize some of the classifications made by the model. Finally, the Section 5 reviews the contents of this paper, and some areas of research that could expand the results are proposed.

2. Related Work

Many diagnostic systems have utilized CNNs and deep learning to diagnose lung cancer and have proven to be effective. A number of them used LIDC-IDRI to evaluate their models, such as Zhao et al., who developed a new transformer-inspired model for lung nodule classification based on the LIDC-IDRI dataset; the BiCFormer enhanced feature extraction using a novel multi-layer GAN for data augmentation and a bi-level coordinate (BiC) encoder. This model achieved 97.4% of accuracy [37]. Similarly, Gopinath et al. presented a novel classification method, Deep Fused Features-Based Cat-Optimized Networks (DFF-CON) based on LIDC-IDRI to enhance lung cancer diagnosis from CT scans, and they employed a saliency map to highlight important features. DFF-CON achieved high performance across metrics, with 99.89% accuracy and 99.88% F1-score, outperforming existing models [38]. Saied et al. developed AI approaches to classify pulmonary nodules efficiently from CT scans. Using DenseNet-121 and SVM based on the LIDC-IDRI dataset, they focused on extracting features and then applied PCA for feature selection, and DenseNet-121 achieved the best results combined with SVM. They also tried VGG-16 and VGG-19, and DenseNet-169. DenseNet-121 combined with SVM achieved an accuracy of 90.39%, sensitivity of 90.32%, and specificity of 93.65% [39]. Meng et al. proposed a machine learning model using CT features and serum biomarkers based on two local datasets by highlighting nodule diameter and average CT value as key predictive features, aiding early intervention and improving early lung cancer diagnosis. The Gradient Boosting Machine (GBM) model showed superior accuracy and AUC of 99% and 93.1%, respectively, with 85.7% and 95.5% in the external phase. The study also included the use of Shapley additive explanations (SHAP) to explain the features on which the model’s decision is based [40]. Utilizing the Kaggle CT scan of lung cancers, Lanjewar et al. suggested a modified DenseNet201 architecture to classify the four types of lung cancer. They conducted five experimental scenarios, including one-upping the architecture of DenseNet201 in a bid to make it less trainable and also using it as a feature extractor with a Support Vector Machine, Logistic Regression, Random Forest, Decision Tree, Gaussian Naive Bayes, and KNN. Two performance improvement techniques were used to achieve a feasible number of extracted features, ETC and MRMR. The method proposed achieved good results, with the highest accuracy of 100%, an AUC value of 99.25%, and a Kappa score of 93%. It should be mentioned that 5-fold cross-validation achieves an accuracy level of 95%. However, despite these promising results, the authors stated that there are limitations, such as the small dataset and the necessity to validate and further test on larger datasets, for the model’s improvement in robustness and generalizability [41]. Shen et al. proposed a WS-LungNet based on the LIDC-IDRI dataset, a weakly supervised deep learning framework to segment 3D pulmonary nodules based on unlabeled data and to address scarcity and inconsistency in computer-aided lung cancer diagnosis (CAD) from CT images. By using semi-supervised segmentation with adversarial learning and cross-nodule attention mechanisms, WS-LungNet achieved 82.99% competition performance metric (CPM), an 88.63% area under the curve (AUC), and DROC of 87.12% [42].
Wahab et al. presented a deep learning-based model for lung cancer detection using DenseNet-121 CNN for feature extraction, deep autoencoders to minimize the feature dimensionality, and MobileNetV3-Small for classification. The proposed approach was evaluated using the lung-PET-CT-Dx dataset, and achieved 98.6% accuracy with reduced computational requirements. They employed techniques such as quantization-aware training and early stopping to optimize the performance. Despite this performance, the authors declared that there is a lack of highly balanced datasets, and the dataset used was of low-quality images [43]. On the contrary, in [44], the authors developed Lung-EffNet, a transfer learning-based model using EfficientNet variants for accurate lung cancer classification from the IQ-OTH/NCCD dataset augmented using the GAN technique; the EfficientNetB1-based Lung-EffNet achieved 99.1% accuracy and a score of 97% to 99% of ROC. The model outperformed other CNNs, offering efficiency and lower training demands, making it a promising tool for clinical deployment. They also employed a Grad-cam to assess model interpretability. Despite this performance, the dataset used is very small, even with data augmentation; the model suffers from data dependency, which limits its generalizability, as shown in Table 1.
Despite the remarkable results achieved by previous studies, most of these models have been evaluated for their effectiveness using only a single dataset. This can lead to a certain level of data dependency, especially because these datasets are homogeneous, gathered from the same demographic sample, and use the same imaging equipment. Consequently, the models have a narrow range of characteristics they can identify, making them even more non-generalizable. However, most previous research has only tested such models on the available test samples from the same database and never assessed the models on external demographic factors to determine their generalizability. Furthermore, while deep learning models are effective and accurate, they are considered “black boxes”, as the processes leading to their classifications are uninterpretable. Owing to the existing circumstances, there is a feeling of uneasiness among practitioners, particularly considering the nature of the undertaking. Many previous studies have not been sufficient in this area of work, and a large gap remains. Table 1 summarizes the related work.

3. Materials and Methods

3.1. Problem Statement

Physicians and radiologists require a lung cancer detection system that can accurately screen patients in different cases [45] and minimize false positive and false negative rates [46]. In addition, the system should be sufficiently reliable to explain their results [47] as an additional support; hence, it functions as a dependable second opinion.

3.2. Research Objectives

This study aims to achieve a set of objectives, which are as follows:
  • To explore a new lung cancer CADx.
  • To reduce data dependency and bias by using five training datasets.
  • To improve the model’s robustness and generalization capability by employing a mixup augmentation technique and curriculum learning strategy.
  • To evaluate the model generalizability using an external dataset.
  • To enhance trustworthiness and interpretability.

3.3. Datasets and Preprocessing

It is worth mentioning that all datasets used in this work are publicly available. We used a subset of the LIDC-IDRI dataset along with four other datasets for training and validation of the model and an internal test [29,48,49,50,51]. The preprocessing included converting images to grayscale and resizing them to 224 × 224-pixel dimensions to fit the model input shape and reduce image complexity while keeping important information. The first three datasets were subjected to primary data augmentation using oversampling to eliminate class bias and were then split into training and validation datasets. Downsampling was used for the fourth and fifth datasets, with the data being split into training, validation, and testing datasets. This ensured that the model in the internal test would not view the test scans during training or validation. The resulting splits were as follows: 9328, 492, and 238 images were used for training from the first three datasets. For validation, we used 2488 images from the first dataset and 492 and 238 images from the other two datasets. For the fourth and fifth datasets, 160 and 126 images were used for training, 18 and 126 for validation, and 162 were reserved for testing from these two datasets, as shown in Figure 2.
To further evaluate the generalizability of the model, we tested it on a completely unseen external dataset, the Lung Cancer CT (LDCT) dataset [52].

3.4. The Utilized Techniques and Proposed Methodology

3.4.1. MobileNetV3 Small

MobileNetV3 Small is a lightweight CNN that is designed for resource-efficient mobile applications. It incorporates concepts from MobileNetV2, such as depthwise separable convolutions, while introducing new elements, such as squeeze-and-excitation modules and neural architecture search to optimize the network’s structure. This model is ideal for achieving a balance between performance and computational efficiency in tasks like image classification and object detection on devices with limited resources [53].

3.4.2. Resnet50

ResNet50 is a deep residual network architecture comprising 50 layers, and it is well known for introducing residual learning, which helps in training very deep networks by mitigating the vanishing gradient problem. The model uses identity mappings to add shortcut connections between layers, facilitating efficient backpropagation and improving model convergence. ResNet-50 is widely used in various computer vision tasks, such as image recognition, object detection, and feature extraction, owing to its high accuracy and performance on large datasets like ImageNet [54].

3.4.3. Classification Layers

The categorization step includes three dense layers (using ReLU activation) within ‘l2’ normalization, each one of them followed by batch normalization, then dropout (20%) to avoid overfitting, and finally, the output layer with one unit and using the sigmoid function. Figure 3 illustrates the feature extraction stage using both MobileNetV3Small and ResNet50 and the classification phase using dense layers.

3.4.4. Mixup Methods

Mixup is a data augmentation technique that blends pairs of input data and their labels to create new synthetic samples. This reduces overfitting, improves model robustness, and enhances generalization to unseen data by producing smoother decision boundaries. Mixup has been effective across various domains, boosting model performance in tasks such as image recognition and natural language processing [55]. Figure 4 represents the application of mixup to random images from the used dataset.

3.4.5. Curriculum Learning

Curriculum learning is a training strategy in which the model is gradually exposed to increasingly complex data, mimicking how humans learn. The approach begins with simpler examples and progressively introduces harder examples, allowing the model to develop more reliable and stable representations. It has been applied to fields such as language modeling, image recognition, and reinforcement learning, demonstrating improvements in training speed, convergence, and overall performance [56].

3.4.6. Grad-CAM

Gradient-weighted Class Activation Mapping (Grad-CAM) was proposed by Selvaraju et al. to produce visual explanations of the decision-making process in convolutional neural networks. It leverages the gradient information from the last convolutional layer, which typically strikes the best balance between high-level semantics and detailed spatial features. The importance of each neuron in the model’s decision making is assessed. Grad-CAM generates a heatmap that highlights the regions of the input image that most influence the model’s prediction, helping to make the typically opaque nature of CNNs more understandable and interpretable [57].

3.5. The Proposed Methodology

The solution proposed in this work can be used to diagnose lung cancer with high efficiency and accuracy and with considerable generalizability. It employs several distinct CT datasets, making it accessible to the model to eliminate data dependence and bias. This strategy presents a hybrid feature extraction model that integrates MobileNetV3Small and ResNet50. In addition, data augmentation is employed using a mixup, thereby making the model more robust and improving the generalization performance. Curriculum learning mimics the human learning process by training the model using simple and complex patterns. Finally, the Grad-CAM technique is applied to the proposed model to improve its comprehensibility and reliability, as shown in Figure 5.

3.6. Pseudocode for Mixup Augmentation and Training Phase Using Curriculum Learning

To enhance the reusability of the model, this section includes the pseudocode of the three main techniques used in the proposed model.
  • Mixup augmentation (Algorithm 1): This method generates an image and labels forged together by combining two images and their respective labels using a random interpolation factor (λ) sampled from a Beta distribution.
  • Mixup data generator (Algorithm 2): Augmented data batches are constructed by mixing augmented input sample batches, which creates an infinite variety of fresh augmented datasets in training.
  • Training with mixup augmentation (Algorithm 3): In the training phase, curriculum earning was adopted. During each of the five phases, data and labels were processed to create augmented batches. The model was trained for 100 epochs per phase, with callbacks such as learning rate reduction and early stopping. The training procedure was monitored by updating the total number of epochs and providing training history records.
Algorithm 1: The pseudocode of Mixup Augmentation Method (MixupAug)
(1) procedure MixupAugmentation(image1, image2, label1, label2, alpha)
(2) λ ⟵ RandomBeta(alpha, alpha) // generate random lambda parameter
(3) MixedImage ⟵ λ * image1 + (1 - λ) * image2 // perform mixup on images
(4) MixedLabel ⟵ λ * label1 + (1 - λ) * label2 // perform mixup on labels
(5) Result ⟵ Clip(MixedLabel, 0, 1) // clip mixed label values
(6) Return(MixedImage, Result) // return mixed image and label end procedure
Algorithm 2: The pseudocode of Mixup Data Generator (MixupGen)
(1) procedure MixupDataGen(x_data, y_data, batch_size, alpha)
(2)   While(True) // start infinite loop for generating batches
(3)      Shuffle(indices) // shuffle the data indices
(4)      For(i ϵ Range(0, len(x_data), batch_size)) // iterate over data batches
(5)            x_batch, y_batch ⟵ CurrentBatch(x_data, y_data, indices, i,
                    batch_size) // get the current batch
(6)            AugBatch ⟵ ApplyMixupToBatch(x_batch, y_batch, alpha) //             
                   apply mixup augmentation
(7)            Shuffle(AugBatch) // shuffle the augmented batch
(8)            Return(AugBatch) // return the augmented batch end procedure
Algorithm 3: Training Model with Mixup Augmentation (TrainMixup)
(1) procedure TrainMixup(datasets, model, batch_size, epochs, callbacks)
(2)       For (x_data, y_data, phase_name) ϵ datasets // iterate through datasets
(3)            TrainGen ⟵ MixupDataGen(x_data, y_data, batch_size,
                   alpha=0.01) // initialize mixup generator
(4)            Steps ⟵ ComputeSteps(len(x_data)) // compute steps per epoch
(5)           TrainModel(model, TrainGen, epochs, Steps, callbacks)// train model  

4. Results and Discussion

The training of the model was performed using five different datasets, beginning with a high-resolution dataset (LIDC-IDRI), and the remaining four datasets were given to the model in order. In addition, the Adam optimizer was used with a learning rate of 0.01 and a batch size of 32 suitable for all five datasets, and ReduceLROnPlateau was used to reduce the learning rate when the performance of the model did not increase for a while.
The proposed approach achieved comparable results. Figure 6 and Figure 7 show the accuracy and loss curves during training and validation. The mixup technique made the model somewhat heavy. It was noted that when the model was supplied with a new dataset each time, the accuracy decreased by approximately 40% because the new samples were unfamiliar to the model. However, the model adapted quickly and improved its performance. The same was observed for the loss, where only a very small change occurred, which was soon corrected. Moreover, the model achieved excellent results during testing, with an accuracy of 99.38%, precision and specificity of 100%, sensitivity of 98.76%, an F1-score of 99.37%, and 100% AUC and ROC. The analysis in Table 2 and Figure 8 shows the confusion matrix of the proposed model for the testing stage and proves that its performance surpasses the majority of the models available in the literature, with the exception of the model of Gopinath et al., which achieved an accuracy of 99.89%, and F1-score, sensitivity, and precision of 99.8% using DCNN. However, the authors used only one dataset, the LIDC-IDRI dataset, without any augmentation. Therefore, the model could be considered biased toward this dataset, particularly because it has a high resolution and is homogeneous [38]. In addition, the proposed model by Lanjewar et al. stated that their proposed model reached 100% accuracy with an AUC of 99.25%. Despite this high performance, the results obtained in the model testing only included data from the same demographic sample, meaning that the studied features were similar. This creates a problem and challenge in terms of generalizability [41].
However, it should be mentioned that such models have been trained and evaluated using only a single dataset, which introduces some bias and data dependence. On the contrary, the proposed model in this paper performed well not only on the internal training and testing datasets but also on entirely different external datasets, shown in Table 2 and Figure 9, where 100% was registered across all metrics, as well as 0% false negative and false negative rates, indicating excellent generalization potential. Furthermore, as shown in Figure 10 and Figure 11, the performance of the assessed model was consistent across internal and external datasets. However, additional tests should be performed using additional external datasets that vary in nature. Additionally, the complexity of the model during training and validation still requires further experimentation to enhance the speed while lowering the complexity. Other methods, such as CutMix or Cutout, may be more effective.
To improve the model’s reliability and interpretability, the Grad-CAM technique was used with an alpha value of 0.1, as shown in Figure 12. Eight images were randomly selected to visualize the features and areas the model focused on when making its classifications. In one instance, the model incorrectly classified an image as benign, as shown in Figure 12a; the model’s attention was primarily on the outer wall of the lungs, excluding the tumor region. In contrast, the model correctly classified the remaining images, concentrating on specific areas along the lung wall and on impurities within the images, as highlighted by the color intensities.

5. Conclusions

This study provides a model for diagnosing lung cancer that is interpretable, robust, and accurate with good generalization. The proposed model used a hybrid combination.
MobileNetV3Small and ResNet50 models were used for feature extraction and attention was injected into the dense layers for classification. To enhance the strength of the model, we used mixup augmentation based on the idea of overlaying the two images and their labels. Furthermore, the model adopted a curriculum learning paradigm during training for all five datasets. The results were excellent, with 99.08% accuracy, 0% false positive rate, 1.23% false negative rate, and 100% in ROC, AUC, precision, and specificity scores. An entirely external dataset was used to assess the model’s generalization ability, achieving 100% across all metrics with no false positives or negatives. Furthermore, GRAM-CAM was applied to enhance the model’s trustworthiness and comprehensibility. Although the results obtained are commendable, additional experiments are still needed. In the future, we will seek to experiment with different augmentation techniques, such as CutMix, Cutout, or RandAugment, which could improve the model’s performance and generalization ability. In addition, testing the proposed model on a broader range of external datasets with diverse demographic characteristics is crucial. The model complexity experienced during the training phase remains a myth that must be alleviated by optimizing the model, which will also be explored. Moreover, including multi-modal data in a curriculum learning approach could allow for stepwise learning of different imaging datasets. Therefore, the generalization of the model trained on different datasets from various imaging modalities will be explored.

Author Contributions

A.B. (Amira Bouamrane) and M.D. have conceived and designed the work, and was responsible for the writing—original draft. A.B. (Akram Bennour) and M.A.-S. have contributed to the supervision, data curation, writing—original draft, and validation. N.A.K. has contributed to the supervision, software, visualization, and validation. T.A.E.E. and A.-H.M.E. have carried out the critical revision of the article, visualization, validation, and the acquisition of funds, and the data curation. All authors reviewed the finalized manuscript and were responsible for the writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

The Deanship of Research and Graduate Studies at King Khalid University funded this research through a Large Research Project under grant number RGP2/259/45.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

The data supporting the findings of this study are based on publicly available datasets in the following repositories:

Acknowledgments

The authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Research Project under grant number RGP2/259/45.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kadhim, Y.A.; Khan, M.U.; Mishra, A. Deep learning-based computer-aided diagnosis (cad): Applications for medical image datasets. Sensors 2022, 22, 8999. [Google Scholar] [CrossRef] [PubMed]
  2. Cellina, M.; Cacioppa, L.M.; Cè, M.; Chiarpenello, V.; Costa, M.; Vincenzo, Z.; Pais, D.; Bausano, M.V.; Rossini, N.; Bruno, A.; et al. Artificial intelligence in lung cancer screening: The future is now. Cancers 2023, 15, 4344. [Google Scholar] [CrossRef] [PubMed]
  3. Dizon, D.S.; Kamal, A.H. Cancer statistics 2024: All hands on deck. CA Cancer J. Clin. 2024, 74, 12–49. [Google Scholar] [CrossRef] [PubMed]
  4. Yang, C.Y.; Lin, Y.T.; Lin, L.J.; Chang, Y.H.; Chen, H.Y.; Wang, Y.P.; Shih, J.Y.; Yu, C.J.; Yang, P.C. Stage shift improves lung cancer survival: Real-world evidence. J. Thorac. Oncol. 2023, 18, 47–56. [Google Scholar] [CrossRef]
  5. Suryani, A.I.; Chang, C.W.; Feng, Y.F.; Lin, T.K.; Lin, C.W.; Cheng, J.C.; Chang, C.Y. Lung tumor localization and visualization in chest X-ray images using deep fusion network and class activation mapping. IEEE Access 2022, 10, 124448–124463. [Google Scholar] [CrossRef]
  6. Krupinski, E.A.; Kallergi, M. Choosing a radiology workstation: Technical and clinical considerations. Radiology 2007, 242, 671–682. [Google Scholar] [CrossRef]
  7. Lau, S.C.; Pan, Y.; Velcheti, V.; Wong, K.K. Squamous cell lung cancer: Current landscape and future therapeutic options. Cancer Cell 2022, 40, 1279–1293. [Google Scholar] [CrossRef]
  8. Nooreldeen, R.; Bach, H. Current and future development in lung cancer diagnosis. Int. J. Mol. Sci. 2021, 22, 8661. [Google Scholar] [CrossRef]
  9. Wu, S.; Pan, Y.; Mao, Y.; Chen, Y.; He, Y. Current progress and mechanisms of bone metastasis in lung cancer: A narrative review. Transl. Lung Cancer Res. 2021, 10, 439. [Google Scholar] [CrossRef]
  10. Zheng, H.; Zhang, Q.; Gong, Y.; Liu, Z.; Chen, S. Identification of prognostic biomarkers for stage iii non-small cell lung carcinoma in female nonsmokers using machine learning. arXiv 2024, arXiv:2408.16068. [Google Scholar]
  11. Lam, S.; Bai, C.; Baldwin, D.R.; Chen, Y.; Connolly, C.; de Koning, H.; Heuvelmans, M.A.; Hu, P.; Kazerooni, E.A.; Lancaster, H.L.; et al. Current and future perspectives on computed tomography screening for lung cancer: A roadmap from 2023 to 2027 from the International Association for the Study of Lung Cancer. J. Thorac. Oncol. 2024, 19, 36–51. [Google Scholar] [CrossRef] [PubMed]
  12. Withers, P.J.; Bouman, C.; Carmignato, S.; Cnudde, V.; Grimaldi, D.; Hagen, C.K.; Maire, E.; Manley, M.; Du Plessis, A.; Stock, S.R. X-ray computed tomography. Nat. Rev. Methods Primers 2021, 1, 18. [Google Scholar] [CrossRef]
  13. Sartorio, C.; Milanese, G.; Ledda, R.E.; Tringali, G.; Balbi, M.; Milone, F.; Sverzellati, N.; Silva, M. Diameter versus volumetry: A narrative review on current recommendations to measure and monitor screening detected lung nodules. Shanghai Chest 2021, 5. [Google Scholar] [CrossRef]
  14. Sim, A.J.; Kaza, E.; Singer, L.; Rosenberg, S.A. A review of the role of MRI in diagnosis and treatment of early stage lung cancer. Clin. Transl. Radiat. Oncol. 2020, 24, 16–22. [Google Scholar] [CrossRef]
  15. Yoshida, H.; Dachman, A. CAD techniques, challenges, andcontroversies in computed tomographic colonography. Abdom. Imaging 2004, 30, 26–41. [Google Scholar] [CrossRef]
  16. El-Baz, A.; Beache, G.M.; Gimel farb, G.; Suzuki, K.; Okada, K.; Elnakib, A.; Soliman, A.; Abdollahi, B. Computer-aided diagnosis systems for lung cancer: Challenges and methodologies. Int. J. Biomed. Imaging 2013, 2013, 942353. [Google Scholar] [CrossRef]
  17. Bouamrane, A.; Derdour, M.; Alksas, A.; El-Baz, A. Hybrid Deep Learning Approach with Feature Engineering for Enhanced Pulmonary Nodule Diagnosis. SN Comput. Sci. 2024, 5, 890. [Google Scholar] [CrossRef]
  18. Kanwal, N.; Pérez-Bueno, F.; Schmidt, A.; Engan, K.; Molina, R. The devil is in the details: Whole slide image acquisition and processing for artifacts detection, color variation, and data augmentation: A review. IEEE Access 2022, 10, 58821–58844. [Google Scholar] [CrossRef]
  19. Kalantar, R.; Lin, G.; Winfield, J.M.; Messiou, C.; Lalondrelle, S.; Blackledge, M.D.; Koh, D.M. Automatic segmentation of pelvic cancers using deep learning: State-of-the-art approaches and challenges. Diagnostics 2021, 11, 1964. [Google Scholar] [CrossRef]
  20. You, Y.; Bai, Z.; Zhang, Y.; Li, Z. Contour-induced parallel graph reasoning for liver tumor segmentation. Biomed. Signal Process. Control. 2024, 92, 106111. [Google Scholar] [CrossRef]
  21. Lee, A.L.S.; To, C.C.K.; Lee, A.L.H.; Li, J.J.X.; Chan, R.C.K. Model architecture and tile size selection for convolutional neural network training for non-small cell lung cancer detection on whole slide images. Inform. Med. Unlocked 2022, 28, 100850. [Google Scholar] [CrossRef]
  22. Bouamrane, A.; Derdour, M. Enhancing lung cancer detection and classification using machine learning and deep learning techniques: A comparative study. In Proceedings of the 2023 International Conference on Networking and Advanced Systems (ICNAS), Algiers, Algeria, 21–23 October 2023; IEEE: New York, NY, USA, 2023; pp. 1–6. [Google Scholar]
  23. Bouamrane, A.; Derdour, M.; Alksas, A.; El-Baz, A. A Transfer Learning Framework for Lung Cancer Classification Using EfficientV2-L: Generalizability Assessment. In Proceedings of the 2024 6th International Conference on Pattern Analysis and Intelligent Systems (PAIS), EL OUED, Algeria, 24–25 April 2024; IEEE: New York, NY, USA, 2024; pp. 1–6. [Google Scholar]
  24. Huang, X.; Lei, Q.; Xie, T.; Zhang, Y.; Hu, Z.; Zhou, Q. Deep transfer convolutional neural network and extreme learning machine for lung nodule diagnosis on CT images. Knowl.-Based Syst. 2020, 204, 106230. [Google Scholar] [CrossRef]
  25. Nishio, M.; Sugiyama, O.; Yakami, M.; Ueno, S.; Kubo, T.; Kuroda, T.; Togashi, K. Computer-aided diagnosis of lung nodule classification between benign nodule, primary lung cancer, and metastatic lung cancer at different image size using deep convolutional neural network with transfer learning. PLoS ONE 2018, 13, e0200721. [Google Scholar] [CrossRef]
  26. Khan, M.A.; Rajinikanth, V.; Satapathy, S.C.; Taniar, D.; Mohanty, J.R.; Tariq, U.; Damaševicius, R. VGG19 network assisted joint segmentation and classification of lung nodules in CT images. Diagnostics 2021, 11, 2208. [Google Scholar] [CrossRef]
  27. Bilal, A.; Shafiq, M.; Fang, F.; Waqar, M.; Ullah, I.; Ghadi, Y.Y.; Long, H.; Zeng, R. IGWO-IVNet3: DL-based automatic diagnosis of lung nodules using an improved gray wolf optimization and InceptionNet-V3. Sensors 2022, 22, 9603. [Google Scholar] [CrossRef]
  28. Shafi, I.; Din, S.; Khan, A.; Díez, I.D.L.T.; Casanova, R.d.J.P.; Pifarre, K.T.; Ashraf, I. An effective method for lung cancer diagnosis from ct scan using deep learning-based support vector network. Cancers 2022, 14, 5457. [Google Scholar] [CrossRef]
  29. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef]
  30. Armato, S.G., III; Meyer, C.R.; McNitt-Gray, M.F.; McLennan, G.; Reeves, A.; Croft, B.Y.; Clarke, L.P.; Group, R.R. The Reference Image Database to Evaluate Response to therapy in lung cancer (RIDER) project: A resource for the development of change- analysis software. Clin. Pharmacol. Ther. 2008, 84, 448–456. [Google Scholar] [CrossRef]
  31. Ho, T.K.K.; Gwak, J.; Prakash, O.; Song, J.I.; Park, C.M. Utilizing pretrained deep learning models for automated pulmonary tuberculosis detection using chest radiography. In Proceedings of the Intelligent Information and Database Systems: 11th Asian Conference, ACIIDS 2019, Yogyakarta, Indonesia, 8–11 April 2019; Proceedings, Part II 11. Springer: Berlin/Heidelberg, Germany, 2019; pp. 395–403. [Google Scholar]
  32. Al-Yasriy, H.F.; Al-Husieny, M.S.; Mohsen, F.Y.; Khalil, E.A.; Hassan, Z.S. Diagnosis of lung cancer based on CT scans using CNN. In Proceedings of the IOP Conference Series: Materials Science and Engineering, Thi-Qar, Iraq, 15–16 July 2020; IOP Publishing: Bristol, UK, 2020; Volume 928, p. 022035. [Google Scholar]
  33. Setio, A.A.A.; Traverso, A.; De Bel, T.; Berens, M.S.; Van Den Bogaard, C.; Cerello, P.; Chen, H.; Dou, Q.; Fantacci, M.E.; Geurts, B.; et al. Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: The LUNA16 challenge. Med. Image Anal. 2017, 42, 1–13. [Google Scholar] [CrossRef]
  34. Liu, X.; Li, K.W.; Yang, R.; Geng, L.S. Review of deep learning based automatic segmentation for lung cancer radiotherapy. Front. Oncol. 2021, 11, 717039. [Google Scholar] [CrossRef]
  35. Mastouri, R.; Khlifa, N.; Neji, H.; Hantous-Zannad, S. Deep learning-based CAD schemes for the detection and classification of lung nodules from CT images: A survey. J. X-Ray Sci. Technol. 2020, 28, 591–617. [Google Scholar] [CrossRef] [PubMed]
  36. Chenyang, L.; Chan, S.C. A joint detection and recognition approach to lung cancer diagnosis from CT images with label uncertainty. IEEE Access 2020, 8, 228905–228921. [Google Scholar] [CrossRef]
  37. Zhao, X.; Xu, J.; Lin, Z.; Xue, X. BiCFormer: Swin Transformer based model for classification of benign and malignant pulmonary nodules. Meas. Sci. Technol. 2024, 35, 075402. [Google Scholar] [CrossRef]
  38. Gopinath, A.; Gowthaman, P.; Venkatachalam, M.; Saroja, M. Computer aided model for lung cancer classification using cat optimized convolutional neural networks. Meas. Sens. 2023, 30, 100932. [Google Scholar] [CrossRef]
  39. Saied, M.; Raafat, M.; Yehia, S.; Khalil, M.M. Efficient pulmonary nodules classification using radiomics and different artificial intelligence strategies. Insights Imaging 2023, 14, 91. [Google Scholar] [CrossRef]
  40. Meng, L.; Zhu, P.; Xia, K. Application value of the automated machine learning model based on modified CT index combined with serological indices in the early prediction of lung cancer. Front. Public Health 2024, 12, 1368217. [Google Scholar] [CrossRef]
  41. Lanjewar, M.G.; Panchbhai, K.G.; Charanarur, P. Lung cancer detection from CT scans using modified DenseNet with feature selection methods and ML classifiers. Expert Syst. Appl. 2023, 224, 119961. [Google Scholar] [CrossRef]
  42. Shen, Z.; Cao, P.; Yang, J.; Zaiane, O.R. WS-LungNet: A two-stage weakly-supervised lung cancer detection and diagnosis network. Comput. Biol. Med. 2023, 154, 106587. [Google Scholar] [CrossRef]
  43. Wahab Sait, A.R. Lung Cancer Detection Model Using Deep Learning Technique. Appl. Sci. 2023, 13, 12510. [Google Scholar] [CrossRef]
  44. Raza, R.; Zulfiqar, F.; Khan, M.O.; Arif, M.; Alvi, A.; Iftikhar, M.A.; Alam, T. Lung-EffNet: Lung cancer classification using EfficientNet from CT-scan images. Eng. Appl. Artif. Intell. 2023, 126, 106902. [Google Scholar] [CrossRef]
  45. Porto-Álvarez, J.; Barnes, G.T.; Villanueva, A.; García-Figueiras, R.; Baleato-González, S.; Huelga Zapico, E.; Souto-Bayarri, M. Digital medical x-ray imaging, cad in lung cancer and radiomics in colorectal cancer: Past, present and future. Appl. Sci. 2023, 13, 2218. [Google Scholar] [CrossRef]
  46. de Margerie-Mellon, C.; Chassagnon, G. Artificial intelligence: A critical review of applications for lung nodule and lung cancer. Diagn. Interv. Imaging 2023, 104, 11–17. [Google Scholar] [CrossRef] [PubMed]
  47. Najjar, R. Redefining radiology: A review of artificial intelligence integration in medical imaging. Diagnostics 2023, 13, 2760. [Google Scholar] [CrossRef] [PubMed]
  48. Armato III, S.G.; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Zhao, B.; Aberle, D.R.; Henschke, C.I.; Hoffman, E.A.; et al. The lung image database consortium (LIDC) and image database resource initiative (IDRI): A completed reference database of lung nodules on CT scans. Med. Phys. 2011, 38, 915–931. [Google Scholar] [CrossRef] [PubMed]
  49. Rathi, D. CT Scan Images for Lung Cancer. 2023. Available online: https://www.kaggle.com/datasets/dishantrathi20/ct-scan-images-for-lung-cancer (accessed on 14 October 2024).
  50. JAYAPRAKASHPONDY. Lung Cancer Dataset. 2023. Available online: https://www.kaggle.com/datasets/jayaprakashpondy/lung-cancer-dataset (accessed on 14 October 2024).
  51. Hany, M. Chest CT-Scan images Dataset. 2020. Available online: https://www.kaggle.com/datasets/mohamedhanyyy/chest-ctscan-images (accessed on 14 October 2024).
  52. Dharpure, H. DLCTlUNGDetectNet-Lung Tumor Dataset. 2023. Available online: https://www.kaggle.com/datasets/harshaldharpure/dlctlungdetectnet-lung-tumor-dataset (accessed on 14 October 2024).
  53. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  54. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  55. Zhang, H.; Cisse, M.; Dauphin, Y.; Lopez-Paz, D. mixup: Beyond empirical risk management. In Proceedings of the 6th International Conference Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–13. [Google Scholar]
  56. Bengio, Y.; Louradour, J.; Collobert, R.; Weston, J. Curriculum learning. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; pp. 41–48. [Google Scholar]
  57. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. CADx Phases.
Figure 1. CADx Phases.
Diagnostics 15 00001 g001
Figure 2. Dataset distribution: training and validation. (a) Training; (b) Validation.
Figure 2. Dataset distribution: training and validation. (a) Training; (b) Validation.
Diagnostics 15 00001 g002
Figure 3. Feature extraction and classification dense layers.
Figure 3. Feature extraction and classification dense layers.
Diagnostics 15 00001 g003
Figure 4. Visualization of mixup augmentation on image samples. The three subfigures (ac) represent the application of mixup augmentation. Each subfigure contains two images, both annotated with labels by radiologists. The third image in each subfigure shows the result of applying the mixup augmentation technique to these two images and their corresponding labels.
Figure 4. Visualization of mixup augmentation on image samples. The three subfigures (ac) represent the application of mixup augmentation. Each subfigure contains two images, both annotated with labels by radiologists. The third image in each subfigure shows the result of applying the mixup augmentation technique to these two images and their corresponding labels.
Diagnostics 15 00001 g004aDiagnostics 15 00001 g004b
Figure 5. Proposed approach: a simplified illustration of the model.
Figure 5. Proposed approach: a simplified illustration of the model.
Diagnostics 15 00001 g005
Figure 6. Training performance: loss and accuracy.
Figure 6. Training performance: loss and accuracy.
Diagnostics 15 00001 g006
Figure 7. Validation performance: loss and accuracy.
Figure 7. Validation performance: loss and accuracy.
Diagnostics 15 00001 g007
Figure 8. Confusion matrix: internal test.
Figure 8. Confusion matrix: internal test.
Diagnostics 15 00001 g008
Figure 9. Confusion matrix: external test.
Figure 9. Confusion matrix: external test.
Diagnostics 15 00001 g009
Figure 10. Testing of the model across internal and external datasets.
Figure 10. Testing of the model across internal and external datasets.
Diagnostics 15 00001 g010
Figure 11. Testing of the model across internal and external datasets: FP and FN.
Figure 11. Testing of the model across internal and external datasets: FP and FN.
Diagnostics 15 00001 g011
Figure 12. GRAD-CAM heatmaps: visualizing model attention in predictions. The eight subfigures (ah) represent the original image with its true label and the predicted label by the proposed model. The second part of each subfigure shows the Grad-CAM image, which highlights the important features that the model focuses on during prediction.
Figure 12. GRAD-CAM heatmaps: visualizing model attention in predictions. The eight subfigures (ah) represent the original image with its true label and the predicted label by the proposed model. The second part of each subfigure shows the Grad-CAM image, which highlights the important features that the model focuses on during prediction.
Diagnostics 15 00001 g012
Table 1. Related work comparison.
Table 1. Related work comparison.
#StudyYearApproachDatasetStrengthsLimitation
01Zhao et al. [37]2024BiCFormerLIDC-IDRIAccuracy = 97.4%Lack of interpretability, homogenous dataset, limited dataset size
02Meng et al. [40]2024Gradient Boosting MachineLocal hospital datasetsAccuracy = 99%,
AUC = 93.1%,
External validation:
Accuracy = 85.7%,
AUC = 95.5%
Homogenous dataset
03Gopinath et al. [38]2023DFF-CON using DCNNLIDC-IDRIAccuracy = 99.89%,
F1-score = 99.88,
Sensitivity = 99.8%,
Specificity = 99.76%,
Precision = 99.8%
Limited dataset size, homogeneity, potential for bias, limited reliability
04Saied et al. [39]2023DenseNet-121 and SVMLIDC-IDRIAccuracy = 90.39%,
Sensitivity = 90.32%,
Specificity = 93.65%
Small, homogenous dataset, potential overfitting, limited generalizability
05Lanjewar et al. [41]2023SVM, LR, RF, DT, GNB, KNNChest-CT Kaggle datasetAccuracy = 100%,
AUC = 99.25%,
Kappa = 93%
Limited generalizability, feature dependency, limited interpretability
06Wahab et al. [43]2023DenseNet-121 and MobileNetV3-SmallLung-PET-CT-Dx datasetAccuracy = 98.6%,
Precision = 97.9%,
Recall = 98.1%,
F1-Score = 98,
Kappa = 95.8%
Imbalanced dataset, limited interpretability
07Raza et al. [44]2023EfficientNetB1-based Lung-EffNetIQ-OTH/NCCDAccuracy = 99.10%,
Precision = 99.22%,
Recall = 97.22%,
F1score = 98.16%
Very small, homogenous dataset
08Shen et al. [42]2023WS-LungNetLIDC-IDRICPM = 82.99%,
AUC = 88.63%,
DROC = 87.12%
Lack of interpretability, small and homogenous dataset
Table 2. Model results across the internal and external datasets.
Table 2. Model results across the internal and external datasets.
Proposed ModelAccuracyPrecisionSpecificitySensitivityF1-ScoreAUCFalse PositiveFalse Negative
Internal test dataset99.38%100%100%98.76%99.37%100%00%1.23%
External Dataset100%100%100%100%100%100%00%00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bouamrane, A.; Derdour, M.; Bennour, A.; Elfadil Eisa, T.A.; M. Emara, A.-H.; Al-Sarem, M.; Kurdi, N.A. Toward Robust Lung Cancer Diagnosis: Integrating Multiple CT Datasets, Curriculum Learning, and Explainable AI. Diagnostics 2025, 15, 1. https://doi.org/10.3390/diagnostics15010001

AMA Style

Bouamrane A, Derdour M, Bennour A, Elfadil Eisa TA, M. Emara A-H, Al-Sarem M, Kurdi NA. Toward Robust Lung Cancer Diagnosis: Integrating Multiple CT Datasets, Curriculum Learning, and Explainable AI. Diagnostics. 2025; 15(1):1. https://doi.org/10.3390/diagnostics15010001

Chicago/Turabian Style

Bouamrane, Amira, Makhlouf Derdour, Akram Bennour, Taiseer Abdalla Elfadil Eisa, Abdel-Hamid M. Emara, Mohammed Al-Sarem, and Neesrin Ali Kurdi. 2025. "Toward Robust Lung Cancer Diagnosis: Integrating Multiple CT Datasets, Curriculum Learning, and Explainable AI" Diagnostics 15, no. 1: 1. https://doi.org/10.3390/diagnostics15010001

APA Style

Bouamrane, A., Derdour, M., Bennour, A., Elfadil Eisa, T. A., M. Emara, A. -H., Al-Sarem, M., & Kurdi, N. A. (2025). Toward Robust Lung Cancer Diagnosis: Integrating Multiple CT Datasets, Curriculum Learning, and Explainable AI. Diagnostics, 15(1), 1. https://doi.org/10.3390/diagnostics15010001

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop