Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Fetal Safety in MRI During Pregnancy: A Comprehensive Review
Previous Article in Journal
Diagnostic Potential of CTRP5 and Chemerin for Coronary Artery Disease: A Study by Coronary Computed Tomography Angiography
Previous Article in Special Issue
Can Artificial Intelligence Revolutionize the Diagnosis and Management of the Atrial Septal Defect in Children?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Transfer Learning for Classification of Late Gadolinium Enhancement Cardiac MRI Images into Myocardial Infarction, Myocarditis, and Healthy Classes: Comparison with Subjective Visual Evaluation

1
Technology and Medical Imaging Laboratory LR12ES06, Faculty of Medicine of Monastir, University of Monastir, Monastir 5019, Tunisia
2
Faculty of Sciences of Monastir, University of Monastir, Monastir 5019, Tunisia
3
LR18-SP08 Department of Radiology, University Hospital of Monastir, Monastir 5019, Tunisia
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(2), 207; https://doi.org/10.3390/diagnostics15020207
Submission received: 31 October 2024 / Revised: 25 December 2024 / Accepted: 30 December 2024 / Published: 17 January 2025
(This article belongs to the Special Issue Diagnostic AI and Cardiac Diseases)

Abstract

:
Background/Objectives: To develop a computer-aided diagnosis (CAD) method for the classification of late gadolinium enhancement (LGE) cardiac MRI images into myocardial infarction (MI), myocarditis, and healthy classes using a fine-tuned VGG16 model hybridized with multi-layer perceptron (MLP) (VGG16-MLP) and assess our model’s performance in comparison to various pre-trained base models and MRI readers. Methods: This study included 361 LGE images for MI, 222 for myocarditis, and 254 for the healthy class. The left ventricle was extracted automatically using a U-net segmentation model on LGE images. Fine-tuned VGG16 was performed for feature extraction. A spatial attention mechanism was implemented as a part of the neural network architecture. The MLP architecture was used for the classification. The evaluation metrics were calculated using a separate test set. To compare the VGG16 model’s performance in feature extraction, various pre-trained base models were evaluated: VGG19, DenseNet121, DenseNet201, MobileNet, InceptionV3, and InceptionResNetV2. The Support Vector Machine (SVM) classifier was evaluated and compared to MLP for the classification task. The performance of the VGG16-MLP model was compared with a subjective visual analysis conducted by two blinded independent readers. Results: The VGG16-MLP model allowed high-performance differentiation between MI, myocarditis, and healthy LGE cardiac MRI images. It outperformed the other tested models with 96% accuracy, 97% precision, 96% sensitivity, and 96% F1-score. Our model surpassed the accuracy of Reader 1 by 27% and Reader 2 by 17%. Conclusions: Our study demonstrated that the VGG16-MLP model permits accurate classification of MI, myocarditis, and healthy LGE cardiac MRI images and could be considered a reliable computer-aided diagnosis approach specifically for radiologists with limited experience in cardiovascular imaging.

1. Introduction

Cardiac magnetic resonance imaging (MRI) is an effective method for non-invasively assessing cardiovascular morphology, myocardial perfusion, ventricular function, and tissue characterization [1]. In cardiac MRI, late gadolinium enhancement (LGE) sequences offer a reliable diagnosis of myocardial infarction (MI) and myocarditis due to their ability to detect areas of myocardial necrosis and fibrosis with high spatial resolution [2]. In MI, LGE is typically subendocardial or transmural, located in a region corresponding to the perfusion area of a coronary artery [3]. Meanwhile, myocarditis is associated with epicardial and midmyocardial LGE patterns [4]. In clinical practice, LGE regions are assessed visually, potentially leading to inconsistencies within and between different observers [5]. Furthermore, fine subepicardial enhancement in myocarditis may be misinterpreted as epicardial fat, while fine subendocardial enhancement in MI could be mistaken for the ventricular cavity, which may result in the misclassification of pathological LGE images as normal. Moreover, in cases of quasi-transmural enhancement, distinguishing myocarditis from MI is particularly challenging. Such uncertainty is frequently noted among radiologists or cardiologists with limited experience in cardiovascular imaging. Recently, machine learning has proven to have significant potential in this field [6]. It has been recognized as effective in predicting and detecting various cardiovascular diseases [7,8,9]. According to a recent study, machine learning can help to differentiate between MI and myocarditis with an accuracy of 89% using LGE cardiac images [5]. Deep learning, an advanced subset of machine learning, reveals its potential in several fields, particularly in cardiovascular MRI by aiding clinicians in image analyses, image interpretation, and patient assessment [10]. It has the capability to extract relevant features from images, which is highly valuable to the analysis of medical images. Furthermore, in contrast to machine learning, which requires human intervention for feature extraction, deep learning offers the advantage of automatically identifying and extracting key features in an end-to-end process [11,12]. Deep learning includes a class of algorithms known as neural networks. Notably, convolutional neural networks (CNNs) are currently acknowledged as the state of the art for image prediction [13]. CNNs have been recently implemented to address radiological issues [14] and have demonstrated the ability to classify medical images at a level that rivals or even surpasses that of experts in various fields [15]. A study has demonstrated that CNN networks can effectively categorize subjects with normal, hypertrophic cardiomyopathy and dilated cardiomyopathy on cine MRI images [16]. Another study utilized deep learning for the binary classification of patients with and without cardiac amyloidosis based on clinical and imaging features extracted from LGE images [17]. One of the most popular architectures in CNNs is Visual Geometry Group-16 (VGG16). It demonstrates a considerable capacity for feature extraction, enabling it to achieve significant results in image classification [18]. To avoid inconsistencies between readers and to automate redundant tasks, our study aimed to develop a computer-aided diagnosis (CAD) approach for the classification of LGE cardiac MRI images into MI, myocarditis, and healthy classes using a fine-tuned VGG16 CNN model hybridized with multi-layer perceptron (MLP) (VGG16-MLP) and to assess our model’s performance in comparison to various pre-trained base models and cardiac MRI readers.

2. Materials and Methods

2.1. Study Population

This retrospective study included LGE images from 163 patients who underwent cardiac MRI between 2015 and 2022 in the Medical Imaging Department. Study selection was made by a 15-year-experienced radiologist specializing in cardiovascular imaging, evaluating and labeling LGE images based on their representative appearance as healthy, MI, and myocarditis. The dataset is described in Table 1.

2.2. MRI Acquisitions

All cardiac MRI acquisitions were performed on a Philips Ingenia 1.5 Tesla MRI (Philips Healthcare, Best, The Netherlands) using a torso phased-array coil coupled to a spine coil. A standard cardiac protocol was used, including cine steady-state free precession in short-axis, long-axis, and four-chamber views; a black blood short-time inversion recovery sequence; and an LGE sequence. LGE images were acquired 10 min after the administration of a 0.2 mmol/kg bolus of adulterate dimeglumine (Dotarem, Guerbet, Roissy, France). Short-axis sections of LGE were obtained with a 15 s breath-hold using the following parameters: flip angle, 25°; repetition time, 6 ms; echo time, 3 ms; field of view, 25 cm; slice thickness, 10 mm.

2.3. Image Preparation

2.3.1. Extraction of the Region of Interest (ROI): Left Ventricle Segmentation

A segmentation of the left ventricle, which includes the left myocardium and left ventricle cavity, was performed using a deep learning method developed in our laboratory [19]. The procedure is described in Figure 1.

Image Preprocessing for Segmentation

The left ventricle in LGE images was manually segmented to create the ground truth. Four operations were conducted on cardiac LGE images and their corresponding ground truths to enhance image contrast: smoothing, gamma correction, intensity normalization, and histogram equalization. A patch extraction was applied, dividing the image into small patches to increase the effective number of training data. In total, 20% of the data were used for the test.

Concat-U-Net Segmentation

For the deep learning segmentation process, Concat-U-Net was employed, harnessing the strengths of three distinct architectures: U-Net, ResNet, and DenseNet [19]. This architecture is similar to U-Net. However, it has the advantage of using fewer parameters to reduce the risk of overfitting.

Superposition

The mask was superposed to the original image to obtain an image of the ROI.

2.3.2. Image Resizing

The images were adjusted from the size of (320 × 320) to (224 × 224) in order to meet the requirements of the model.

2.3.3. Data Augmentation

To increase the diversity of the training dataset, geometric data augmentation was applied during training, including a 10% horizontal shift range and a 15% zoom range.

2.4. Deep Learning Classification

2.4.1. Model Description

Our model combined a fine-tuned VGG16 model with MLP. For feature extraction from input images, an ImageNet-pre-trained VGG16 model was performed. All layers in the base model were frozen except for the last 3 layers. This means that the weights of most layers in the base model were not be updated during training. The VGG16 model consists of five blocks of convolutional layers with small 3 × 3 filters, each followed by a max-pooling layer to reduce the spatial dimension. A spatial attention mechanism was implemented to enhance the model’s ability to focus on relevant features. This was achieved using two successive DepthwiseConv2D layers with 7 × 7 and 5 × 5 kernels, capturing multiscale spatial information while significantly reducing the number of parameters and computational cost. These layers were followed by two standard Conv2D layers with ReLU and Sigmoid activation functions, respectively. ReLU was used after the first Conv2D layer to introduce non-linearity while preserving and propagating important spatial features, while Sigmoid activation was applied after the second Conv2D layer to generate attention weights between 0 and 1. This combination enables the model to extract meaningful spatial features and create a normalized attention map that determines which spatial regions to emphasize or suppress when multiplied by the original features. Then, a GlobalAveragePooling2D layer was used to reduce the spatial dimensions of the feature maps into a single vector, which served as input for the MLP. The MLP architecture used for the classification task included five dense layers integrated with dropout for regularization, which formed the network’s latter part, with a final dense layer producing a 3-unit output. The architecture of our model VGG16-MLP is presented in Figure 2.

2.4.2. Model Assessment: Training, Validation, and Test

The training dataset consisted of 669 images (80% of the dataset). After data augmentation using 10% horizontal shift range and 15% zoom range, 720 images were generated, resulting in 1389 images used for training. A 5-fold cross-validation technique was used to avoid overfitting and to assess the model’s performance by evaluating it on five random subsets of the data (80% training, 20% validation). A holdout test set with 168 images was used for the test. The majority voting technique was used to identify the most frequently predicted class from the five folds, which was then selected as the final prediction. The batch size, indicating the number of training examples utilized in each training iteration, was set to 60. The number of epochs, representing how many times the complete training dataset would traverse the model during training, was set to 60. The learning rate was 0.0001.

2.4.3. Evaluation Metrics

The performance metrics reported in this study were obtained using the hold-out test set. Four metrics were used for the evaluation of segmentation: the dice coefficient, which measures the similarity between the predicted segmented regions generated by the algorithm and the actual ground truth segmentation; the accuracy (1); the precision (2); and the sensitivity (3). The performance of a classification model is assessed using various metrics. The most widely used metrics are accuracy (1), precision (2), sensitivity (3), and F1-score (4) [5,20,21,22,23,24]. In our study, we employed these metrics to assess the classification task.
Accuracy = Number of Correctly Predicted Instances/Total Number of Instances
Precision = True Positives/(True Positives + False Positives)
Sensitivity = True Positives/(True Positives + False Negatives)
F1-score = 2 × (Precision×sensitivity)/(Precision + sensitivity)
A confusion matrix was used to represent the performance of the tested models. To illustrate the performance of the three classes on the VGG16-MLP model, the receiver operating characteristic (ROC) curve, representing the true-positive rate as a function of the false-positive rate, and the precision–recall curve were employed.

2.4.4. Computational Setting

The experiments were conducted on 16 GB memory and a single NVIDIA GeForce GTX 1650 GPU. The CNN model was generated using the Keras library (https://keras.io accessed on 11 April 2024). The TensorFlow backend was used in this process (https://www.tensorflow.org accessed on 10 April 2024). Numpy library was applied for numerical computing (https://numpy.org accessed on 11 April 2024). OpenCV was employed for preprocessing and image analysis (https://opencv.org accessed on 11 April 2024) and Matplotlib was utilized for data visualization (https://matplotlib.org accessed on 11 April 2024).

2.5. Evaluation of Comparative Models

To compare the VGG16-MLP model’s performance in feature extraction, various pre-trained base models integrating a spatial attention mechanism with the two Depthwiseconv2D layers and combined with MLP were evaluated: VGG19-MLP, DenseNet121-MLP, DenseNet201-MLP, MobileNet-MLP, InceptionV3-MLP, and InceptionResNetV2-MLP. For the classification, the Support Vector Machine (SVM) classifier combined with the fine-tuned VGG16 model (VGG16-SVM) was evaluated and compared to VGG16-MLP.

2.6. Contribution Evaluation of Preprocessing, Spatial Attention Mechanism, and Depthwiseconv2D Layers

For VGG16-MLP model, the contributions of image segmentation, data augmentation, spatial attention mechanism, and Depthwiseconv2D layers were evaluated.

2.7. Visual Reader Analysis

A subjective visual analysis was performed by two blinded independent readers to classify healthy, MI, and myocarditis LGE images. The first reader was a radiology resident with 4 years of experience, and the second was a radiology resident with 3 years of experience.

2.8. Statistical Analysis

To describe the study population, the two-sample t-test was used to compare the age distribution between the different classes, and the chi-square test was used for the sex distribution. A Wilcoxon signed-rank test was used to compare the performance of VGG16-MLP with the other tested models, based on the accuracy values of the 5-fold cross-validation. The significance level was set at p < 0.05 for all tests. IBM SPSS version 26.0 was used for statistical analysis.

3. Results

3.1. Left Ventricle Segmentation

For the segmentation task, the dice coefficient was 92%, the accuracy was 99%, the sensitivity was 93%, and the precision was 92% (Figure 3 and Figure 4).

3.2. Performance of Models on Segmented Images

The accuracy, precision, sensitivity, and F1-score of VGG16-MLP, VGG19-MLP, DenseNet121-MLP, DenseNet201-MLP, MobileNet-MLP, InceptionV3-MLP, InceptionResNetV2-MLP, and VGG16-SVM are presented in Table 2. The VGG16-MLP model achieved the highest accuracy at 96%, precision at 97%, sensitivity at 96%, and F1-score at 96%. Most models demonstrated a statistically significant difference compared to VGG16-MLP (Table 2). The confusion matrices of tested models are presented in Figure 5. The training and inference times of the tested models (Table 2) demonstrated promising results, highlighting computational performance and feasibility for practical implementation.

VGG16-MLP Performance

The VGG16-MLP model metrics for the MI, myocarditis, and healthy classes are presented in Table 3 and Figure 6. The accuracy reached 97% for MI, 98% for myocarditis, and 98% for the healthy class. The precision was 95% for MI, 100% for myocarditis, and 96% for the healthy class. The sensitivity was 99% for MI, 91% for myocarditis, and 98% for the healthy class. The F1-score achieved 97% for MI, 95% for myocarditis, and 97% for the healthy class. The confusion matrix is presented in Figure 5. The ROC curves (Figure 7) show the high performance of the three classes, as true positives are close to one and false positives are close to zero. The areas under the curve (AUCs) for the MI, myocarditis, and healthy classes were 1.00, 0.99, and 1.00, respectively. These values were superior to 0.9, which means that the diagnostic performance of the three classes was considered excellent [25]. The precision–recall curves (Figure 7) demonstrate high values of precision and recall for the three classes. The AUCs for the MI, myocarditis, and healthy classes were 1.00, 0.98, and 0.99, respectively.

3.3. Contribution Evaluation of Preprocessing, Spatial Attention Mechanism, and Depthwiseconv2D Layers

3.3.1. Effect of Extracting the Region of Interest

For the VGG16-MLP model, the accuracy, precision, sensitivity, and F1-score of the original images before applying the segmentation were 87%, 89%, 86%, and 86%, respectively. So, in comparison to the segmented image results, the accuracy value increased by 9%, the precision by 8%, and the sensitivity and F1-score by 10% (Table 4).

3.3.2. Effect of Data Augmentation

The accuracy, precision, sensitivity, and F1-score of our model without data augmentation were 86%, 86%, 87%, and 86%, respectively. Compared to the model results with data augmentation, the accuracy and F1-score increased by 10%, the precision by 11%, and the sensitivity by 9% (Table 4).

3.3.3. Effect of Spatial Attention Mechanism

The accuracy, precision, sensitivity, and F1-score of VGG16-MLP without applying the spatial attention mechanism were 86%, 86%, 87%, and 86%, respectively. In comparison to the results using this function, the accuracy and F1-score increased by 10%, the precision by 11%, and the sensitivity by 9% (Table 4).

3.3.4. Effect of Depthwiseconv2D Layers

The accuracy, precision, sensitivity, and F1-score of our model without applying the two layers of Depthwiseconv2D were 89%, 89%, 89%, and 88%, respectively. Compared to the model results using Depthwiseconv2D layers, the accuracy and sensitivity increased by 7%, and the precision and F1-score by 8% (Table 4).

3.4. Visual Reader Analysis

The first reader achieved an accuracy of 69%, precision of 95%, sensitivity of 71%, and F1-score of 81%. The second reader’s performance metrics were as follows: 79% accuracy, 97% precision, 80% sensitivity, and 87% F1-score. (Figure 8).

4. Discussion

CAD is expanding rapidly and is becoming widely used in radiological assessment [26]. The visual evaluation of MI, myocarditis, and healthy LGE images may be subjective and lead to incoherencies between readers. Hence, we decided to use deep learning to develop a CAD and enhance objectivity. Our study demonstrates that the VGG16-MLP model allows for distinguishing between MI, myocarditis, and healthy cardiac LGE images with a high accuracy level of 96% on automatically segmented images, which surpasses the two MRI readers. A recent study [5] aimed to determine if radiomics features extracted from LGE areas in cardiac MRI scans allow the differentiation between MI and myocarditis. This study was based on a machine learning algorithm. The highest accuracy was 89% using the SVM classifier. The classical medical image classification methods have not been sufficient to fulfill the current requirements due to the high complexity of medical images. In recent times, the evolution of deep learning theory has introduced a technical framework to answer the challenges of medical image classification [27]. According to our results, the use of deep learning improves the accuracy of the classification task and outperforms the subjective visual analysis. Deep learning architectures have the advantage of automatically extracting key features from datasets without explicit human intervention, as opposed to traditional machine learning, which requires experts’ manual selection of relevant features [11,12]. Another important advantage over machine learning is that it incorporates a greater number of learning layers and creates more advanced and generalized data representations through a superior level of abstraction [28]. In our study, VGG16-MLP was the best model for the classification of LGE cardiac MRI images compared to the other models combined with MLP (VGG19, DenseNet121, DenseNet201, MobileNet, InceptionV3, and InceptionResNetV2). The VGG16 model was introduced by K. Simonyan and A. Zisserman from Oxford University [18]. This architecture has gained considerable popularity in the research field [29] due to its simple structure of neural networks [23,29]. Thus, it has enhanced generalization capabilities and the ability to adapt to diverse datasets [18]. Several studies based on VGG16 for the classification task demonstrate the great performance of this model in various medical imaging modalities [16,18,24,30,31,32,33,34,35]. Tommaso et al. [5] evaluated the SVM model for the classification task of MI and myocarditis LGE images. In our work, SVM was assessed and compared with MLP for the classification task of MI, myocarditis, and healthy LGE images. The VGG16-MLP surpasses the VGG16-SVM in terms of accuracy, precision, sensitivity, and F1-score. It should be noted that image preprocessing is an important factor in improving the classification results. According to our study, segmenting the left ventricle significantly improves the accuracy of our model. Reducing unnecessary information in the images enables a more effective extraction of features. For instance, a study published in 2021 demonstrated that accuracy was higher when the image matrix was centered on the cardiac region of interest [16]. Furthermore, we used automatic segmentation using deep learning, which is characterized by greater consistency, reproducibility, and speed compared to manual segmentation, which is subjective and laborious. CNNs typically need to be trained on large numbers of images to achieve accurate performance [36]. To overcome the problem of limited data, we used the transfer learning technique first. It is the practice of developing high-performing learners who are pre-trained using readily available data from different domains [37]. It permits the exploitation of knowledge from the existing labeled data to improve a model’s performance with no labeled or limited data [38]. Secondly, we opted for the data augmentation technique. It is used to artificially expand the training set by performing small transformations on the existing data [39]. In our work, we applied a 15% zoom and a 10% horizontal shift range. We found that our model’s results were better using data augmentation. In fact, increasing the diversity of training data improves the robustness and performance of deep learning models, as well as their ability to generalize to the data [40]. To avoid overfitting and provide more reliable performance estimates, we used the K-fold cross-validation technique. It consists of dividing a dataset into k subsets; then, iteratively, some of them are used to learn the model, while others are used to evaluate its performance [41]. In addition to cross-validation, holdout test data were used. It allows the evaluation of the performance of our model on completely unseen data, which assesses its real-world efficacy on new data. It should be noted that deep learning faces challenges in medical image classification, including the difficulty of constructing high-performance models adapted to the characteristics of medical images and the limited adaptability of current network structures and training strategies. Thus, we used the spatial attention mechanism. In our work, this function considerably ameliorates the results of classification. Incorporating an attention mechanism into a deep learning model improves information extraction efficiency for medical image analysis and enhances the precision of reasoning [42]. Spatial attention can be seen as an adaptive mechanism for selecting spatial regions, inspired by human visual capabilities. This mechanism can be treated as a dynamic selection process for focusing on the most important features within an image [43]. Attention mechanisms integrated with CNNs were used in numerous studies for medical image classification [20,21,22,44]. The major limitation of our study is the limited dataset. Large datasets ensure greater performance for deep learning models. Secondly, other revealing images of myocarditis and MI, such as cine post-gadolinium and T2-weighted images, could be added to the dataset. Thirdly, we could consider including more types of cardiomyopathies in our model. Finally, in our study, all patients were included without exclusion during the data collection period, minimizing selection bias and ensuring a diverse representation of the clinical population. Additionally, all images were acquired using a single MRI scanner to reduce technical variability. The retrospective nature of the study may still introduce potential biases such as variations in image quality and patient demographics, which we will address by expanding the dataset. As a future perspective, our approach is to integrate clinical data into the algorithm to increase the diversity of the input data and thus improve the accuracy of the classification.

5. Conclusions

In conclusion, the VGG16-MLP model permits accurate classification of MI, myocarditis, and healthy LGE cardiac MRI images and could be considered a reliable CAD model. In order to enhance the robustness of our model, it would be advantageous to incorporate various datasets. The generalizability of our model to other cardiac diseases would require expanding the dataset with additional annotated data.

Author Contributions

A.B.K. (Amani Ben Khalifa): Conceptualization, Methodology, and Writing—original draft. M.M. (Manel Mili): Methodology and Software. M.M. (Mezri Maatouk): Conceptualization, Investigation, Data Curation, and Writing—Review and Editing. A.B.A. (Asma Ben Abdallah): Methodology, Validation, and Visualization. M.A.: Conceptualization and Methodology. S.G.: Validation and Visualization. A.B.A. (Azza Ben Ali): Formal Analysis. Y.L.: Formal Analysis. M.H.B.: Methodology, Validation, Visualization, Supervision, and Writing—Review and Editing. A.Z.: Supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study protocol was approved by the local research ethics committee at the Faculty of Medicine of Monastir, Tunisia (IORG 0009738 N°171/OMB 0990-0279, approval date: 06/04/2024) and complied with the Declaration of Helsinki.

Informed Consent Statement

Informed consent was waived for this retrospective study.

Data Availability Statement

Due to privacy and confidentiality concerns, the data supporting the findings of this study are not publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Russo, V.; Lovato, L.; Ligabue, G. Cardiac MRI: Technical basis. La Radiol. Med. 2020, 125, 1040–1055. [Google Scholar] [CrossRef] [PubMed]
  2. Sechtem, U.; Mahrholdt, H.; Vogelsberg, H. Cardiac magnetic resonance in myocardial disease. Heart 2007, 93, 1520–1527. [Google Scholar] [CrossRef]
  3. Nucifora, G.; Miani, D.; Di Chiara, A.; Piccoli, G.; Artico, J.; Puppato, M.; Slavich, G.; De Biasio, M.; Gasparini, D.; Proclemer, A. Infarct-like acute myocarditis: Relation between electrocardiographic findings and myocardial damage as assessed by cardiac magnetic resonance imaging. Clin. Cardiol. 2013, 36, 146–152. [Google Scholar] [CrossRef]
  4. Valevičienė, N.; Petrulionienė, Z.; Petrauskienė, B.; Lauraitis, G.; Glaveckaitė, S.; Palionis, D.; Tamošiūnas, A.; Laucevičius, A. Differentiation of acute myocarditis and acute myocardial infarction by the regional distribution of myocardial irreversible injury using cardiovascular magnetic resonance imaging. Medicina 2012, 48, 132–137. Available online: https://pubmed.ncbi.nlm.nih.gov/22588344/ (accessed on 8 November 2023). [CrossRef]
  5. Di Noto, T.; von Spiczak, J.; Mannil, M.; Gantert, E.; Soda, P.; Manka, R.; Alkadhi, H. Radiomics for Distinguishing Myocardial Infarction from Myocarditis at Late Gadolinium Enhancement at MRI: Comparison with Subjective Visual Analysis. Radiol. Cardiothorac. Imaging 2019, 1, e180026. [Google Scholar] [CrossRef]
  6. Al’Aref, S.J.; Anchouche, K.; Singh, G.; Slomka, P.J.; Kolli, K.K.; Kumar, A.; Pandey, M.; Ma-liakal, G.; van Rosendael, A.R.; Beecy, A.N.; et al. Clinical applications of machine learning in cardiovascular disease and its relevance to cardiac imaging. Eur. Heart J. 2019, 40, 1975–1986. [Google Scholar] [CrossRef] [PubMed]
  7. Krittanawong, C.; Virk, H.U.H.; Bangalore, S.; Wang, Z.; Johnson, K.W.; Pinotti, R.; Zhang, H.; Kaplin, S.; Narasimhan, B.; Kitai, T.; et al. Machine learning prediction in cardiovascular diseases: A meta-analysis. Sci. Rep. 2020, 10, 16057. [Google Scholar] [CrossRef] [PubMed]
  8. Nagavelli, U.; Samanta, D.; Chakraborty, P. Machine Learning Technology-Based Heart Disease Detection Models. J. Healthc. Eng. 2022, 2022, 7351061. [Google Scholar] [CrossRef] [PubMed]
  9. Kawasaki, T.; Kidoh, M.; Kido, T.; Sueta, D.; Fujimoto, S.; Kumamaru, K.K.; Uetani, T.; Tanabe, Y.; Ueda, T.; Sakabe, D.; et al. Evaluation of Significant Coronary Artery Disease Based on CT Fractional Flow Reserve and Plaque Characteristics Using Random Forest Analysis in Machine Learning. Acad. Radiol. 2020, 27, 1700–1708. [Google Scholar] [CrossRef]
  10. Leiner, T.; Rueckert, D.; Suinesiaputra, A.; Baeßler, B.; Nezafat, R.; Išgum, I.; Young, A.A. Machine learning in cardiovascular magnetic resonance: Basic concepts and applications. J. Cardiovasc. Magn. Reson. 2019, 21, 61. [Google Scholar] [CrossRef]
  11. Sun, Z.; Shi, Z.; Xin, Y.; Zhao, S.; Jiang, H.; Li, J.; Li, J.; Jiang, H. Contrast-Enhanced CT Imaging Features Combined with Clinical Factors to Predict the Efficacy and Prognosis for Transarterial Chemoembolization of Hepatocellular Carcinoma. Acad. Radiol. 2023, 30, S81–S91. [Google Scholar] [CrossRef] [PubMed]
  12. Hausmann, D.; Lerch, A.; Hitziger, S.; Farkas, M.; Weiland, E.; Lemke, A.; Grimm, M.; Kubik-Huch, R.A. AI-Supported Autonomous Uterus Reconstructions: First Application in MRI Using 3D SPACE with Iterative Denoising. Acad. Radiol. 2023, 31, 1400–1409. [Google Scholar] [CrossRef]
  13. Henglin, M.; Stein, G.; Hushcha, P.V.; Snoek, J.; Wiltschko, A.B.; Cheng, S. Machine Learning Approaches in Cardiovascular Imaging. Circ. Cardiovasc. Imaging 2017, 10, e005614. [Google Scholar] [CrossRef]
  14. Shi, Y.-J.; Zhu, H.-T.; Li, X.-T.; Zhang, X.-Y.; Liu, Y.-L.; Wei, Y.-Y.; Sun, Y.-S. Histogram array and convolutional neural network of DWI for differentiating pancreatic ductal adenocarcinomas from solid pseudopapillary neoplasms and neuroendocrine neoplasms. Clin. Imaging 2023, 96, 15–22. [Google Scholar] [CrossRef]
  15. Liu, X.; Faes, L.; Kale, A.U.; Wagner, S.K.; Fu, D.J.; Bruynseels, A.; Mahendiran, T.; Moraes, G.; Shamdas, M.; Kern, C.; et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: A systematic review and meta-analysis. Lancet Digit. Health 2019, 1, e271–e297. [Google Scholar] [CrossRef]
  16. Germain, P.; Vardazaryan, A.; Padoy, N.; Labani, A.; Roy, C.; Schindler, T.H.; El Ghannudi, S. Classification of Cardiomyopathies from MR Cine Images Using Convolutional Neural Network with Transfer Learning. Diagnostics 2021, 11, 1554. [Google Scholar] [CrossRef] [PubMed]
  17. Martini, N.; Aimo, A.; Barison, A.; Della Latta, D.; Vergaro, G.; Aquaro, G.D.; Ripoli, A.; Emdin, M.; Chiappino, D. Deep learning to diagnose cardiac amyloidosis from cardiovascular magnetic resonance. J. Cardiovasc. Magn. Reson. 2020, 22, 84. [Google Scholar] [CrossRef] [PubMed]
  18. Guan, Q.; Wang, Y.; Ping, B.; Li, D.; Du, J.; Qin, Y.; Lu, H.; Wan, X.; Xiang, J. Deep convolutional neural network VGG-16 model for differential diagnosing of papillary thyroid carcinomas in cytological images: A pilot study. J. Cancer 2019, 10, 4876–4882. [Google Scholar] [CrossRef]
  19. Messaoud, N.H.; Mansour, A.; Aissi, M.; Ayari, R.; Frih, M.; Ben Abdallah, A.; Bedoui, M.H. Automated segmentation of multiple sclerosis lesions based on convolutional neural networks. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2023, 11, 1359–1377. [Google Scholar] [CrossRef]
  20. Bodapati, J.D.; Shaik, N.S.; Naralasetti, V. Composite deep neural network with gated-attention mechanism for diabetic retinopathy severity classification. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 9825–9839. [Google Scholar] [CrossRef]
  21. Jiang, M.; Yan, B.; Li, Y.; Zhang, J.; Li, T.; Ke, W. Image Classification of Alzheimer’s Disease Based on External-Attention Mechanism and Fully Convolutional Network. Brain Sci. 2022, 12, 319. [Google Scholar] [CrossRef]
  22. Li, J.; Dou, Q.; Yang, H.; Liu, J.; Fu, L.; Zhang, Y.; Zheng, L.; Zhang, D. Cervical cell multi-classification algorithm using global context information and attention mechanism. Tissue Cell 2022, 74, 101677. [Google Scholar] [CrossRef] [PubMed]
  23. Kong, L.; Cheng, J. Classification and detection of COVID-19 X-Ray images based on DenseNet and VGG16 feature fusion. Biomed. Signal Process. Control. 2022, 77, 103772. [Google Scholar] [CrossRef] [PubMed]
  24. Saini, D.; Khosla, A.; Chand, T.; Chouhan, D.K.; Prakash, M. Automated knee osteoarthritis severity classification using three-stage preprocessing method andVGG16architecture. Int. J. Imaging Syst. Technol. 2023, 33, 1028–1047. [Google Scholar] [CrossRef]
  25. Nahm, F.S. Receiver operating characteristic curve: Overview and practical use for clinicians. Korean J. Anesthesiol. 2022, 75, 25–36. [Google Scholar] [CrossRef]
  26. Huang, Z.; Xiao, J.; Wang, X.; Li, Z.; Guo, N.; Hu, Y.; Li, X.; Wang, X. Clinical Evaluation of the Automatic Coronary Artery Disease Reporting and Data System (CAD-RADS) in Coronary Computed Tomography Angiography Using Convolutional Neural Networks. Acad. Radiol. 2022, 30, 698–706. [Google Scholar] [CrossRef]
  27. Mahmood, T.; Rehman, A.; Saba, T.; Nadeem, L.; Bahaj, S.A.O. Recent Advancements and Future Prospects in Active Deep Learning for Medical Image Segmentation and Classification. IEEE Access 2023, 11, 113623–113652. [Google Scholar] [CrossRef]
  28. Taye, M.M. Understanding of Machine Learning with Deep Learning: Architectures, Workflow, Applications and Future Directions. Computers 2023, 12, 91. [Google Scholar] [CrossRef]
  29. Nash, W.; Drummond, T.; Birbilis, N. A review of deep learning in the study of materials degradation. npj Mater. Degrad. 2018, 2, 37. [Google Scholar] [CrossRef]
  30. Han, Y.; Ma, Y.; Wu, Z.; Zhang, F.; Zheng, D.; Liu, X.; Tao, L.; Liang, Z.; Yang, Z.; Li, X.; et al. Histologic subtype classification of non-small cell lung cancer using PET/CT images. Eur. J. Nucl. Mol. Imaging 2020, 48, 350–360. [Google Scholar] [CrossRef] [PubMed]
  31. da Rocha, D.A.; Ferreira, F.M.F.; Peixoto, Z.M.A. Diabetic retinopathy classification using VGG16 neural network. Res. Biomed. Eng. 2022, 38, 761–772. [Google Scholar] [CrossRef]
  32. Sitaula, C.; Hossain, M.B. Attention-based VGG-16 model for COVID-19 chest X-ray image classification. Appl. Intell. 2021, 51, 2850–2863. [Google Scholar] [CrossRef]
  33. Bakasa, W.; Viriri, S. VGG16 Feature Extractor with Extreme Gradient Boost Classifier for Pancreas Cancer Prediction. J. Imaging 2023, 9, 138. [Google Scholar] [CrossRef]
  34. Naveen, P.; Diwan, B. Pre-trained VGG-16 with CNN architecture to classify X-Rays images into normal or pneumonia. In Proceedings of the 2021 International Conference on Emerging Smart Computing and Informatics, ESCI 2021, Pune, India, 5–7 March 2021; pp. 102–105. [Google Scholar] [CrossRef]
  35. Nishio, M.; Noguchi, S.; Matsuo, H.; Murakami, T. Automatic classification between COVID-19 pneumonia, non-COVID-19 pneumonia, and the healthy on chest X-ray image: Combination of data augmentation methods. Sci. Rep. 2020, 10, 17532. [Google Scholar] [CrossRef]
  36. Landsmann, A.; Ruppert, C.; Borkowski, K.; Hejduk, P.; Ciritsis, A.; Wieler, J.; Rossi, C.; Boss, A. Detection of microcalcifications in photon-counting dedicated breast-CT using a deep convolutional neural network: Proof of principle. Clin. Imaging 2023, 95, 28–36. [Google Scholar] [CrossRef]
  37. Weiss, K.; Khoshgoftaar, T.M.; Wang, D.D. A survey of transfer learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef]
  38. Day, O.; Khoshgoftaar, T.M. A survey on heterogeneous transfer learning. J. Big Data 2017, 4, 29. [Google Scholar] [CrossRef]
  39. Athalye, C.; Arnaout, R. Domain-guided data augmentation for deep learning on medical imaging. PLoS ONE 2023, 18, e0282532. [Google Scholar] [CrossRef] [PubMed]
  40. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  41. Anguita, D.; Ghelardoni, L.; Ghio, A.; Oneto, L.; Ridella, S. The “K” in K-fold Cross Validation. In Proceedings of the 20th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN), Bruges, Belgium, 25–27 April 2012; pp. 441–446. [Google Scholar]
  42. An, F.; Li, X.; Ma, X. Medical Image Classification Algorithm Based on Visual Attention Mechanism-MCNN. Oxidative Med. Cell. Longev. 2021, 2021, 6280690. [Google Scholar] [CrossRef] [PubMed]
  43. Guo, M.-H.; Xu, T.-X.; Liu, J.-J.; Liu, Z.-N.; Jiang, P.-T.; Mu, T.-J.; Zhang, S.-H.; Martin, R.R.; Cheng, M.-M.; Hu, S.-M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  44. Yao, H.; Zhang, X.; Zhou, X.; Liu, S. Parallel Structure Deep Neural Network Using CNN and RNN with an Attention Mechanism for Breast Cancer Histology Image Classification. Cancers 2019, 11, 1901. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Framework for the extraction of the region of interest (ROI).
Figure 1. Framework for the extraction of the region of interest (ROI).
Diagnostics 15 00207 g001
Figure 2. Descriptive architecture of our deep learning model VGG16-MLP composed of fine-tuned Visual Geometry Group (VGG16), spatial attention mechanism, and multi-layer perceptron (MLP).
Figure 2. Descriptive architecture of our deep learning model VGG16-MLP composed of fine-tuned Visual Geometry Group (VGG16), spatial attention mechanism, and multi-layer perceptron (MLP).
Diagnostics 15 00207 g002
Figure 3. A visual representation of the input images, ground truth masks, and predicted masks.
Figure 3. A visual representation of the input images, ground truth masks, and predicted masks.
Diagnostics 15 00207 g003
Figure 4. Bar charts of accuracy, dice coefficient, sensitivity, and precision of segmentation of the region of interest.
Figure 4. Bar charts of accuracy, dice coefficient, sensitivity, and precision of segmentation of the region of interest.
Diagnostics 15 00207 g004
Figure 5. Confusion matrices of VGG16-MLP, VGG19-MLP, DenseNet121-MLP, DenseNet201-MLP, MobileNet-MLP, InceptionV3-MLP, InceptionResNetV2-MLP, and VGG16-SVM.
Figure 5. Confusion matrices of VGG16-MLP, VGG19-MLP, DenseNet121-MLP, DenseNet201-MLP, MobileNet-MLP, InceptionV3-MLP, InceptionResNetV2-MLP, and VGG16-SVM.
Diagnostics 15 00207 g005
Figure 6. Bar charts of VGG16-MLP model’s accuracy, precision, sensitivity, and F1-score for myocardial infarction (class 0), myocarditis (class 1), and healthy (class 2). For the myocardial infarction, myocarditis, and healthy classes, the accuracy was 97%, 98%, and 98%; the precision was 95%, 100%, and 96%; the sensitivity was 99%, 91%, and 98%; and the F1-score was 97%, 95%, and 97%, respectively.
Figure 6. Bar charts of VGG16-MLP model’s accuracy, precision, sensitivity, and F1-score for myocardial infarction (class 0), myocarditis (class 1), and healthy (class 2). For the myocardial infarction, myocarditis, and healthy classes, the accuracy was 97%, 98%, and 98%; the precision was 95%, 100%, and 96%; the sensitivity was 99%, 91%, and 98%; and the F1-score was 97%, 95%, and 97%, respectively.
Diagnostics 15 00207 g006
Figure 7. The receiver operating characteristic (ROC) curve and precision–recall curve of the myocardial infarction (class 0), myocarditis (class 1), and healthy (class 2) classes obtained from the VGG16-MLP model. In the ROC curves for the three classes, the true-positive rates are close to one, the false-positive rates are near zero, and the area under the curve (AUC) exceeds 0.9. The precision–recall curves indicate high values of precision and recall for the three classes, with AUCs of 1.00, 0.98, and 0.99 for the myocardial infarction, myocarditis, and healthy classes, respectively.
Figure 7. The receiver operating characteristic (ROC) curve and precision–recall curve of the myocardial infarction (class 0), myocarditis (class 1), and healthy (class 2) classes obtained from the VGG16-MLP model. In the ROC curves for the three classes, the true-positive rates are close to one, the false-positive rates are near zero, and the area under the curve (AUC) exceeds 0.9. The precision–recall curves indicate high values of precision and recall for the three classes, with AUCs of 1.00, 0.98, and 0.99 for the myocardial infarction, myocarditis, and healthy classes, respectively.
Diagnostics 15 00207 g007
Figure 8. Bar charts of the accuracy, precision, sensitivity, and F1-score of our model VGG16-MLP, reader 1, and reader 2.
Figure 8. Bar charts of the accuracy, precision, sensitivity, and F1-score of our model VGG16-MLP, reader 1, and reader 2.
Diagnostics 15 00207 g008
Table 1. Dataset description.
Table 1. Dataset description.
HealthyMIMyocarditisp Value
Number of patients278947
Age42.25 (±14.62)58.50 (±11.57)29.06 (±10.19)<0.001
Sex (F/M)8/1915/746/410.433
Number of images 254361222
Total number of images 837
Table 2. The accuracy, precision, sensitivity, F1-score, training time, and inference time of the tested models and the p-values of the statistical comparison of VGG16-MLP with other tested models.
Table 2. The accuracy, precision, sensitivity, F1-score, training time, and inference time of the tested models and the p-values of the statistical comparison of VGG16-MLP with other tested models.
Tested ModelsAccuracy (%)Precision (%)Sensitivity (%)F1-Score (%)p ValueTraining Time (s)Inference
Time (s)
VGG16-MLP96 *97 *96 *96 *-23720.738
VGG19-MLP868586860.84229210.954
DenseNet121-MLP828282810.042 24500.606
DenseNet201-MLP798080790.042 24810.870
MobileNet-MLP777978770.043 20880.210
InceptionV3-MLP777877770.043 23830.510
InceptionResNetV2-MLP898888880.19726100.984
VGG16-SVM898989890.59324220.702
* Highest performance. : p < 0.05.
Table 3. VGG16-MLP model accuracy, precision, sensitivity, and F1-score for myocardial infarction, myocarditis, and healthy classes.
Table 3. VGG16-MLP model accuracy, precision, sensitivity, and F1-score for myocardial infarction, myocarditis, and healthy classes.
ClassesAccuracy (%)Precision (%)Sensitivity (%)F1-Score (%)
Myocardial infarction97959997
Myocarditis981009195
Healthy98969897
Table 4. Accuracy, precision, sensitivity, and F1-score of VGG16-MLP without image segmentation, data augmentation, spatial attention mechanism, or Depthwiseconv2D layers.
Table 4. Accuracy, precision, sensitivity, and F1-score of VGG16-MLP without image segmentation, data augmentation, spatial attention mechanism, or Depthwiseconv2D layers.
Accuracy (%)Precision (%)Sensitivity (%)F1-Score (%)
VGG16-MLP 96979696
VGG16-MLP without image segmentation87898686
VGG16-MLP without data augmentation86868786
VGG16-MLP without spatial attention mechanism86868786
VGG16-MLP without Depthwiseconv2D layers89898988
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ben Khalifa, A.; Mili, M.; Maatouk, M.; Ben Abdallah, A.; Abdellali, M.; Gaied, S.; Ben Ali, A.; Lahouel, Y.; Bedoui, M.H.; Zrig, A. Deep Transfer Learning for Classification of Late Gadolinium Enhancement Cardiac MRI Images into Myocardial Infarction, Myocarditis, and Healthy Classes: Comparison with Subjective Visual Evaluation. Diagnostics 2025, 15, 207. https://doi.org/10.3390/diagnostics15020207

AMA Style

Ben Khalifa A, Mili M, Maatouk M, Ben Abdallah A, Abdellali M, Gaied S, Ben Ali A, Lahouel Y, Bedoui MH, Zrig A. Deep Transfer Learning for Classification of Late Gadolinium Enhancement Cardiac MRI Images into Myocardial Infarction, Myocarditis, and Healthy Classes: Comparison with Subjective Visual Evaluation. Diagnostics. 2025; 15(2):207. https://doi.org/10.3390/diagnostics15020207

Chicago/Turabian Style

Ben Khalifa, Amani, Manel Mili, Mezri Maatouk, Asma Ben Abdallah, Mabrouk Abdellali, Sofiene Gaied, Azza Ben Ali, Yassir Lahouel, Mohamed Hedi Bedoui, and Ahmed Zrig. 2025. "Deep Transfer Learning for Classification of Late Gadolinium Enhancement Cardiac MRI Images into Myocardial Infarction, Myocarditis, and Healthy Classes: Comparison with Subjective Visual Evaluation" Diagnostics 15, no. 2: 207. https://doi.org/10.3390/diagnostics15020207

APA Style

Ben Khalifa, A., Mili, M., Maatouk, M., Ben Abdallah, A., Abdellali, M., Gaied, S., Ben Ali, A., Lahouel, Y., Bedoui, M. H., & Zrig, A. (2025). Deep Transfer Learning for Classification of Late Gadolinium Enhancement Cardiac MRI Images into Myocardial Infarction, Myocarditis, and Healthy Classes: Comparison with Subjective Visual Evaluation. Diagnostics, 15(2), 207. https://doi.org/10.3390/diagnostics15020207

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop