A Novel Robust CNN Model For MRI Brain Tumor
A Novel Robust CNN Model For MRI Brain Tumor
A Novel Robust CNN Model For MRI Brain Tumor
https://doi.org/10.33472/AFJBS.6.5.2024. 7113-7126
AfricanJournalofBiological
Sciences
A Novel Robust CNN Model for MRI Brain Tumor
Classification
Sulekha Das1, *Dr.Avijit Kumar Chaudhuri2, Dr.Partha Ghosh3, Swagato Sikdar4
1
Research Scholar, Information Technology, GCECT, Kolkata, West Bengal, India
2
Associate Professor, Computer Science and Engineering, Brainware University, Barasat, West Bengal, India
3
Assistant Professor, Computer Science and Engineering, GCECT, Kolkata, West Bengal, India
4
Student –CSE-AIML(2nd year), Techno Engineering College Banipur
shu7773sea@gmail.com,c.avijit@gmail.com,parth_ghos@rediffmail.com,sikdarswagato0@gmail.co
m
Abstract
Brain tumors rank as the 10th leading cause of death for men and
women, affecting both adults and children1. This underscores the urgent
need for effective strategies in prevention, diagnosis, and treatment to
Article History
address this significant health challenge. Magnetic resonance imaging
Volume 6, Issue 5, 2024 (MRI) is the preferred method for identifying brain tumors. Recent
Received: 15 May 2024
Accepted: 22 May 2024 advancements in image classification technology, particularly
doi:10.33472/AFJBS.6.5.2024. 7113-7126
Convolutional Neural Networks (CNNs), have greatly improved tumor
classification accuracy. The findings of this study have significant
implications for clinicians specializing in the early detection of brain
tumors. By leveraging advanced neural network models, healthcare
professionals can potentially improve the accuracy and efficiency of
tumor diagnosis, leading to better patient outcomes and possibly earlier
interventions. This underscores the importance of leveraging cutting-edge
technology, such as deep learning and neuroimaging, in medical
diagnostics. In this study, CNNs are used for brain tumor classification,
successfully categorizing brain images into two classes: benign and
malignant with an impressive accuracy of 97%.
1
www.cancer.net
Introduction
Brain tumors rank as the 10th leading cause of death for men and women, affecting both
adults and children. This underscores the urgent need for effective strategies in prevention,
diagnosis, and treatment to address this significant health challenge. Magnetic resonance
imaging (MRI) is the preferred method for identifying brain tumors. Recent advancements in
image classification technology, particularly the brain is, by far, the master organ of the
human body, controlling the intricately intertwined processes of all other systems and making
critical choices in our lives (Saba, 2020). Being the coordinator of the central nervous system,
it controls both the voluntary activities and the involuntary processes that are a vital part of
our body's functioning (Özyurt et al., 2020). While the heart serves as a critical organ that
keeps us alive, the development of tumors within this vital structure poses considerable
threats to our health and wellness.
Brain and central nervous system cancers represent a growing, frightful global public health
challenge because of their high death rate, inflating costs to individuals and society, poor
survival rate, and extremely low quality of life. As specified in the GLOBOCAN 2020 data,
brain and central nervous system cancer is a significant share of the global causes of sickness
(Ilic&Ilic, 2023). They are the 19th most frequent cancer, accounting for 1.9% of all cancers,
and the 12th leading cause of cancer-related deaths, amounting to 2.5% of overall cancer
deaths. Comprehensive, in-depth knowledge of the many aspects of brain tumors, like their
symptoms, stage, and criticality, is very important for early diagnosis and solution. Magnetic
Resonance Imaging (MRI) has been a base for clinicians to uncover internal subtle
information about the brain structure and composition, including tumors (Sharif et al.,
2021). Advanced computational methods, including deep learning, have demonstrated their
capability to improve the accuracy and efficiency of tumor detection and classification.
On the contrary, it seems that a lot of articles have been written on this subject to summarize
the current trends in DL, but they usually center around a very narrow part of the issue, which
in the end leads to a situation in which we know too little about the field in its
totality. Therefore, people do not have much idea about the scope of DL developments and
how much they have grown up in this time.
Page 7115 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
In this regard, Convolutional Neural Networks (CNNs) have become an essential instrument
for discerning and classifying different types of brain tumors, which include gliomas,
meningiomas, pituitary tumors, and normal brain tissue. CNNs include multiple layers of
interconnected neurons. When all of them are connected, they form a network that can learn
complex patterns and features from data (Hu &Razmjooy, 2021). Datasets with labeled brain
images are the means of adjusting neurons' weights and biases across various layers for
CNNs during the training process. Through repeated comparisons of predicted outputs and
ground truth labels, the network refines its parameters to reduce errors and achieve the
highest accuracy. This pattern of iterative learning occurs through the activation of functions,
which introduces non-linearity and thus allows the network to capture internal relationships
within the data.
The structure of CNNs is mainly the input layer, output layer, and at least one hidden layer is
present. These layers act as key components in the feature extraction and abstraction
processes, which enables the network to detect these sensitive distinctions. Moreover, trial
and error with various parameters of the machine learning model like the number of layers,
input filters, and learning rates positively impacts the diagnosis of different tissues
(Mehnatkesh et al., 2023, Kumar &Mankame, 2020). Despite their computational
complexity, CNNs deliver a unique function in dealing with and processing medical images
and their capabilities to excel in the field of neuroimaging research cannot be devalued.
Researchers constantly attempt to increase the efficacy of these deep learning models to
enhance diagnostic precision, simplify clinical procedures, and finally, ensure the
improvement of patient outcomes in the diagnosis and treatment of brain cancer.
The architecture of the experimental layout of this research and model construction is
presented in Fig.1. A CNN in PyTorch is formulated from constituent
layers which include convolutional, pooling, and fully connected layers, to build the model
structure. These levels are necessary to build up hierarchical representations of features from
input data consideration, especially for tasks like image classification. The process consists
of creating an initial Python class consisting of the CNN model. All these layers are set as an
initial layer where parameters are defined, such as the kernel size, the number of filters, and
activation functions. For instance, the authors start with four convolution layers followed by
pooling layers to down sample the feature map and simply connected layers for classification.
After the creation of these layers, they are passed into the forward function of the class one
Page 7116 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
by one, which produces the outputs of the specific network. It is a key part of the network
where data undergoes convolutional operations, followed by pooling, and then steps through
fully connected layers ultimately. With this connected architecture, information flows in a
predefined order during both prediction and training (backward pass) to facilitate the efficient
propagation of information through the network.
The model is trained with image datasets after constructing architecture. This is done by
feeding the input images through the network one after another, computing loss between
predicted and actual labels, and then using techniques like backpropagation and optimization
algorithms (like Stochastic Gradient Descent, Adam) to update network parameters
iteratively. There is an automatic learning functionality enabled that allows the convolutional
and pooling layers to learn effective feature representations from the input images. This is
done through the process of optimization of parameters. One of the biggest strengths of
CNNs is their ability to learn detailed features directly from data making them suitable for
achieving outstanding performance across several computer vision tasks.
The authors employed several strategies to enhance the performance of their Convolutional
Neural Network (CNN) model for vision tasks:
Adding More Layers: Through an increase in the depth of the CNN architecture, the
model is essentially capable of learning more complex elements and features from the input
data. This leads to the acquisition of better precision and increases the overall quality of the
model.
Training for More Epochs: A rise of epochs means that the model will see the training data
more times, which helps higher convergence to fine-tuned features. Thus, the model can
create more complex embeddings that manage to encapsulate a greater diversity of
features that can be found in the input data.
Experimentation and Iterative Improvement: The authors use the cycle of trial and error
to improve the model by iteratively adjusting the model architecture, training procedures, and
parameters depending on the feedback. Therefore, this process involves systematically
exploring different configuration options and strategies to find the best-performing model to
match the specific vision tasks.
The authors will implement these strategies and will never cease refining their model through
experimentation to achieve the accuracy and robustness of the vision tasks given. Therefore,
this iterative process shows that a model should meet high initial standards and also be
improved and optimized during the whole learning process. The architecture for the proposed
classifier for brain cancer prediction has been depicted in Figure 1 below.
Relevant Literature
Afshar et al., 2018 indicate, that the early and precise diagnosis of brain tumors is
complicated based on brain tumors' different severity and incidence across different age
groups. Convolution neural networks (CNNs) are suitable for this type of task, but they
require trained data and also cannot handle input transformations. Capsule Network
(CapsNets), the breakthrough in machine learning architectures, is a natural choice here since
it is more robust to transformations and requires a smaller training set, which makes it perfect
for such a small dataset and brain MRI images.
Experimental data show that CapsNet has a CNN (convolutional neural network) structure,
performing better than the others in the deep learning area, including imaging tasks,
especially in the field of medicine.
Sajjad et al., 2019 proposed a novel multi-grade classification system for brain tumor
diagnosis based on the CNN approach, which aims to alleviate the work burden of
radiologists and better utilize MRI data. The suggested system is verified through data
augmentation and the original data. It shows a better performance than existing ones. The
results have demonstrated the efficiency of this method in accurate and reliable brain tumor
classification. Thus, radiologists can work properly based on medical imaging analysis.
Deepak et al., 2019 present a new approach to brain tumor classification in which three
different types of tumors are identified: Glioma, meningioma, and pituitary are the major
three diseases to be detected by the deep convolutional transfer learning approach. The
methodology includes the use of the GoogLeNetmodel that has been pre-trained to produce
the image attributes from the MRI images of the brain. Subsequently, these features shall be
given to trained classifier models so that classification can be performed. This study
randomly applies a patient-level, five-fold cross-validation process to MRI data collected
through Figshare. The obtained results reveal a highly accurate mean classification accuracy
of 98%. Performance metrics, e.g., AUC, Precision, Recall, F-score, and Specificity are also
Page 7119 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
computed. On the other hand, this paper reveals the practicality of small training samples,
based on transfer learning, which is shown in this case. The outcomes prove that deep
transfer learning is an effective method in medical image analysis, especially when the data is
less available.
Deb et al.(2024) in their proposed work, make use of deep neural network form, which is a type
of convolutional neural network (CNN) to classify images into TUMOUR DETECTED and
TUMOUR NOT DETECTED. The model has achieved outstanding results in terms of
accuracy, with a mean accuracy score of 96.08% and a fscore of 97.3. This suggests that such
models can detect tumors precisely in MRIs and hence demonstrate the path with deep
learning methods used in the field of medical image analysis and diagnosis.
Ayadi et al., 2021 noted that the main aim of brain tumors classification results from their
grave consequences as well as different classifications. Moreover, the new CNN-based
model, which can differentiate between brain tumors in MRI images, is illustrated in the first
instance. This model has been made up of numerous layers of the network to achieve decent
classification accuracy for brain tumors.
Table 1. Summary of recent work that has been performed in brain cancer detection and in
this research using machine learning and deep learning algorithms
Specificity = NA
AUC = NA
Sensitivity = 91.84
Specificity = 99.57
AUC = NA
Specificity = 95
AUC = 98.6
Sensitivity = 81.8
Specificity = 90
AUC = 94.1
Accuracy = 98
SRC, KNN
SPC = 81
AUC = 91
Sensitivity = 86.26
Specificity = 90.90
Page 7121 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
Sensitivity = 88.41
Specificity = 96.12
Sensitivity = 96
Specificity = 98
Kappa Score = 94
AUC/ ROC = 97
Methodology
Dataset Description
The dataset used in this study for the brain tumor classification was acquired from Kaggle,
containing around 3762 MRI images. The dataset contains 2079 images without any tumors
and 1683 images with tumors. These images, which are segmented into training and testing,
are used for the creation of models. The training data corresponds to the part of input
information used for the learning process, whereas the testing data is used to test
performance. The dataset has been divided into training and testing groups in various ratios,
such as 60/40, 70/30 and 80/20 train/test partitions, and 10-fold cross-validation. In cross-
validation, the dataset is split into mutually exclusive subsets, and the training and testing are
done through the different subset combinations. To illustrate this, Figure # demonstrates
some data that will be used for experimental purposes. The visualization part of the process
can, therefore, be of help in understanding the general qualities and features of the
dataset that is being examined.
Page 7122 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
A Convolutional Neural Network (CNN) is an exemplary deep learning architecture that was
developed for dealing with structured grid data frameworks such as images. The structure
consists of several layers, wherein each layer performs specific functions and is stacked in
hierarchical order to extract the salient features from the input data.
Input Layer: The input layer receives the raw image data, typically represented as a matrix
of pixel values.
Pooling Layer:The pooling layer manages the spatial dimensions of the feature maps
obtained from the convolutional layers by downsizing. Max pooling, a widely used method,
picks the maximum from each of the local areas (e.g., 2x2) of the feature map. One of the
roles of Max pooling is to reduce computational complexity, make the network more tolerant
of minor input differences, and minimize over-fitting. In addition, pooling layers capture the
most important features and dispose of unrelated information while minimizing sensitivity to
the input translations.
Fully Connected (FC) Layer: The fully connected layers integrate the detected features by
the convolutional and pooling layers for the purpose of classification or regression. Every
single neuron in a fully connected layer is connected to every neuron in the previous layer, so
it forms a tightly packed structure. Unlike CNN, FC layers focus on learning the
interrelationships between features and making the final predictions. We may use activation
functions such as Softmax to assign class probabilities from the fully connected layer output.
Output Layer: The output layer generates the final predicted or classified results. For the
classification tasks, it mainly uses the softmax activation function to convert raw scores into
the degree of probability of belonging to classes. The network's output shows the confidence
interval for the output classes.
Page 7123 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
Feature Map: Feature maps, as the output of convolutional layers, encode learned features.
They referred to as convolutional neural network (CNN) is structured grids of activation
values, where each value represents either the existence or the lack of a specific feature at a
given position in the input. Feature maps are partitioned into the overlapped sets to perform
pooling operations, which are considered as the most important information either with the
max pooling or the average pooling. In summary, CNNs make use of convolutional and
pooling operations to extract hierarchical features from input images and finally fully
connected layers to make decisions based on these features. This architecture obtained
remarkable result in numerous computer vision tasks, such as image classification, object
detection, and segmentation. (Ting et al., 2019; Dang et al., 2019). Figure 2 demonstrate the
CNN architecture.
CNNs have been of interest to many researchers in their ability to process and analyze images
as well as recognize patterns. Here are some key reasons why CNNs are extensively used in
cancer research:
Page 7124 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
CNNs can learn hierarchical image representations automatically and consequently can
extract intricate features from MRI and CT images as a result. These features are particularly
important in detecting even small abnormalities suggestive of cancer(Litjens et al., 2017).
CNNs are capable of classifying medical images as belonging to various categories that in
turn may be utilized for different cancer diagnoses. They are trained to learn discriminative
features and then are able to give a precise classification of the available data to the
healthcare professionals for informative decisions(Esteva et al., 2017).
The CNNs can segment tumour regions in medical images highlighting the edges of
cancerous tissues. Precise segmentation is vital for treatment planning, monitoring disease
progression and checking the effectiveness of treatment(Litjens et al., 2016)..
CNNs can forecast the response of tumors to specific treatment with help of pre-treatment
imaging data. Through the analysis of tumor characteristics change, CNNs play role in
tailoring treatments to specific patients to achieve better outcomes(Parmar et al., 2017).
CNNs can analyze medical images to extract prognostic biomarkers indicative of patient
outcomes. By identifying imaging features associated with disease aggressiveness, CNNs can
aid in predicting patient survival and stratifying risk(Wulczyn et al., 2020).
Accuracy: Accuracy can measure the overall correctness of the classifier. It records the ratio
of correctly classified instances (both positive and negative) that are classified correctly to the
total number of instances, as depicted in the equation (1)
(1)
Sensitivity(Recall): Sensitivity, which for most cases is known as predictive value or true
positive rate, is defined as the probability of getting the correct disease class and it is depicted
in the formula (2).
(2)
Specificity: Specificity is defined as the proportion of actual true negative cases that are
classified as negative by the method. It is depicted as shown in the formula (3).
. (3)
Precision: The measures of precision are the number of instances of positive test results
correctly diagnosed over all the positive tests. It is shown in the formula (4).
(4)
False Positive Rate (FPR): The FPR is a ratio of the proportion of true negative to that of the
corresponding false positive and represented by the mathematical function (5).
(5)
F1 Score: F1 score is the harmonic mean (HM) of precision and recall rates. It gives a
balance between precision and recall, and used when classes are imbalanced. It is represented
by the mathematical function (6).
(6)
Page 7126 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
(7)
(8)
The accuracy, recall, specificity, Kappa statistics and AUCs are the basis of comparison and
are depicted in Tables 2-5, respectively.
Sensitivity 92
Specificity 93
AUC 92
Kappa Score 84
70-30 Accuracy 92
Sensitivity 91
Specificity 92
AUC 91
Kappa Score 83
80-20 Accuracy 91
Sensitivity 86
Specificity 95
AUC 90
Page 7127 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
Kappa Score 80
Sensitivity 82
Specificity 97
AUC 97
Kappa Score 80
Sensitivity 86
Specificity 97
AUC 90
Kappa Score 83
Sensitivity 92
Specificity 95
AUC 95
Kappa Score 87
Sensitivity 92.34
Specificity 96.22
Page 7128 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
AUC 94
Sensitivity 93
Specificity 95
AUC 92
Kappa Score 87
6th Accuracy 93
Sensitivity 87
Specificity 96
AUC 92
Kappa Score 85
Sensitivity 89
Specificity 95.4
AUC 94
Kappa Score 85
8th Accuracy 94
Sensitivity 91
Specificity 96
AUC 93
Kappa Score 87
9th Accuracy 91
Page 7129 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
Sensitivity 87
Specificity 94
AUC 94
Kappa Score 81
10th Accuracy 97
Sensitivity 96
Specificity 98
AUC 93
Kappa Score 94
This research provides insight into how MRI technology plays a crucial role in detecting
brain tumors, with the main focus being to interpret MRI images using deep learning
algorithms and detect cancer cells. The authors aimed to integrate these algorithms into
medical practice and in particular use data from open sources and multiple medical centres.
The authors used a CNN to successfully achieve their goals which they implemented using
the PyTorch machine learning library and coded in Python. We split the data into the 10
fold cross validation and various train-test splits (60-40, 70-30, and 80-20) and retrained the
model for the 30 epochs, with batch size 32 for every epoch.
The results were promising and the model illustrated considerable accuracy rates, especially
97%accuracy(as shown in Table 3) for the final epoch with 10-fold cross-validation. This
accuracy was also similar irrespective of the train-test splits. Moreover, the model manifested
excellent sensitivity, specificity, and kappa scores, approximately 96%, 98%, and 94%
respectively, particularly for the discrimination between benign and malignant nodes. These
performance metrics still remain stable despite the variation of data splits.
Page 7130 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
One of the interesting findings of the research was the recognition of difficult cases where
benign nodules looked like malignant ones. In addition to these difficulties, the algorithm
proved to be quite robust in assignments, displaying its capabilities to recognize benign and
malignant nodules even in visually unclear cases.
Table 6. Accuracies for 80/ 20 train/ test partitions with reduced convolutional and max-
pooling layers
80-20 Sensitivity 32
Specificity 07
Using Explainable Artificial Intelligence (XAI), we could explain the changes in model
performance that are observed when convolutional layers and max-pooling layers are reduced
by 1 as illustrated in Table 6. These techniques enable one to see the impact of changes in the
architecture of the model on different aspects of performance.
Here's a breakdown of the observed changes and potential explanations using XAI principles:
Page 7132 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
Explanation: The marked reduction in the recall shows that the proposed
model with fewer convolutional and max-pooling layers is not as successful in
identifying malignant instances. This is likely because:
Insufficient Feature Learning: However, this may lead to the fact that the
model does not learn rich or deep features that would help to distinguish
between malignant cases and others.
Loss of Fine-Grained Details: As one can gather, the model might neglect
important patterns necessary for classifying malignant cases due to the
simplification of the network architecture.
Feature Overlap: It is possible that there are features that can assist in
discriminating malignancy that are not well learned or represented in the
model, thus misclassifying benign cases as malignant ones.
In general, this research reveals the ability of deep learning algorithms integrated into MRI
technology to correctly diagnose and classify brain cancer as a good instrument for medical
professionals in diagnosis and treatment planning.
Conclusion:
This paper demonstrates how transfer learning, especially the use of a deep convolutional
neural network (CNN), can help with the precise diagnosis of different types of brain cancer
through magnetic resonance imaging (MRI) scans. The model shown in the given paper uses
sigmoid and ReLU activation functions and successfully gets an outstanding score on every
evaluation metric that can be used to measure accuracy. Moreover, this shows how powerful
the model is in correctly classifying the images into their proper classes. The highest
accuracy that was achieved here manifests the model's success in amplifying early diagnosis
and customizing treatment plans for brain cancer patients for clinicians. Such screening is
expected to bring many benefits to the process, starting with the quick and precise
recognition of different types of tumors from MRI scans. This method has the prospect of
optimizing therapeutic outcomes and contributing to early detection that might ultimately
result in improved patient prognosis.
Bringing together several performance metrics like accuracy, sensitivity, specificity, kappa
and AUC (Area Under the ROC Curve) while analyzing brain cancer models is a significant
advance in the field of model evaluation. Normally, researchers usually pay attention to only
one or two metrics that typically comprise of accuracy. On the other hand, such a step might
leave out several other aspects of model performance, including a situation where different
metrics demonstrate various facets of the model's behaviour, especially in medical
applications.
Through the employment of all these metrics, our CNN model of brain cancer detection
offers a holistic assessment of its performance. Achieving the best accuracy across all the
metrics demonstrates our model's capability to excel in different modes of evaluation. This
integrational strategy is not only the evidence of model distinction but also the proof of
feasibility of such concept in practical health-care domain.
Page 7134 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
Finally, the use of advanced deep learning techniques in medical image analysis as presented
in the research indicates the likelihood of the development of early diagnosis and better living
of patients in the context of oncology, especially in the case of brain cancer. This advanced
approach shows us how to harness the potential of artificial intelligence to improve health
systems and thus patients' well-being.
Future Scope
The actual accuracy score of around 98% by the developed model with the help of all
performance metrics suggests best possibilities for future applications in the field of medical
image analysis. The efficacy of this model demonstrates its potential is not limited to other
disease datasets. Instead, it shows the way to the early identification and classification of
numerous other medical conditions.
In the upcoming period, the extent of this research can be augmented by working jointly with
healthcare agencies and hospitals to collect varied datasets. Researching different ways to do
the tasks, including applying reinforcement learning methods, can help to raise the level of
model performance. These may entail tweaking the current model architecture, implementing
new neural network machines, or application of advanced algorithms for the best outcomes.
Furthermore, this model can be developed for present-day diagnosis and treatment
monitoring, which helps in the creation of more personalized and timely interventions. The
integration of artificial intelligence into healthcare is a rapidly growing field, and the success
illustrated in this study gives the basis for further innovation and testing of various solutions
to enhance early disease diagnosis and improve patients' outcomes.
Data Availability Statement (DAS): The dataset used in this study for the brain tumor
classification was acquired from kaggle.com and this does not violate the protection of
human subjects, or other valid ethical, privacy, or security concerns
Data Availability Statement (DAS): The dataset used in this study for the brain tumor
classification was acquired from kaggle.com and this does not violate the protection of
human subjects, or other valid ethical, privacy, or security concerns
References
Afshar, P., Mohammadi, A., &Plataniotis, K. N. (2018, October). Brain tumor type
Ayadi, W., Elhamzi, W., Charfi, I., &Atri, M. (2021). Deep CNN for brain tumor
Cheng, J., Huang, W., Cao, S., Yang, R., Yang, W., Yun, Z., ... & Feng, Q. (2015). Enhanced
Dang, L. M., Hassan, S. I., Im, S., & Moon, H. (2019). Face image manipulation detection
156-168.
Deb, D., Khang, A., & Chaudhuri, A. K. (2024). Fuzzy Thresholding-Based Brain Image
Global.
Deepak, S., & Ameer, P. M. (2019). Brain tumor classification using deep CNN features via
Esteva, A., Kuprel, B., Novoa, R. A., Ko, J., Swetter, S. M., Blau, H. M., &Thrun, S. (2017).
542(7639), 115-118.
Hu, A., &Razmjooy, N. (2021). Brain tumor diagnosis based on metaheuristics and deep
Huang, Z., Du, X., Chen, L., Li, Y., Liu, M., Chou, Y., & Jin, L. (2020). Convolutional
neural network based on complex networks for brain tumor image classification with
Ilic, I., &Ilic, M. (2023). International patterns and trends in the brain cancer incidence and
9(7).
Kumar, S., &Mankame, D. P. (2020). Optimization driven deep convolution neural network
1190-1204.
Litjens, G., Kooi, T., Bejnordi, B. E., Setio, A. A. A., Ciompi, F., Ghafoorian, M., ... &
Litjens, G., Sánchez, C. I., Timofeeva, N., Hermsen, M., Nagtegaal, I., Kovacs, I., ... & Van
Der Laak, J. (2016). Deep learning as a tool for increased accuracy and efficiency of
Machhale, K., Nandpuru, H. B., Kapur, V., &Kosta, L. (2015, May). MRI brain cancer
Mehnatkesh, H., Jalali, S. M. J., Khosravi, A., &Nahavandi, S. (2023). An intelligent driven
deep residual learning framework for brain tumor classification using MRI images.
Özyurt, F., Sert, E., &Avcı, D. (2020). An expert system for brain tumor detection: Fuzzy C-
means with super resolution and convolutional neural network with extreme learning
Parmar, C., Grossmann, P., Bussink, J., &Lambin, P. (2017). Machine learning methods for
125.
Page 7137 of 7136
Sulekha Das / Afr.J.Bio.Sc. 6(5) (2024). 7113-7126
Paul, J. S., Plassard, A. J., Landman, B. A., &Fabbri, D. (2017, March). Deep learning for
Molecular, Structural, and Functional Imaging (Vol. 10137, pp. 253-268). SPIE.
Ryu, Y. J., Choi, S. H., Park, S. J., Yun, T. J., Kim, J. H., & Sohn, C. H. (2014). Glioma:
Saba, T. (2020). Recent advancement in cancer detection using machine learning: Systematic
Sajjad, M., Khan, S., Muhammad, K., Wu, W., Ullah, A., &Baik, S. W. (2019). Multi-grade
brain tumor classification using deep CNN with extensive data augmentation. Journal
Sasikala, M., &Kumaravel, N. (2008). A wavelet-based optimal texture feature set for
198-205.
Sharif, M. I., Khan, M. A., Alhussein, M., Aurangzeb, K., & Raza, M. (2021). A decision
support system for multimodal brain tumor classification using deep learning.
Skogen, K., Schulz, A., Dormagen, J. B., Ganeshan, B., Helseth, E., & Server, A. (2016).
Ting, F. F., Tan, Y. J., & Sim, K. S. (2019). Convolutional neural network improvement for
Verma, R., Zacharaki, E. I., Ou, Y., Cai, H., Chawla, S., Lee, S. K., ... &Davatzikos, C.
966-977.
Wulczyn, E., Steiner, D. F., Xu, Z., Sadhwani, A., Wang, H., Flament-Auvigne, I., ...
Zacharaki, E. I., Wang, S., Chawla, S., Soo Yoo, D., Wolf, R., Melhem, E. R., &Davatzikos,
C. (2009). Classification of brain tumor type and grade using MRI texture and shape
Zia, R., Akhtar, P., & Aziz, A. (2018). A new rectangular window based image cropping