Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Gastro-Esophageal Cancer: Can Radiomic Parameters from Baseline 18F-FDG-PET/CT Predict the Development of Distant Metastatic Disease?
Previous Article in Journal
Clinical and Hemodynamic Features of Aneurysm Rupture in Coil Embolization of Intracranial Aneurysms
Previous Article in Special Issue
A Novel Framework for Data Assessment That Uses Edge Technology to Improve the Detection of Communicable Diseases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Medical Image Retrieval with UMLS-Integrated CNN-Based Text Indexing

1
Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakaka 72388, Saudi Arabia
2
Information Retrieval and Knowledge Management Research Laboratory, York University, Toronto, ON M3J 1P3, Canada
3
Research Laboratory on Development and Control of Distributed Applications (REDCAD), National Engineering School of Sfax, Sfax University, Sfax 3029, Tunisia
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(11), 1204; https://doi.org/10.3390/diagnostics14111204
Submission received: 3 May 2024 / Revised: 30 May 2024 / Accepted: 4 June 2024 / Published: 6 June 2024
(This article belongs to the Special Issue Medical Data Processing and Analysis—2nd Edition)

Abstract

:
In recent years, Convolutional Neural Network (CNN) models have demonstrated notable advancements in various domains such as image classification and Natural Language Processing (NLP). Despite their success in image classification tasks, their potential impact on medical image retrieval, particularly in text-based medical image retrieval (TBMIR) tasks, has not yet been fully realized. This could be attributed to the complexity of the ranking process, as there is ambiguity in treating TBMIR as an image retrieval task rather than a traditional information retrieval or NLP task. To address this gap, our paper proposes a novel approach to re-ranking medical images using a Deep Matching Model (DMM) and Medical-Dependent Features (MDF). These features incorporate categorical attributes such as medical terminologies and imaging modalities. Specifically, our DMM aims to generate effective representations for query and image metadata using a personalized CNN, facilitating matching between these representations. By using MDF, a semantic similarity matrix based on Unified Medical Language System (UMLS) meta-thesaurus, and a set of personalized filters taking into account some ranking features, our deep matching model can effectively consider the TBMIR task as an image retrieval task, as previously mentioned. To evaluate our approach, we performed experiments on the medical ImageCLEF datasets from 2009 to 2012. The experimental results show that the proposed model significantly enhances image retrieval performance compared to the baseline and state-of-the-art approaches.

1. Introduction

Medical information retrieval has a range of applications and solutions connected with better health care. At a basic level, it encompasses image retrieval, retrieval of reports, and natural language queries to databases containing both images and text. However, image retrieval is a challenging task as it can be very subjective, requiring high-level cognitive processing. There are two main types of image retrieval used clinically. One is where the medical professional has a clear idea of what they are looking for and uses the image to seek specific information. The second is a case in which the medical professional has an image and desires to find all similar images to aid diagnosis or as a teaching aid. Step one carries manual work, as tags need to be manually attached to the image usually as metadata. When the images are stored in large databases such as Picture Archiving and Communication Systems (PACS), this can be a highly disorganized and time-consuming task. Step two involves searching using the image as the query and algorithmic methods using visual features of the image attempt to retrieve similar images. As technology advances, there has been increasing support to move to automatic image annotation and content-based retrieval. It is within content-based image retrieval (CBIR) that the model CNNMIR seeks to improve the current state of the art.
Usually in medical domain, images constitute a reference set of previously evaluated cases, that physicians may use to make the right decisions. With the massive growth of medical images, it becomes hard for domain experts to find relevant images in large medical datasets. Thus, the need for an efficient and effective medical image retrieval system becomes urgent [1]. Two main approaches for medical image retrieval are widely used: text- and content-based retrievals. These approaches search for relevant images by using different principles: the text-based approach relies on the high-level semantic features of the images, however, the content-based approach relies on the low-level visual features (e.g., color, shape, and texture) of the image. Comparing both approaches, the Content-Based Medical Image Retrieval (CBMIR) performance is less favorable due to the gap between low-level visual features and high-level semantic features [2,3]. Therefore, several medical image retrieval systems apply the Text-Based Medical Image Retrieval (TBMIR) approach to search for images [4]. Most of these approaches are: either, traditional simple keyword-based approaches; where the meanings of medical entities are ignored, or concept-based approaches; that are time and disk-space consuming. According to our previous works [5,6], the presence of specific medical information, namely Medical-Dependant Features (MDF) in the textual descriptions of medical images has a positive impact on the performance of TBMIR approaches.
In these last years, Convolutional Neural Network (CNN) [7,8] models have shown significant performance improvement in several fields as Natural Language Processing (NLP) [9] and computer vision [10]. Given their success in such fields, it seems to be efficient for image retrieval. Unfortunately, until now, the CNN models have not a significant positive impact on medical image retrieval, especially on text-based medical image retrieval (TBMIR) [11]. It may be due to the complexity of the ranking process: it is not obvious how to consider TBMIR tasks as an image retrieval task [12] and not as a traditional information retrieval task, nor an NLP task. Indeed, the traditional information retrieval systems identify the relevance of a document to a given query; however, the NLP systems deduce the semantic relations between the query and the document. These two systems do not take into account the specificity of images in their processes.
In our previous work [13], we proposed a personalized CNN model that considers the specificity of images in its retrieval process. In that model, we consider the Word2Vec model for word embedding. However, it is well known that the Word2Vec model considers general terminologies, which are not specific for any domain. As our work fits in the medical image retrieval field, we believe that using medical semantic resources, such as UMLS, for converting textual words is more appropriate.
In this paper, we propose a deep matching process for TBMIR that is different from traditional information retrieval and NLP described as follows: first, it takes into account the specificity of images, by mapping the textual queries and the image metadata (document) into MDF. Second, it extracts the semantic relations between MDF, using UMLS, to build a good representation of query and document, and third, it computes the document relevance to the query by using the extracted relations.
In the literature, a variety of deep matching models have been proposed; however, most of them are designed for NLP, rather than information retrieval. Indeed, they consider the semantic matching instead of relevance matching. These models can be categorized, according to their architecture, into two types [14]: the first one is the interaction-focused models [15,16]. These models extract the relationships between queries and documents and then integrate them into a deep neural network to create new matching models. The second one is the representation-focused models [15,17]. These models apply the deep neural network to extract the best representations for both query and document, and then integrate them into a matching process.
In this paper, we propose a new medical image re-ranking process based on a deep matching model (DMM) for TBMIR. Overall, our model is a new representation-focused model that builds a good representation of queries and documents using MDF, UMLS and a personalized CNN for relevance matching. Specifically, we first create the semantic similarity matrix by extracting the UMLS relationships between each pair of MDF. Each query/document MDF is mapped into a similarity vector representing the relationships between the corresponding MDF and all MDF. As each query/document is composed of MDF, the resulting representation of the query/document will be a similarity matrix. Using these matrixes, our model tries to find the best representations for both query and document. Indeed, It applies a personalized CNN which is composed of retrieval filters taking into account some ranking features. Finally, an overall matching score is computed.
We evaluate the effectiveness of the proposed DMM using the ImageCLEF datasets from 2009 to 2012. For comparison, we take into account three well-known traditional retrieval models. The empirical results show that our model significantly outperforms the baseline models in terms of all the evaluation metrics.
This paper is structured as follows: Section 2 summarizes related work. Section 3 describes the proposed DMM model. Section 4 describes, first, how to represent MDF as a semantic similarity matrix using UMLS similarity, and second, our personalized CNN model with specific filters and finally the matching function. Experiments and results are presented and discussed in Section 7. Finally, Section 8 concludes the paper and gives some future work.

2. Related Work

Text-based indexing opens the possibility of indexing medical images by utilizing the associated reports, thereby providing a crucial way to access the exponentially growing clinical image databases. Current methods to retrieve images from medical databases are either based on attribute image content or non-image content. The content-based image retrieval (CBIR) method relies on image features such as shape, texture, or (in most successful cases) a previously assigned semantic feature to retrieve similar images. Non-image content retrieval methods often deal with text-based searches of databases, where a search query is submitted to an image database and text associated with images is compared to the query. However, both methods do not consider the actual medical knowledge relevant to an image and more traditional methods of organizing medical records which are indexed by keywords. Text-based image retrieval by and large builds upon retrieval technique that aims at finding all images relevant to a given query out of large database of images. In most textbook studies in proposed system, a specific image is given as an entry and the user wants to retrieve all images relevant to the given query. Though it is actually a subtype of text-based image retrieval where actual text-based images are not available and it is limited by the availability of some keyword-based annotation of the image. But this method has been shown to be very effective in retrieving images relevant to a given query and hence can be generalized to retrieval systems where text is the main modality. The creation of ’indexes of associated features’ from images is effectively the creation of a searchable database that links images to text.
In the literature, several works studied the use of CNN and semantics in medical image retrieval. This section briefly summarizes some of these works.

2.1. CNN for Medical Image Retrieval

The use of CNN models in the medical image retrieval domain has received great attention [18,19,20]. Authors in [21] used the CNN model based on the bag-of-word (BoW) technique to index biomedical articles. In this particular model, the input is a matrix of numbers that stand for the various medical terms that are contained in the input text. After that, a system of hidden layers is utilized to assign categories to the document. The authors of [22] developed yet another method for the classification of medical texts that may be put to use for retrieval operations. It does this by employing CNN training to extract the semantics of an input sentence; more specifically, it uses the Word2vec technique to represent the input sentences. This method is based on the use of CNNs. During the training of the CNN model, which is comprised of numerous hidden layers, it additionally maintains the list of stop-words. The CNN model was employed by the authors in the cited article [23] to remove the background noise from clinical notes that were going to be used for medical literature retrieval. They represented the input questions by using GloVe vectors, which are cited in the following reference: [24]. The CNN model’s primary purpose in this study is to make predictions about the relative relevance of search query phrases.
Despite the success of CNN for computer vision and NLP, employing CNN to search for relevant documents in TBMIR is not effective; and this may be due to the complexity of the ranking process. Moreover, most existing CNN models represent queries and documents without taking into account the specificity of the medical domain. This latter requires semantic extraction using external medical resources.

2.2. Semantics in Medical Image Retrieval

The integration of semantic knowledge in the medical image retrieval domain has received great attention, such as [25,26,27]. Authors in [28] used UMLS meta-thesaurus in the medical domain to improve queries and converting words to medical terms. They integrated the semantics in the retrieval process to map the text into concepts using UMLS meta-thesaurus [29]. The authors of the cited paper [30] developed a retrieval method to discover discriminative qualities between various medical photos using a Pruned Dictionary that was based on a description of a Latent Semantic Topic. They did this by calculating the topic-word relevance, which allowed them to make a prediction about the word’s relationship to the underlying topic. The latent themes are learned based on the association between the images and the words, and they are used to bridge the gap between low-level visual features and high-level semantic characteristics. This is accomplished by bridging the gap between low-level visual features and high-level semantic features. Moreover, in [31], an image retrieval framework that is based on semantic features has been proposed by the authors. This framework relies on (1) the automatic prediction of ontological terms that define the image content and (2) the retrieval of similar images by analyzing the similarity between annotations. The study of this system demonstrated that it is beneficial to make use of ontology while retrieving medical images.
Despite the large number of works using CNN and semantic resources in medical image retrieval, there is a lack of studies that investigate the integration of semantic knowledge on the CNN model to enhance the medical image retrieval performance. Therefore, we propose a new deep matching model based on personalized CNN and semantic resources (MDF, UMLS) to improve retrieval accuracy.

3. Overview of Our Approach

It is well known that medical images and their associated reports are not usually in agreement. For example, a patient who has a slipped disc, but displays no symptoms, will not have many abnormalities in his MR scan but will have many associated words or phrases about his condition. This inconsistency is a major problem for medical image retrieval systems, which rely on the images and associated text being “about the same thing”. With this in mind, and the fact that we have indexed the images and text separately, we need to devise a way where a text query can be used to aid the image query and vice-versa, without the user having to switch between the two.
Our proposed solution leverages relevance feedback to integrate information from both image and text queries using the other modality. For instance, if a user seeks to locate an image corresponding to a report about ‘left lung cancer’, the current system requires them to separately index the text using a natural language processing (NLP) tool and formulate a query, then repeat the process for the image. This method is inefficient and requires users to switch between modalities. In contrast, our system enables users to use an NLP tool to index the text or query, subsequently identifying relevant images that correspond to the text. Modality-specific technology subsequently ranks the images based on their similarity to the text. This approach automates the task of ’finding images matching this report,’ enhancing efficiency and accuracy.
Due to the positive impact of MDF on both retrieval performance [6,13] and query classification [32,33], we choose to integrate them into a deep matching model. In this study, we utilized the Unified Medical Language System (UMLS) as our semantic resource to construct a semantic similarity matrix, which represents the relationships between pairs of Medical Dependent Features (MDF). The literature [34,35,36] widely recognizes UMLS as a comprehensive thesaurus and ontology of biomedical concepts, designed to link various biomedical terminologies. By leveraging UMLS, we ensure a robust semantic framework for our analysis. Additionally, our system allows users to index text or queries using natural language processing (NLP) tools, facilitating more accurate and efficient retrieval of relevant medical information.
A new personalized CNN model using MDFs is proposed to build the best representation for both queries and documents, that are used to compute their matching score. In this paper, Figure 1 presents an overview of our approach:
  • The preliminary step:
We represent each query/document as a set of features, then, for each MDF ( F i ), we assign the corresponding vector extracted from the similarity matrix. Hence, each query/document is represented as a Matrix.
  • The Deep Matching Model Process:
First, we build a good representation of the query/document with a personalized CNN model that takes into account the interaction between query and document. Indeed, several personalized filters have been proposed and integrated into this model. Then, a matching function is applied to measure the matching degree between the query and the document representations. More precisely, we use the cosine similarity function as a matching function.
Second, we combine the obtained score with the corresponding baseline score to form a new re-ranking score. The re-ranking process is achieved by sorting the images according to their new scores.

4. Deep Matching Model: Preliminary Step

4.1. Medical Dependent Features

As our work falls into the medical image retrieval field, we propose to integrate the MDFs [6,32], that are a set of categorical medical features, into a new deep matching model to enhance the retrieval performance. A medical dependent features presented in the Figure 2.
Each MDF f i has m associated values v defined by f i = v 1 , v 2 , v m . The set of MDF used in our work is detailed as follows:
  • Radiology = “Ultrasound Imaging”, “Magnetic Resonance Imaging”, “Computerized Tomography”, “X-Ray”, “2D Radiography”, “Angiography”, “PET”, “Combined modalities in one image”, “Coronarography”, “Cystography”, “Scintigraphy”, “Mammography”, “Bone Densitometry”, “Radiotherapy”, “Urography”, “Pelvic Ultrasound”, “Myelography”, “FibroScan”
  • Microscopy = “Light Microscopy”, “Electron Microscopy”, “Transmission Microscopy”, “Fluorescence Microscopy”, “Biopsy”, “Stool Microscopy”, “Capillaroscopy”, “Trophoblast Biopsy”, “Cytology”
  • Visible light photography = “Dermatology”, “Skin”, “Endoscopy”, “Other organs”, “Colposcopy”, “Cystoscopy”, “Hysteroscopy”
  • Printed signals and waves = “Electroencephalography”, “Electrocardiography”, “Electromyography”, “Holter”, “Audiometry”, “Urodynamic Assessment”
  • Generic Biomedical Illustrations = “modality tables and forms”, “program listing”, “statistical figures”, “graphs”, “charts”, “screen shots”, “flowcharts”, “system overviews”, “gene sequence”, “chromatography”, “gel”, “chemical structure”, “mathematics formula”, “non-clinical photos”, “hand-drawn sketches”
  • Dimensionality = “macro”, “micro”, “small”, “gross”, “combined dimensionality”
  • V-Spec = “brown”, “black”, “white”, “red”, “gray”, “green”, “yellow”, “blue”, “colored”
  • T-spec = “finding”, “pathology”, “differential diagnosis”, “Amniocentesis”, “Hemogram”, “Non-Invasive Prenatal Screening”, “Urinalysis”, “Lumbar Puncture”, “Seminogram”, “Triple Test”
  • C-spec = “Histology”, “Fracture”, “Cancer”, “Benign”, “Malignant”, “Tumor”, “Pregnancy”, “Antibiogramme”

4.2. Semantic Matrix Construction

In this section, we present a new semantic mapping method using two semantic resources: MDF and UMLS. Frequently in NLP, the text data is converted into a vector of numbers, which deep models can process as input. Several approaches, such as Word2Vec, Glove, and one-hot-encoding, have been proposed for word embedding. Usually, these models consider general terminologies, which are not specific for any domain, to derive similarities and relations between words. As our work fits in the medical image retrieval field, we believe that retrieval performance could be improved if we use medical semantic resources such as UMLS for converting textual words. We represent the queries and documents as a set of MDF to keep only semantic information related to the medical domain. Then, each MDF is transformed into a concept using the MetaMap tool, then the UMLS Similarity tool [37] is used to calculate the similarities between each pair of concepts and then construct the similarity matrix as shown in Figure 1.
As shown in the preliminary step of Figure 1, all features are transformed into a similarity matrix and thus by following the next steps:
  • Step 1: the MetaMap tool [38] is used to transform each MDF into a concept.
  • Step 2: the similarities between each pair of medical concepts are calculated using the UMLS Similarity tool [37,39]. These semantic similarity scores are arranged in a semantic matrix. More precisely, we use the Resnik measure to determine the semantic relations between extracted concepts, as according to [40], it performs better than Path-based measures.

5. Deep Matching Model Construction

The new DMM model is a representation focused model that should build a good representation for a query and document with a deep neural network and conduct matching between the corresponding representations. Moreover, this model should take into account the specificity of information retrieval, NLP and medical image retrieval.
The inputs to our DMM model are a set of queries and documents presented with MDFs; each MDF is transformed into concepts then to a vector of numbers to be processed by the subsequent layers of the network. In the following, we detail the main components of our DMM model: the query/document matrix extraction, the personalized CNN and the matching function.

5.1. Query and Document Matrix Extraction

As our work fits in the medical image retrieval field, we represent the queries and the documents as a set of MDF in order to keep only semantic information related to the medical domain. In this paper, we propose to convert each query and document into an MDF vector. Then, each vector is converted to a semantic similarity matrix as presented in Figure 3:
  • Step 1: For each query/document vector, we assign a binary value for each MDF depending on whether the query/document contains the feature value or not. The length of the resulting vector V equals n where n is the number of MDFs. This vector is transformed into a n n matrix M / i n , j n , M [ i ] [ j ] = V [ i ] where i represents the row index and j represents the column index.
  • Step 2: we multiply the resulting matrix M with the semantic similarity matrix SSM to obtain a new query matrix NQM as follows:
    N Q M [ i ] [ j ] = M [ i ] [ j ] S S M [ i ] [ j ]
    The illustration of the calculation is done in Figure 3.

5.2. Personalized CNN

We present, in this section, the personalized CNN that explicitly addresses the three specificities mentioned above. Indeed, the filters are designed to extract the best representation of queries and documents. In each representation, the network considers several retrieval features such as the MDF co-occurrence, the document ranking, and the IPM score. Moreover, it considers the NLP features for each query/document representation as it extracts the interaction between document and query. Indeed, according to [14], most of NLP models extract the interaction between two texts.
In the following, we present the layers of our network: convolutional, activations, pooling and fully connected layers.
Figure 4 Presents the architecture of the personalized CNN model.

5.2.1. Convolutional Layer

In this layer, a set of filters F R d are applied to the query and document vectors to produce different feature maps. In our model, the query filters are distinct from the document filters. Below, we provide detailed information on the filters used for each component (document and query).
  • Query Filters:
The query filters aim to extract the best representation of the queries by considering the relationship between the document and the query. The more relevant the document is to the query, the higher the resulting vector values will be.
  • Confidence Query Filter (CoQF): The idea consists of calculating the co-occurrences of query MDFs with all MDFs.
    C o Q F = j Q i D f r ( f i , f j ) i D f r ( f i )
    where Q is the query MDF, D is the document MDFs, f r ( f j ) is the cooccurrence of query MDFs in the collection, f r ( f i ) is the cooccurrence of document MDFs in the collection and f r ( f i , f j ) is the cooccurrence of query MDF and document MDF in the collection.
    In order to take into consideration, the length of the document, we use this filter. A document having only the query MDF should be more relevant than a document having other MDF in addition to the query ones. In fact, both documents are specific but the first document is more exhaustive. For that, we propose to divide the number of MDF in both document and query, with the number of document MDF. If the document did not include any query MDF, then the value will 0.
  • Length Query Filter (LQF): For each query, if the document contains all query MDF, then we divide the number of MDF in both document and query, with the number of document MDF. Else, the value will be equal 0.
    L Q F = M D F ( Q , D ) M D F D
    where M D F ( Q , D ) is the number of MDFs in both query MDF Q and document MDF D and M D F D is the number of MDF in the document containing all query MDF.
  • Rank Query Filter (RQF): We calculate the inverse document rank. If the document did not appear in the first search, the RQF will be equal.
    R Q F = 1 d o c r a n k
  • Proximity Query Filter (PQF): IIn the event that a document has query MDFs, we will compute the inverse of the distances that separate these MDFs in the document. In this instance, the distance between two features is represented by the total number of features that are located between them.
    P Q F = 1 1 + d i s t M D F D
    where d i s t M D F D is the distances between document MDFs.
  • PMI Query Filter (PMIQF): The PMI (Pointwise Mutual Information) [41] is a proposed metric to find features with a close meaning. Indeed, the PMI of the MDFs f i and f j is defined using the occurrences of f i ( f r ( f i ) ) and f j ( f r ( f j ) ), the co-occurrences f r ( f i , f j ) within a vector of features, and N is the collection size.
    P M I F ( Q F ) = l o g N × f r ( f i , f j ) f r ( f i ) × f r ( f j )
    This equation calculates the semantically closest MDFs of the collection to f i and f j .
  • Feature Difference Query Filter (FDQF): The more the query MDFs not found is small, the more the document is relevant. For each query, we compute the inverse of number of query MDFs not in document MDFs.
    F D Q F = 1 1 + M D F Q D
  • Document Filters:
Similar to the query filters, document filters try to extract the best representation of documents. They are based on the relationship between document and query. The more the document is relevant to the query, the highest is the resulting vector values.
  • Confidence Document Filter (CoDF): This document filter determines the total amount of MDF documents that are included in the query. The relevance of the document will increase in proportion to the number of query MDFs it contains.
    C o D F = f i q ̲ f j d ̲
    where f i q ̲ f j d ̲ is the number of common MDF in query.
  • Length Document Filter (LDF): When it comes to documents, first we determine the number of document MDFs that are included in the related query, and after we have that amount, we divide it by the document length ( L D ). In point of fact, the relevance of the document will increase if it is of a modest size and if it shares several characteristics with the query being conducted.
    L D F = M D F d o c i n q u e r y L D
    where M D F d o c i n q u e r y is the number of MDF in both document and query and L D is the document length using the MDF features.
  • Rank Document Filter (RDF):
    R D F = i q f r ( f i i n d o c ) × γ
    The variable f r ( f i ) represents the frequency of query MDFs in the document, while γ represents the organization factor of the query in the document. The value of γ is 1 if the query preserves its organization in the document, and 0.5 if it does not.
  • Proximity Document Filter (PDF): The more the document’s features existing in the query are closer, the more it is relevant.
    P D F = 1 f i Q
    where F D Q is the documents MDFs in the query.
  • PMI Document Filter (PMIDF): Similar to PMI in query filter, PMI in document filter try to find MDFs with a close meaning. It has the same equation except the N in this filter is the document size.
    P M I F ( D F ) = l o g N × f r ( f i , f j ) f r ( f i ) × f r ( f j )
    This equation calculates the semantically closest MDFs in the document.
  • Feature Difference Document Filter (FDDF): The more the number of document MDFs not in the query is small, the more the document is relevant.
    F D D F = 1 1 + M D F D Q
    where D is the document MDFs and Q is the query MDFs.
The input of the DMM model is a matrix S R n × n , and the convolutional filters are also matrices F R n . It is important to note that these filters have the same dimensionality, denoted as n, as the input matrix. In addition, these filters scan the vector representations and produce an output vector C R m . Each component c i of the vector C is obtained by multiplying a vector V with a filter F, and then summing the resulting values to obtain a single value.
c i = k = 1 n V k F k

5.2.2. Activation Function

Immediately after the convolutional layer comes a non-linear activation function called a l p h a that is applied to the output of the layer that came before it. Through the use of this function, it is possible for a neuron’s input signal to be transformed into an output signal. In the research that has been done, a number of different activation functions have been proposed [42]. One of these functions is called the Rectified Linear Unit (ReLU) function, and it assures that positive values are passed on to the subsequent layer. The authors in [43] demonstrated that this function is effective, uncomplicated, and has the capacity to lower the amount of complexity as well as the amount of time required for calculations. As a result, we have decided to include this function in our model in the capacity of an activation function.

5.2.3. Pooling Layer

The pooling layer’s goal is to do three things: aggregate information, minimize the amount of representation used, and derive global features from the convolutional layer’s local ones. There are two functions that can be found in the body of literary work: (1) the average consists of computing the average of each feature map of the convolutional layer to consider all the elements of the input are even if many of them have low weights [44], and (2) the Max consists of selecting the maximum value of each feature map of the convolutional layer. Both of these operations are performed in order to take into consideration all of the elements of the input are. We have decided to adopt max-pooling for our research because it takes into account only neurons with high activation values, which ultimately results in a high level of semantic abstraction of the input data.

5.2.4. Fully Connected Layer

In order to produce a final vector representation of the query or document, a Fully Connected Layer (FCL) is applied to the vector that was generated as a result of the previous step.

5.3. Matching Function

According to [14], the most significant challenge associated with the retrieval of information is the matching problem, which refers to the process of determining a document’s relevancy in light of a query. If we have a document denoted by d and a query denoted by q, then the matching function is a mechanism for assigning a score to the representation of d and q:
R S V ( d , q ) = F ( Φ ( d ) , Φ ( q ) )
where F stands for the scoring function and P h i is the mapping function that converts each d|q pair into a vector representation. In the research that has been done on the subject, a number of different deep matching models have been suggested for the overall matching process. These silhouettes fall primarily into one of two categories when grouped together. The representation-focused model is the first one, and in this model, P h i is a complicated mapping function while F is a straightforward scoring function. A deep neural network is utilized by this model in order to construct an accurate representation for the document as well as the query. After that, it does some sort of matching between these different representations. The second one is the interaction-focused model where Φ is simple mapping function and F is complex scoring function.
We use a representation-focused model in which P h i is a sophisticated mapping function between representations and F is a straightforward matching function. Since the sophisticated P h i -based mapping function of the individualized CNN is what drives our selection, we resort to the more elementary F-based cosine similarity. The formal definition of a document’s relevance to a query is as follows:
R S V ( Q , D ) = c o s i n e ( Q , D ) = Q . D Q . D
where Q and D are the query and the document vectors respectively. In the IR, for a given query, the documents are ranked by their relevance scores.

6. SemRank: Semantic Re-Ranking Model Based on DMM

In the last part of this article, we discussed our MDF-based deep matching model, which calculates the DMM score of the document d with respect to the query q. However, doing a search of relevant documents by utilizing MDF alone is insufficient; certain phrases could not be mapped to MDF, and as a result, such keywords should be eliminated from the search. As a result, we recommend combining the findings of the DMM with those of the baseline, taking into consideration all of the query terms. To be more specific, we suggest modeling the SemRank score using the most common type of late fusion approach, which is known as a straightforward linear combination. Before adding the two scores together, we first standardize the initial score and the DMM score in the following manner:
S e m R a n k s c o r e = α I n i t i a l S c o r e max I n i t i a l S c o r e s + ( 1 α ) D M M S c o r e max D M M S c o r e
where α is a balancing parameter α [ 0 1 ] , InitialScore represents the initial ranking score of the document and DMM score of the same document. The normalized score is obtained by dividing the relevance score for a given document d by the highest relevance score in the whole collection. As a baseline, we propose to use the BM25 model which is well known for its efficiency and its performance in many retrieval tasks

7. Experiments and Results

In this section, we describe the experimental datasets, then we present our several experiments released to evaluate the accuracy of our model and we compare it to some existing approaches.

7.1. Experimental Datasets

In order to assess the effectiveness of our suggested method, it is imperative to utilize medical image datasets that include both images and textual descriptions, together with queries and ground truth. The majority of medical data sets currently available do not fulfill these criteria. Some sources lack assessment protocols, such as OHSUMED [45], while others focus on textual analysis and evaluation, like TREC. On the other hand, the ImageCLEFmed evaluation campaign offers specific medical picture collections for the purpose of assessing medical image retrieval. From 2011 onwards, the quantity and extent of the collections were comparable to those seen in real-world applications [46]. Due to copyright restrictions, the redistribution of the ImageCLEFmed collections to research groups is only allowed through a special agreement with the original copyright holders [47]. Therefore, we are restricted to conducting experiments using only the five collections for which we have obtained copyrights. The collections are shown in Table 1 and consist of two relatively small data sets: 74,902 and 77,495 images for the 2009 [48] and 2010 [49] data sets, respectively. After the evolution of ImageCLEF, three additional data sets were added: 230,088 images for the 2011 [50] data set, and 306,539 images for both the 2012 [51] and 2013 [52] data sets.
Each image in these data sets is accompanied by a textual description. An image can have the text from its caption or a hyperlink to the HTML page that has the complete text of the article [53], together with the title of the article [48]. Furthermore, the queries were chosen from a collection of themes suggested by physicians and clinicians in order to precisely mimic the information requirements of a clinician involved in diagnostic tasks. The images in the 2009 and 2010 data sets were sourced from the RSNA journals Radiology and Radiographics [48] and consist of a portion of the Goldminer data set. Nevertheless, the photos in the data sets from 2011, 2012, and 2013 are derived from studies that were published in open-access journals and may be accessed through PubMed Central. The later data sets have a wider range of images, including charts, graphs, and other nonclinical images, resulting in more visual variety.
To evaluate the proposed SemRank model, we conducted experiments using the ImageCLEF collections in Medical Retrieval Task from 2009 to 2012. These collections are composed of images and queries. Each image has a textual description. Each query is composed of text representation and a few sample images. In our work, we use only text representation of the queries and textual description of the images. The 2009 and the 2010 datasets are relatively small (74,902 and 77,495 images respectively). The 2011 and the 2012 datasets are significantly bigger (230,088 and 306,539 images respectively). Indeed, these datasets contain a greater image diversity and also include charts, graphs and other non-clinical images [32]. Also, we are limited to these datasets as we do not have the 2013 dataset, the last dataset in the medical image retrieval task of imageCLEF. Figure 5 illustrates 3 images of the used datasets, respectively with their associated MDF, extracted using the Medical-Dependent Features. We observe here that MDF features represent specific characteristics of medical images but not a body part (brain) or a pathology (cancer) and this due to the nature of medical textual queries aiming to find medical images.

7.2. Effectiveness of the SemRank Model in Image Reranking

In this section, we present a set of experiments carried out to the SemRank model. To achieve the best linear combination, we use several values of α . α = 0 means that only the DMM score is used and α = 1 means that only the BM25 score is used. Figure 6 presents the MAP, the P@5 and the P@10 values when: α [ 0 : 1 ] in datasets from 2009 to 2012.
According to Figure 6, we notice that using only DMM model to retrieve relevant documents gives the worst ranking results. However, the combination of the baseline and the DMM models gives better results. This proves our assumption that using only MDF to search for relevant documents is not sufficient; the combining models is a solution. According to MAP, P@5 and P@10 values, the best results are obtained when α [ 0.1 : 0.5 ] . In the remaining experiments, we chose to set α = 0.3.

7.3. Comparison of the SemRank Model with Literature Models

In this section, we propose to compare our model with BM25, DLM (Dirichlet Language Model) [54] and Bo1PRF (Bo1 Pseudo Relevance Feedback) [55] models. Table 1 summarizes this comparison according to the P@5, P@10 and MAP measures. The best result of all models and for each metric is presented in bold.
For the 2009 and the 2010 ImageClef datasets, the results show that our SemRank model performs better than the existing models in terms of MAP, P@5 and P@10. For the 2011 dataset, the SemRank model gives better results than the BM25 and the DLM models in terms of P@5 and P@10, but do not outperform the Bo1PRF model in term of MAP. Moreover, For the 2012 dataset, the Bo1PRF model outperforms our model in terms of MAP, P@5 and P@10. This can be explained by the high number of non-clinical images in these datasets which contain a diversity of images (tables, shapes, graphs); and our retrieval model is specific for medical images. Moreover, the Bo1PRF model is based on the pseudo-relevance feedback technique that improves retrieval results.
The accuracy gain is presented in Table 2. Indeed, we determine the improvement rate and we conduct a statistical t-test (Wilcoxon) [56] to evaluate the results. The gain is considered statistically significant when p < 0.05 . In this work, the results are followed by the ** when p < 0.05 .
Results show that the improvements have been achieved on the majority of datasets.Our model achieves between 12% and 29% on the 2009 dataset, which is a substantial improvement over the performance of existing models. When compared to the DLM and Bo1PRF models, the retrieval performance of the 2010 dataset is significantly enhanced by the SemRank model. This might be explained by the fact that the datasets from 2009 and 2010 contain photos suggested by clinicians and physicians that beat the information that is required.
Although our model is performing worse than the Bo1PR model on the 2011 and the 2012 datasets, it improves the retrieval performance (+4 percentage points) compared to the BM25 model for the 2012 dataset and (+40 percentage points) compared to the DLM model for the 2011 dataset. This variation may be due to the pseudo relevance feedback technique, which adds the first m keywords that appear in the top k retrieved documents. However, our model uses only query features without additional terms and enhances significantly the retrieval performance on the 2009 and the 2010 datasets.
We conclude that using our DMM improves significantly the results compared to the literature models. This validates our assumption that our DMM is a promising technique for improving medical image retrieval performance. In addition, this improvement could be related to the importance of using medical external resources: MDF and UMLS.

8. Conclusions and Future Work

This paper introduces an innovative SemRank model designed to enhance the ranking of medical images. The model leverages two external semantic resources: the Medical-Dependent Features (MDF) terminology and the Unified Medical Language System (UMLS) Metathesaurus. Within this framework, queries and documents are represented as sets of MDF, with the UMLS ontology employed to compute semantic similarity matrices between these sets. These matrices serve as the foundation for constructing matrix representations for each query and document, which are subsequently integrated into a Convolutional Neural Network (CNN) process. The resulting outputs yield vectors used to compute new relevance scores for documents when presented with a query. This innovative approach not only harnesses semantic knowledge from external resources but also employs advanced neural network techniques to improve the accuracy and effectiveness of medical image retrieval.
Our experiments were conducted on the Medical ImageCLEF collections from 2009 to 2012. The findings demonstrate a significant improvement in the re-ranking process when integrating Medical-Dependent Features (MDF) and the Unified Medical Language System (UMLS) into the Deep Matching Model (DMM). Furthermore, a comparative analysis was conducted between our model and various state-of-the-art approaches. The results revealed a noteworthy increase in the accuracy of the re-ranking process, underscoring the efficacy of our proposed methodology.
In our forthcoming research endeavors, we aim to augment the capabilities of the CNN model by integrating supplementary filters that encompass a broader spectrum of retrieval attributes. Furthermore, we plan to enhance the SemRank model by incorporating visual features, thereby elevating the precision of image retrieval.

Author Contributions

Conceptualization, K.G.; methodology, K.G. and H.A.; software, K.G. and H.A.; validation, H.A. and M.T.; formal analysis, K.G., H.A. and M.T.; resources, K.G. and H.A.; data curation, K.G. and H.A.; writing—original draft preparation, K.G. and H.A.; writing—review and editing, M.T. and H.A.; visualization, K.G. and H.A.; funding acquisition, K.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No. (DGSSR-2023-02-02065).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are openly accessible at the following link: https://www.imageclef.org.

Acknowledgments

This work was funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No. (DGSSR-2023-02-02065).

Conflicts of Interest

All authors declare that there are no competing interests.

References

  1. Kelishadrokhi, M.K.; Ghattaei, M.; Fekri-Ershad, S. Innovative local texture descriptor in joint of human-based color features for content-based image retrieval. Signal Image Video Process. 2023, 17, 4009–4017. [Google Scholar] [CrossRef]
  2. Smeulders, A.; Worring, M.; Santini, S.; Gupta, A.; Jain, R. Content-based image retrieval at the end of the early years. IEEE Trans. Patt. Anal. Mach. Intell. 2000, 22, 1349–1380. [Google Scholar] [CrossRef]
  3. Anandh, A.; Mala, K.; Suganya, S. Content based image retrieval system based on semantic information using color texture and shape features. In Proceedings of the International Conference of Computing Technologies and Intelligent Data Engineering (ICCTIDE), Kovilpatti, India, 7–9 January 2016; pp. 1–8. [Google Scholar]
  4. Moon, J.H.; Lee, H.; Shin, W.; Kim, Y.H.; Choi, E. Multi-modal understanding and generation for medical images and text via vision-language pre-training. IEEE J. Biomed. Health Inform. 2022, 26, 6070–6080. [Google Scholar] [CrossRef]
  5. Ayadi, H.; Torjmen, M.K.; Huang, J.X.; Daoud, M.; Ben Jemaa, M. Learning to Re-rank Medical Images Using a Bayesian Network-Based Thesaurus. In Proceedings of the 39th European Conference on IR Research (ECIR), Aberdeen, UK, 8–13 April 2017; pp. 160–172. [Google Scholar]
  6. Ayadi, H.; Torjmen-Khemakhem, M.; Daoud, M.; Huang, J.X.; Ben Jemaa, M. MF-Re-Rank A modality feature-based Re-Ranking model for medical image retrieval. J. Assoc. Inf. Sci. Technol. 2018, 69, 1095–1108. [Google Scholar] [CrossRef]
  7. Krichen, M. Convolutional neural networks: A survey. Computers 2023, 12, 151. [Google Scholar] [CrossRef]
  8. Shamsipour, G.; Fekri-Ershad, S.; Sharifi, M.; Alaei, A. Improve the efficiency of handcrafted features in image retrieval by adding selected feature generating layers of deep convolutional neural networks. Signal Image Video Process. 2024, 18, 2607–2620. [Google Scholar] [CrossRef]
  9. Abas, A.R.; Elhenawy, I.; Zidan, M.; Othman, M. BERT-CNN: A Deep Learning Model for Detecting Emotions from Text. Comput. Mater. Contin. 2022, 71, 2943–2961. [Google Scholar]
  10. Yang, K.; Ding, Y.; Sun, P.; Jiang, H.; Wang, Z. Computer vision-based crack width identification using F-CNN model and pixel nonlinear calibration. Struct. Infrastruct. Eng. 2023, 19, 978–989. [Google Scholar] [CrossRef]
  11. Arkin, E.; Yadikar, N.; Xu, X.; Aysa, A.; Ubul, K. A survey: Object detection methods from CNN to transformer. Multimed. Tools Appl. 2023, 82, 21353–21383. [Google Scholar] [CrossRef]
  12. Ayadi, H.; Torjmen-Khemakhem, M.; Huang, J.X. Term dependency extraction using rule-based Bayesian Network for medical image retrieval. Artif. Intell. Med. 2023, 140, 102551. [Google Scholar] [CrossRef]
  13. Souissi, N.; Ayadi, H.; Torjmen, M.K. Text-based Medical Image Retrieval using Convolutional Neural Network and Specific Medical Features. In Proceedings of the 12th International Conference on Health Informatics HEALTHINF, Prague, Czech Republic, 22–24 February 2019; pp. 78–87. [Google Scholar]
  14. Guo, J.; Fan, Y.; Ai, Q.; Croft, W.B. A deep relevance matching model for ad-hoc retrieval. In Proceedings of the Conference on Information and Knowledge Management (CIKM), Indianapolis, IN, USA, 24–28 October 2016; pp. 55–64. [Google Scholar]
  15. Hu, B.; Lu, Z.; Li, H.; Chen, Q. Convolutional neural network architectures for matching natural language sentences. In Proceedings of the 27th International Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 8–13 December 2014; pp. 2042–2050. [Google Scholar]
  16. Lu, Z.; Li, H. A deep architecture for matching short texts. In Proceedings of the 26th International Conference on Neural Information Processing Systems (NIPS), Lake Tahoe, NV, USA, 5–10 December 2013; pp. 1367–1375. [Google Scholar]
  17. Huang, P.-S.; He, X.; Gao, J.; Deng, L.; Acero, A.; Heck, L. Learning deep structured semantic models for web search using clickthrough data. In Proceedings of the 22nd ACM International Conference on Information and Knowledge Management (CIKM), San Francisco, CA, USA, 27 October–1 November 2013; pp. 2333–2338. [Google Scholar]
  18. Chen, C.; Li, X.; Zhang, B. Research on image retrieval based on the convolutional neural network. In Proceedings of the 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 14–16 October 2017. [Google Scholar]
  19. Qiu, C.; Cai, Y.; Gao, X.; Cui, Y. Medical image retrieval based on the deep convolution network and hash coding. In Proceedings of the 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 14–16 October 2017. [Google Scholar]
  20. Saminathan, K. Content Based Medical Image Retrieval Using Deep Learning Algorithms. J. Data Acquis. Process. 2023, 38, 3868. [Google Scholar]
  21. Rios, A.; Kavuluru, R. Convolutional neural networks for biomedical text classification: Application in indexing biomedical articles. In Proceedings of the 6th ACM Conference on Bioinformatics, Computational Biology and Health Informatics, Atlanta, GA, USA, 9–12 September 2015; pp. 258–267. [Google Scholar]
  22. Hughes, M.; Li, I.; Kotoulas, S.; Suzumura, T. Medical text classification using convolutional neural networks. Stud. Health Technol. Inform. 2017, 235, 246–250. [Google Scholar] [PubMed]
  23. Soldaini, L.; Yates, A.; Goharian, N. Denoising Clinical Notes for Medical Literature Retrieval with Convolutional Neural Model. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 2307–2310. [Google Scholar]
  24. Pennington, J.; Socher, R.; Manning, C. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  25. Guo, K.; Liang, Z.; Tang, Y.; Chi, T. SOR: An optimized semantic ontology retrieval algorithm for heterogeneous multimedia big data. J. Comput. Sci. 2018, 28, 455–465. [Google Scholar] [CrossRef]
  26. Singh, P.; Goudar, R.H.; Rathore, R.; Srivastav, A.; Rao, S. Domain ontology based efficient image retrieval. In Proceedings of the 2013 7th International Conference on Intelligent Systems and Control (ISCO), Coimbatore, India, 4–5 January 2013; pp. 445–452. [Google Scholar]
  27. Kumar, S.; Singh, M.K.; Mishra, M. Efficient Deep Feature Based Semantic Image Retrieval. Neural Process. Lett. 2023, 55, 2225–2248. [Google Scholar] [CrossRef]
  28. Kumar, K.S.; Deepa, K. Medical query expansion using UMLS. Indian J. Sci. Technol. 2016, 9, 1–6. [Google Scholar]
  29. Ivanović, M.; Budimac, Z. An overview of ontologies and data resources in medical domains. Expert Syst. Appl. 2014, 41, 5158–5166. [Google Scholar] [CrossRef]
  30. Zhang, F.; Song, Y.; Cai, W.; Hauptmann, A.G.; Liu, S.; Pujol, S.; Kikinis, R.; Fulham, M.J.; Feng, D.D.; Chen, M. Dictionary pruning with visual word significance for medical image retrieval. Neurocomputing 2016, 177, 75–88. [Google Scholar] [CrossRef] [PubMed]
  31. Kurtz, C.; Depeursinge, A.; Napel, S.; Beaulieu, C.F.; Rubin, D.L. On combining image-based and ontological semantic dissimilarities for medical image retrieval applications. Med. Image Anal. 2014, 18, 1082–1100. [Google Scholar] [CrossRef] [PubMed]
  32. Ayadi, H.; Torjmen, M.K.; Daoud, M.; Ben Jemaa, M.; Huang, J.X. Correlating Medical-dependent Query Features with Image Retrieval Models Using Association Rules. In Proceedings of the 22nd ACM International Conference on Information and Knowledge Management (CIKM), San Francisco, CA, USA, 27 October–1 November 2013; pp. 299–308. [Google Scholar]
  33. Ayadi, H.; Torjmen-Khemakhem, M.; Daoud, M.; Huang, J.X.; Ben Jemaa, M. Mining correlations between medically dependent features and image retrieval models for query classification. J. Assoc. Inf. Sci. Technol. 2017, 68, 1323–1334. [Google Scholar] [CrossRef]
  34. Bodenreider, O. The unified medical language system (UMLS): Integrating biomedical terminology. Nucleic Acids Res. 2004, 32 (Suppl. S1), D267–D270. [Google Scholar] [CrossRef]
  35. Luo, Y.; Uzuner, O. Semi-supervised learning to identify UMLS semantic relations. AMIA Summits Transl. Sci. Proc. 2014, 2014, 67–75. [Google Scholar]
  36. Tran, L.T.T.; Divita, G.; Carter, M.E.; Judd, J.; Samore, M.H.; Gundlapalli, A.V. Exploiting the UMLS Metathesaurus for extracting and categorizing concepts representing signs and symptoms to anatomically related organ systems. J. Biomed. Inform. 2015, 58, 19–27. [Google Scholar] [CrossRef] [PubMed]
  37. McInnes, B.; Liu, Y.; Pedersen, T.; Melton, G.; Pakhomov, S. Umls::Similarity: Measuring the Relatedness and Similarity of Biomedical Concepts; Association for Computational Linguistics: Stroudsburg, PA, USA, 2013. [Google Scholar]
  38. Aronson, A.R. Effective mapping of biomedical text to the UMLS Metathesaurus: The MetaMap program. In Proceedings of the AMIA Symposium American Medical Informatics Association, Washington, DC, USA, 3–7 November 2001; p. 17. [Google Scholar]
  39. Torjmen-Khemakhem, M.; Gasmi, K. Document/query expansion based on selecting significant concepts for context based retrieval of medical images. J. Biomed. Inform. 2019, 95, 103210. [Google Scholar] [CrossRef] [PubMed]
  40. Resnik, P. Semantic similarity in a taxonomy: An information-based measure and its application to problems of ambiguity in natural language. J. Artif. Intell. Res. 1999, 11, 95–130. [Google Scholar] [CrossRef]
  41. Church, K.W.; Hanks, P. Word association norms, mutual information, and lexicography. Comput. Linguist. 1990, 16, 22–29. [Google Scholar]
  42. Severyn, A.; Moschitti, A. Learning to rank short text pairs with convolutional deep neural networks. In Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, Santiago, Chile, 9–13 August 2015; pp. 373–382. [Google Scholar]
  43. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  44. Zeiler, M.D.; Fergus, R. Stochastic pooling for regularization of deep convolutional neural networks. arXiv 2013, arXiv:1301.3557. [Google Scholar]
  45. Hersh, W.R.; Buckley, C.; Leone, T.J.; Hickam, D.H. OHSUMED: An interactive retrieval evaluation and new large test collection for research. In Proceedings of the 17th Annual International ACM-SIGIR Conference on Research and Development in Information Retrieval (SIGIR), Dublin, Ireland, 3–6 July 1994; pp. 192–201. [Google Scholar]
  46. Kalpathy-Cramer, J.; de Herrera, A.G.S.; Demner-Fushman, D.; Antani, S.; Bedrick, S.; Müuller, H. Evaluating performance of biomedical image retrieval systems—An overview of the medical image retrieval task at ImageCLEF 2004–2014. Comput. Med. Imaging Graph. 2014, 39, 55–61. [Google Scholar] [CrossRef]
  47. Grubinger, M.; Nowak, S.; Clough, P. Data sets created in ImageCLEF. In ImageCLEF: Experimental Evaluation in Visual Information Retrieval; Springer: Berlin/Heidelberg, Germany, 2010; pp. 19–43. [Google Scholar]
  48. Müller, H.; Kalpathy-Cramer, J.; Eggel, I.; Bedrick, S.; Radhouani, S.; Bakke, B.; Kahn, C.E., Jr.; Hersh, W. Overview of the CLEF 2009 medical image retrieval track. In Proceedings of the 10th Workshop of the Cross-Language Evaluation Forum (CLEF), Corfu, Greece, 30 September–2 October 2009; pp. 72–84. [Google Scholar]
  49. Müller, H.; Kalpathy-Cramer, J.; Eggel, I.; Bedrick, S.; Reisetter, J.; Kahn, C.E., Jr.; Hersh, W. Overview of the CLEF 2010 medical image retrieval track. In Proceedings of the Workshop of the Cross-Language Evaluation Forum (CLEF), Padova, Italy, 20–23 September 2010. Working Notes. [Google Scholar]
  50. Kalpathy-Cramer, J.; Müller, H.; Bedrick, S.; Eggel, I.; De Herrera, A.G.S.; Tsikrika, T. Overview of the CLEF 2011 medical image classification and retrieval tasks. In Proceedings of the Workshop of the Cross-Language Evaluation Forum (CLEF) 2011, Amsterdam, The Netherlands, 19–22 September 2011; Volume 1177. Working Notes. [Google Scholar]
  51. Müller, H.; Herrera, A.G.S.; Kalpathy-Cramer, J.; Demner Fushman, D.; Antani, S.; Eggel, I. Overview of the imageCLEF 2012 medical image retrieval and classification tasks. In Proceedings of the Workshop of the Cross-Language Evaluation Forum (CLEF) 2012, Rome, Italy, 17–20 September 2012. Working Notes. [Google Scholar]
  52. Herrera, A.G.; Kalpathy-Cramer, J.; Demner Fushman, D.; Antani, S.; Müller, H. Overview of the ImageCLEF 2013 medical tasks. In Proceedings of the Workshop of the Cross-Language Evaluation Forum (CLEF) 2013, Valencia, Spain, 23–26 September 2013; Volume 1179. Working Notes. [Google Scholar]
  53. Wu, H.; Sun, K.; Deng, X.; Zhang, Y.; Che, B. Uestc at imageCLEF 2012 medical tasks. In Proceedings of the Workshop of the Cross-Language Evaluation Forum (CLEF), Valencia, Spain, 23–26 September 2013. Online Working Notes/Labs/Workshop. [Google Scholar]
  54. Yu, G.; Li, X.; Bao, Y.; Wang, D. Evaluating document-to-document relevance based on document language model: Modeling, implementation and performance evaluation. In Proceedings of the International Conference on Intelligent Text Processing and Computational Linguistics, Mexico City, Mexico, 13–19 February 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 593–603. [Google Scholar]
  55. Lioma, C.; Ounis, I. A syntactically-based query reformulation technique for information retrieval. Inf. Process. Manag. 2008, 44, 143–162. [Google Scholar] [CrossRef]
  56. Wilcoxon, F. Individual comparisons by ranking methods. Biom. Bull. 1945, 1, 80–83. [Google Scholar] [CrossRef]
Figure 1. Overview of the Deep Matching Model.
Figure 1. Overview of the Deep Matching Model.
Diagnostics 14 01204 g001
Figure 2. Medical -Dependent Features.
Figure 2. Medical -Dependent Features.
Diagnostics 14 01204 g002
Figure 3. Query matrix extraction process.
Figure 3. Query matrix extraction process.
Diagnostics 14 01204 g003
Figure 4. The architecture of the personalized CNN model.
Figure 4. The architecture of the personalized CNN model.
Diagnostics 14 01204 g004
Figure 5. Some examples of ImageCLEF medical images and their extracted MDF.
Figure 5. Some examples of ImageCLEF medical images and their extracted MDF.
Diagnostics 14 01204 g005
Figure 6. Results according to α using 4 ImageCLEF datasets 2009, 2010, 2011 and 2012.
Figure 6. Results according to α using 4 ImageCLEF datasets 2009, 2010, 2011 and 2012.
Diagnostics 14 01204 g006
Table 1. Comparative results with the previous state-of-the-art approaches using ImageCLEF datasets.
Table 1. Comparative results with the previous state-of-the-art approaches using ImageCLEF datasets.
BM25DLMBo1PRFSemRank ( α = 0.3)
ImageClef-2009P@50.6080.5920.6080.696
P@100.5840.5240.5680.664
MAP0.3790.3270.3710.425
ImageClef-2010P@50.4000.4360.3610.453
P@100.4200.3750.3300.453
MAP0.3120.3130.3050.389
ImageClef-2011P@50.3930.2400.3860.406
P@100.3130.2230.3260.340
MAP0.1930.1380.2110.195
ImageClef-2012P@50.4180.2810.5540.427
P@100.3130.2410.4090.322
MAP0.1930.1460.3610.201
Table 2. Accuracy gain of the SemRank compared to other models.
Table 2. Accuracy gain of the SemRank compared to other models.
2009201020112012
SemRank/BM25+12% **+24%+1%+4%
SemRank/DLM+29% **+24%+40% **+38%
SemRank/Bo1PRF+14% **+27% **--
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gasmi, K.; Ayadi, H.; Torjmen, M. Enhancing Medical Image Retrieval with UMLS-Integrated CNN-Based Text Indexing. Diagnostics 2024, 14, 1204. https://doi.org/10.3390/diagnostics14111204

AMA Style

Gasmi K, Ayadi H, Torjmen M. Enhancing Medical Image Retrieval with UMLS-Integrated CNN-Based Text Indexing. Diagnostics. 2024; 14(11):1204. https://doi.org/10.3390/diagnostics14111204

Chicago/Turabian Style

Gasmi, Karim, Hajer Ayadi, and Mouna Torjmen. 2024. "Enhancing Medical Image Retrieval with UMLS-Integrated CNN-Based Text Indexing" Diagnostics 14, no. 11: 1204. https://doi.org/10.3390/diagnostics14111204

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop