Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
survey
Open access

A Systematic Collection of Medical Image Datasets for Deep Learning

Published: 27 November 2023 Publication History

Abstract

The astounding success made by artificial intelligence in healthcare and other fields proves that it can achieve human-like performance. However, success always comes with challenges. Deep learning algorithms are data dependent and require large datasets for training. Many junior researchers face a lack of data for a variety of reasons. Medical image acquisition, annotation, and analysis are costly, and their usage is constrained by ethical restrictions. They also require several other resources, such as professional equipment and expertise. That makes it difficult for novice and non-medical researchers to have access to medical data. Thus, as comprehensively as possible, this article provides a collection of medical image datasets with their associated challenges for deep learning research. We have collected the information of approximately 300 datasets and challenges mainly reported between 2007 and 2020 and categorized them into four categories: head and neck, chest and abdomen, pathology and blood, and others. The purpose of our work is to provide a list, as up-to-date and complete as possible, that can be used as a reference to easily find the datasets for medical image analysis and the information related to these datasets.

1 Introduction

Medical imaging technology has marked a new era in medicine since its inception [30]. The development of medical imaging began with the introduction of x-rays and has since evolved to include various other imaging techniques [72, 119], such as 3D Computed Tomography (CT) [1], Magnetic Resonance Imaging (MRI) [3, 6], nuclear medicine [2], ultrasound, endoscopy [4], and Optical Coherence Tomography (OCT) [169]. These medical image modalities, directly or indirectly, have contributed to the diagnosis and treatment of various diseases [232] and to the research on the human body’s structure and intrinsic mechanisms [217].
Medical images can provide critical insight into the diagnosis and treatment of many diseases. The human body’s different reactions to imaging modalities are used to produce scans of the body. Reflection and transmission are commonly used in medical imaging because the reflection or transmission ratio of different body tissues and substances are different. Some other methods acquire images by changing the energy transferred to the body, such as magnetic field changes or the rays radiated from a chemical agent.
Before modern Artificial Intelligence (AI) was applied in medical image analysis, radiologists and pathologists needed to manually look for the critical “biomarkers” in the patient’s scans. These biomarkers, such as tumors and nodules, are the basis for the medics to diagnose and devise treatment plans. Such a diagnostic process needs to be performed by medics with extensive medical knowledge and clinical experience. However, problems such as diagnostic bias and the lack of medical resources are prevalent and cannot be avoided. With the recent advancements in AI, which have achieved human-like performance in image recognition [33, 85, 164, 165, 166, 191, 198] and even won games like Go [190] and real-time strategy games [215], there has been a growing interest in developing AI-based automatic medical image analysis algorithms. As a result, AI has become a major research focus in medical image analysis and has achieved significant success in recent years [65, 98, 149, 163, 168, 172, 176].
Many researchers brought their focus to AI-based medical image analysis methods believing that it might be one of the solutions to the challenges (e.g., medical resource scarcity) and taking advantage of the technological progress [106, 137, 143, 189, 216]. Traditional medical image analysis focuses on detecting and classifying biomarkers for diagnosis and treatment. AI imitates the medic’s diagnosis through classification, segmentation, detection, regression, and other AI tasks in an automated or semi-automated way.
AI has achieved a significant performance for many computer vision tasks. This success is yet to be translated to the medical image analysis domain. Deep Learning (DL), a branch of AI, is a data-dependent method because it needs massive training data. However, when DL is applied to medical image analysis, the paucity of labeled data becomes a major challenge and a bottleneck.
Data scarcity is a common problem when applying DL methods to a specific domain, and this problem becomes more severe in the case of medical image analysis. Researchers who apply DL methods to medical image analysis research do not usually have a medical background, commonly computer scientists. They cannot collect data independently because of the lack of access to medical equipment and patients. They cannot also annotate the acquired data either because they lack the relevant medical knowledge. Furthermore, medical data is owned by institutions, which cannot easily make it public due to privacy and ethics restrictions.
To address some of these problems, MICCAI, ISBI, AAPM, and other conferences and institutions have launched many DL-related medical image analysis challenges. These aim to design and develop automatic or semi-automatic algorithms and promote medical image analysis research with computer-aided methods. Concurrently, some researchers and institutions also organize projects to collect medical datasets and publish them for research purposes.
Despite all these works, it is still challenging for novices of medical image analysis to find medical data. Therefore, we present this comprehensive survey of medical datasets and relevant challenges with the aim to help researchers easily find the required datasets for their research.
This article refers to other research papers with a number between square brackets and refers to the datasets listed in the tables with numbers between parentheses.
The article is organized as follows. Section 2 summarizes the datasets and challenges at a macro level and focuses on information relevant to medical data such as the years, body parts, and tasks. Sections 3 through 5 discuss the datasets and challenges of the head and neck, the chest and abdomen organs, and the pathology and blood. For other datasets, such as bone, skin, phantom, and animals, respectively, we place them into supplementary materials, because of the space limitation. We have also created a website with a git repo,1 which shows the list of these datasets and their respective challenges.

2 Medical Image Datasets

In this section, we provide an overview of the datasets and challenges. Our collection contains more than 300 medical image datasets and challenges organized between 2004 and 2020. This article focuses mainly on the ones between 2013 and 2020. Sections 2.1 through 2.4 provide information about the year, body parts, modalities, and tasks, respectively. In Section 2.5, we introduce the sources from where we have collected these datasets and the challenges. Details about the categorization of these datasets and challenges into four groups are provided in subsequent sections. We provide a taxonomy of our work in Figure 1 to help the reader navigate through the different sections.
Fig. 1.
Fig. 1. An overall taxonomy to outline the organization of the article.

2.1 Dataset Timeline

The timeline of these medical image datasets can be split into two, starting from 2013 as the watershed, since the excellent success of AlexNet [116] in the ILSVRC competition from 2012. The success of AlexNet motivated more researchers to solve a variety of problems [60, 66, 115, 132, 139] with different DL-based methods [84, 85, 91, 93, 174, 191]. In addition, it motivated many researchers to create and publish medical datasets. Consequently, many researchers focused on analyzing medical images with DL, which is one of the driving factors and reasons for the release of medical datasets. Another reason for us to mainly focus on datasets released after 2012 is that some of the datasets developed before 2012 are not aimed at computer-aided diagnosis with DL-based methods, such as Alzheimer’s Disease Neuroimaging Initiative (ADNI) 1 (55), although those data could be used for DL.
Figure 2(A) shows the statistics of the datasets and challenges per year between 2013 and 2020. As shown in Figure 2(A), the number of related datasets and challenges increased year by year because of the progress and success of DL in computer vision and medical image analysis. That led more and more researchers to focus on medical image analysis with DL-based methods, and more and more datasets and challenges with different body parts and tasks started to appear.
Fig. 2.
Fig. 2. Summary of medical image datasets and challenges from 2013 to 2020. Panel A shows the number of datasets and challenges published in each year. Panels B through D show the year-by-year trends along with the trends in relative numbers for each of the different categories by year. The numbers listed to the right are the summary count of each category, and the summary counts are not the same as the total numbers, because of (1) some of the categories are not shown and (2) a dataset counts two times if it includes two of the categories. Panels E through G show the most predominant body parts, data modalities, and main tasks with the percentage of their respective dataset.
Figure 2(B) shows the number of datasets related to the brain, heart, lung, liver, and eye. As can be noted, the proportion of the brain-related datasets is initially large. Then, the number of datasets related to these body parts (i.e., heart, lung, liver, and eye) is increased, but the proportion of the brain-related datasets starts to decrease.
Figure 2(C) shows that the number of datasets related to segmentation, classification, detection and other tasks is increased. The research mainly focused on the segmentation (28, 29, 54) task in the early stages. It then diversified to a variety of tasks or combinations of tasks that are also essential for clinical needs, including classification (154, 197, 259), segmentation (1, 2), detection (199, 232), regression (213, 214), generation (14, 19), tracking (220, 245), and registration (24, 25, 166), as time progressed (62).
Figure 2(D) shows no visible changes in the proportion of different modalities over the years, but the number of datasets related to different modalities increased.

2.2 Body Parts

As Figure 2(E) shows, the top-5 focused organs, in these datasets and challenges, include the brain, lung, heart, eye, and liver.
One of reasons datasets and challenges focus on these organs is that the tasks of datasets, such as a classification and segmentation, are helpful in the diagnosis and treatment of cancer (a major threat to human life) or other diseases. For example, eye-related diseases, which cause blindness, incited the collection of eye-related datasets and the release of challenges. Some other datasets and challenges focus on the small organs, such as the prostate, which are challenging to analyze due to the low resolution of images.
As AI has been shown to be more accurate than humans to tackle some complex tasks [79, 86, 188], researchers have attempted to solve the imaging problem of an organ with the AI methods successfully tested on other other organs. That has led the researchers to focus on more and more organs.

2.3 Modalities

There are several types of medical image modalities. As shown in Figure 2(F), the frequently used modalities to acquire medical datasets include MRI, CT, ultrasound, endoscopy, Positron Emission Tomography (PET), Computed Radiography (CR), electrocardiography, and OCT. Limited by space, we introduce the main modalities in the supplementary materials.
Magnetic Resonance (MR), CT, and other modalities are the most commonly used imaging modalities. MR can provide sharp images without harmful radiations of soft tissues. It is therefore widely used in the imaging of brain, heart, and many other small organs. CT is an economical and simple imaging approach, and it is widely used for the diagnosis of cancer (e. g., of the the neck, chest, and abdomen). A pathology image is different from MR and CT, because it is a cell-level imaging method. Pathology is widely used in cancer-related diagnosis.

2.4 Tasks

As Figure 2(G) shows, we categorize these tasks into seven categories: classification, segmentation, detection, regression, generation, registration, and tracking. Limited by space, we introduce these tasks in the supplementary materials.

2.5 Source and Destination

To conduct a systematic search for relevant datasets and challenges, we first reviewed papers that addressed medical image segmentation and classification. This enabled us to identify several sources for datasets, including Grand Challenge, a website that hosts medical imaging related competitions, and The Cancer Imaging Archive(TCIA) [56], a website for cancer-related competitions. Additionally, we utilized Google to search for other relevant websites and collections that featured datasets and challenges related to medical image analysis with DL. Through this process, we discovered Kaggle, Codalab, OpenNeuro, PhysioNet [77], and scientific journals such as Scientific Data, which were inspired by Grand Challenge and TCIA.
Next, we crawled through these websites to gather information about the datasets/competitions featured on these sites, carefully selecting datasets and challenges based on the available information.
The article lists datasets and challenges primarily sourced from Grand Challenge and Codalab, which are predominantly related to medical imaging. These were directly included in our work. However, the datasets and challenges from TCIA, Kaggle, PhysioNet, OpenNeuro, and journals were not all related to medical imaging or DL. Therefore, we filtered these datasets and challenges using specific keywords such as “machine learning,” “deep learning,” “segmentation,” and “classification.” Some datasets were removed because they were not suitable for DL and AI methods, due to the lack of annotations or other factors. It is worth noting that not all medical image analysis tasks require annotations. For example, the generation task, such as generating synthetic CT images from MR ones [44], does not require annotations. Furthermore, for senior researchers, data in TCIA, OpenNeuro, UK Biobank, and other databases can be useful after manual annotations.
We then categorized the remaining datasets and challenges into different groups. Systematically categorizing the datasets and challenges is challenging. Thus, we used an asymmetric categorization to group these datasets and challenges into four groups, as shown in Figure 1. This means that we did not use the same subtaxonomy in each category or subcategory.
First, we simply split the medical datasets and challenges into two groups: body level (Sections 3 and 4) and cell level (Section 5), according to the imaged body part. Second, we grouped the datasets and challenges of the brain, eye, and neck into one group (Section 3), because these are parts of the head. Third, we organized the datasets and challenges related to the chest and abdomen into the same group (Section 4). Finally, for the datasets and challenges that could not be categorized into the preceding groups, we grouped them under “other” and placed them in the supplementary materials due to the limitation of space. These datasets and challenges are related to the skin, bone, phantom, and animals.
The introduction of each group and subgroup includes mainly the type of modality, the task, the disease, and the body part. However, not all groups of datasets can be introduced in that way. For some groups, we introduce the datasets and challenges according to the domain-specific problems. For example, we categorize the pathology datasets into microscopic and macroscopic tasks.

2.6 Notes to the Tables

We have compiled tables in this article to present the datasets and challenges. These tables contain basic information such as the index, dataset/challenge name, release year, and other relevant details. Some information is not available in the table due to either non-provision or unavailability during our data collection process.
For all tables, the information includes the following:
Reference indexThe index to refer to the dataset or challenge in our work, and it is numbered in the order of appearing in the table.
Dataset/ChallengeThe dataset or challenge’s name or abbreviation of the name. The name in the table is the hyperlink to the website or home page of the dataset or challenge.
YearThe year that the dataset or challenge was released. For datasets, it is the year the paper went public or their website was noted. For challenges, it is the year that their “name” was declared, but for some challenges, of which the conferences will be hosted after 2020, we note 2020.
ModalitiesThe modalities of data. For abbreviations, we note the terms in the footnote of the table.
FocusSome words or a sentence to introduce the focus of datasets and challenges.
TasksThe tasks of the dataset or challenge designed or its focus. For abbreviations, we note the terms in the footnote of the table.
Lesion/TumorThe lesion or tumors that the dataset or challenge focused on. For abbreviations, we note the terms in the footnote of the table.
DiseasesThe diseases that the dataset or challenge focused on. For abbreviations, we note the terms in the footnote of the table.
OrgansThe organs that the dataset or challenge focused on. For abbreviations, we note the terms in the footnote of the table.
CategoryThe category of classification or segmentation.
StainThe stain that is used in pathological images.
ResolutionThe resolution of data in the pixel or the voxel. For some datasets and challenges, we note the size of it in millimeters.
ScaleThe number of data in the dataset or challenge. We might note the division of the dataset or challenge if the authors of it mentioned such information.
Multi-centerWhether the dataset or challenge includes multi-center data. We will not use the checkmark (\(\checkmark\)) and crossmark (\(\times\)) if any related information is left in the description of the dataset or challenge.
RankWhether the challenge includes a rank to compare the participants’ algorithms.
Conf.The name of the conference related to the challenge.
LicenseThe license of the dataset. We note the abbreviation in the footnote of the tables.

3 Head and Neck Related Datasets and Challenges

The head and neck are ones of the most significant parts of the human body because many essential organs, glands, and tissues are located there. Several researchers’ image analysis work relates to the head and neck. To make effective use of computers for research, diagnosis, and treatment, many researchers have released datasets and challenges, for example: (i) the analysis of tissue structure and functions (13, 5) and (ii) disease diagnosis (30, 40, 44).
Because the brain controls human emotions and actions/functions of other organs, the datasets and challenges significantly focus on the brain. First, we introduce the datasets and challenges related to the analysis of brain structure, function, imaging, and other basic tasks in Section 3.1. Second, we introduce the datasets and challenges related to brain disease diagnosis in Section 3.2.
Moreover, since the eyes are crucial to our vision, the computer-aided diagnosis of eye-related diseases is also an important research focus. The eye-related datasets and challenges are covered in Section 3.3. Finally, we introduce the remaining datasets and challenges related to the neck and the research of brain behavior and cognition in Section 3.4.

3.1 Structural Analysis Tasks of the Brain

The basic analysis and processing of the brain medical images are clinically critical for diagnosis, treatment, and other brain-related analysis tasks. The datasets and challenges we discuss are mainly for the segmentation tasks and center around the brain structure. In contrast, some datasets focus on imaging, including MRI acceleration (e.g., (14, 16)), the non-linear registration of different modalities (e.g., (24, 25)), and tractography (e.g., (28, 29)).
Table 1 lists the datasets and challenges related to basic brain image analysis, such as the segmentation of White Matter (WM) and Gray Matter (GM) and registration of images. One of the most popular tasks in Table 1 is the segmentation of WM, GM, and Cerebrospinal Fluid (CSF), and their respective datasets and challenges are introduced in Section 3.1.1. Meanwhile, other tissues and functional areas’ segmentation are also the focus of research, and their related datasets and challenges are discussed in Section 3.1.2. Section 3.1.3 describes the other basic tasks. Table 1 presents an overview of the datasets and challenges related to medical image analysis of the head and neck. The majority of these datasets provide MR images, whereas some offer CT, PET, or ultrasound data. Approximately half of the datasets and challenges in Table 1 include multi-center data, and some do not specify whether they contain multi-center data. Most of the challenges listed in Table 1 are affiliated with conferences such as MICCAI, ISBI, and MIDL, whereas others do not have an affiliation. All of the challenges shown in Table 1 provide a ranking of participants’ methods. Some of the challenges require registration to access the data and submit methods, whereas others are public under license.
Table 1.
Table 1. Summary of Datasets and Challenges for the Basic Brain Image Analysis

3.1.1 Segmentation of WM and GM.

The segmentation of WM, GM, CSF, and other parts of brain tissues is always a focus for researchers working on brain-related analysis and diagnosis [96, 150]. Similarly, for AI algorithms, it is also of significance to understand the human brain’s structure. Therefore, MICCAI and others have held many challenges (1–8) with this research focus, and researchers could design automatic algorithms to segment MR images into different parts. We introduce these datasets and challenges related to their modalities and tasks in the following.
Modality. Since these datasets and challenges focus on the segmentation of brain tissues, the data usually provided is MR images. Challenges (1–6) provide mainly two modalities: T1 (T1-weighted MRI, based on longitudinal relaxation time [8], providing structural information about tissues and commonly used for anatomical imaging), T2 (T2-weighted MRI, based on transversal relaxation time [8], highlighting fluid and helping to identify edema or swelling), whereas datasets (7, 8) only provide T1 for the WM hyperintensities segmentation task. Note that MR scans are sensitive to the hydrogen atom, and such a feature can effectively help image analysts distinguish between different tissues and parts of the image. Moreover, due to the color of the tissue imaged by MR, these scans are named as “white matter” and “gray matter.”
Task. The main focus of these datasets and challenges is the segmentation of WM, GM, and CSF. However, they do not only focus on that. Challenges (3–7) also provide the annotation of other parts of the brain, including basal ganglia, WM lesions, cerebellum, and infarction. Challenges (1, 2, 4, 6) use MR images of the neonatal brain and consider tissue volumes as an indicator of long-term neurodevelopmental performance [96]. Challenges (1, 3) increased regions to segment and provided different data compared with the previous challenges (2, 5). Challenge (8) focuses on the segmentation of WM hyperintensities in the brain.
Performance Metric. For the segmentation task, the Dice score is one of the most commonly used metrics, and all of these datasets and challenges adopt it as a performance measure. Besides the Dice score, datasets (3, 5, 8) also use Hausdorff distance and volumetric similarity as metrics; datasets (1, 2) use the average the Hausdorff distance and the average surface distance as one of their metrics; and moreover, dataset (8) also uses sensitivity and F1-score as metrics for performance evaluation.

3.1.2 Segmentation of Functional Areas and Other Tissues.

The segmentation of functional areas and tissues also has an essential meaning for brain-related research and computer-aided diagnosis. In this subsection, we introduce the datasets and challenges related to the segmentation of functional areas and tissues.
Tissues Segmentation. WM, GM, and CSF were introduced in Section 3.1.1; however, the segmentation of other brain tissues is also an active research area. Challenges (3, 5–7) aim to segment brain images into different tissues, including ventricles, cerebellum, brainstem, and basal ganglia. These challenges provide MR images and the voxel-level annotations of the regions of interest with 30 or 40 scans. Because these regions are essential for brain health, researchers need to overcome the challenges related to their size and shape to segment them. Dataset (9) is designed for the segmentation of the cerebellum from Diffusion-Weighted Imaging (DWI) (a.k.a. dMRI), which is a type of MRI that utilizes the movement of water molecules to generate images [7] and measures the random movement of water molecules within tissues and is particularly sensitive to acute strokes. However, dataset (10) focuses on the segmentation of the caudate from brain MR images.
Functional Areas. The segmentation of the human brain cortex into different functional areas is of great significance in education, clinical research, treatment, and other applications. Datasets (11, 12) provide images and annotations for the design of automatic algorithms to segment the brain cortex into different functional areas—in other words, brain atlas. Brain atlas is a map to annotate different functional areas or tissues. Dataset (11) uses the DTK protocol [111], which is modified from the DK protocol [62], and the DTK protocol includes 31 labels, details of which are listed at https://mindboggle.readthedocs.io/en/latest/labels.html. Dataset (12) is a commercial dataset for research in the segmentation of functional areas of the brain cortex.
Dataset (13) focuses on the segmentation of the hippocampus, which can be used to detect the progress of Alzheimer’s Disease (AD), according to challenge (13). It provides MR images from AD patients as the training set and the test set.

3.1.3 Imaging-Related Tasks.

In addition to the segmentation tasks of the brain tissues and the functional areas, some of the datasets and challenges also focus on the generation, registration, and tractography.
Generation. Datasets and challenges (14–16) aim to accelerate MRI or generate high-resolution MR images from low-resolution ones. Usually, high-resolution imaging requires higher cost, whereas low-resolution imaging is cheaper but affects analytical judgment and may lead to an incorrect diagnosis. These challenges provide many scans at low resolution to allow researchers to design algorithms to convert or map low-resolution images onto higher-resolution ones. The datasets and challenges mainly focus on the generation tasks. Another focus is cranioplasty (17–19) to generate a part of broken skull from CT images of the models of the broken skull. Additionally, challenges (20, 21) focus on another type of generation task, the reconstruction of diffusion MR images. Compared with challenge (21), challenge (20) focused on the effect of the local reconstruction accuracy on the quality of connectivity reconstruction. Challenge (22) focuses on the reconstruction of WM to answer the questions such as “What model (either biophysical or signal model) better explains the underlying tissue environment?” and “What dMRI sequence enables the most accurate estimation of these parameters?” Dataset (23) focuses on face reconstruction from PET and CT images, which aims to augmented reality in surgery or other tasks.
Registration. Registration is a research area that focuses on aligning images from different modalities. Challenges (24, 25) are centered on the registration of ultrasound data and MR images of the brain. Cross-modality registration can be challenging because the subject is not completely stationary, and ultrasound is a 2D imaging modality while MR is a 3D volume imaging modality. Therefore, these challenges aim to establish the topological relationship between the pre-operative MR image and intra-operative ultrasound. Challenge (24) provides registration tasks for pre- and post-resection, whereas challenge (25) does not. Challenge (26) focuses on the registration of diffusion MR images to eliminate variations caused by different vendors’ hardware devices and protocols.
Tractography. Tractography is another segmentation task and focuses on the segmentation and imaging of the fiber in the WM. Dataset (27) aims to segment the fiber bundles from brain images, including phantom, squirrel monkey, and macaque, whereas challenges (28, 29) focus on the tractography with Diffusion Tensor Imaging (DTI), which is a kind of DWI) that provides additional information about the direction of water movement within tissues, allowing for the assessment of WM tracts in the brain [7]. The details of the challenges’ tasks are different. Challenge (28) focused on the segmentation of the corticospinal tract from a set of neurosurgical cases, and challenge (29) focused on the segmentation and the analysis based on the segmentation.

3.2 Brain Disease Related Datasets and Challenges

Besides the structural analysis and image processing tasks, computer-aided diagnosis is also a research focus in healthcare. Medical image analysis plays a critical role in clinical research, diagnosis, and treatment. The datasets and challenges we have included are for two tasks: (i) the segmentation of lesions and (ii) the classification of diseases. For the segmentation task, the respective datasets and challenges focus on the lesion segmentation of the human brain and mark the lesion’s contour for diagnosis and treatment, and the relevant details are shown in Section 3.2.1. For classification tasks, the datasets and challenges have been used for the development of automatic algorithms to classify or predict diseases from medical images, and these datasets and challenges are presented in Section 3.2.2.

3.2.1 Datasets for Segmentation of Lesions.

Lesions in the brain affect a human’s healthy life and safety, and image analysis is an effective way to diagnose relevant diseases. In this subsection, related datasets and challenges are introduced, and they are reported in Table 2.
Table 2.
Table 2. Summary of Datasets and Challenges for the Brain Lesion Segmentation Tasks
The tasks of these datasets mainly include the segmentation of brain tumors (e. g., glioma), the segmentation of cerebral aneurysm, the segmentation of stroke lesion, the segmentation of Intra-Cranial Hemorrhage (ICH), and sclerosis lesion. Most of the datasets and challenges use MR images that include different submodalities, whereas some are using CT. Nearly half of these datasets and challenges listed in Table 2 are reported including multi-center data, whereas a few of them are reported as not included. The rest of them do not report whether multi-center data are used. The challenges in Table 2 provide rankings of the participants’ methods, and most of them are associated with MICCAI. Most of the challenges in Table 2 are required to register when accessing data, whereas other datasets and challenges are opened under specific licenses.
Glioma Datasets and Challenges. Gliomas are one of the most common and deadly brain malignancies [12]. Therefore, many challenges and datasets focus on the segmentation of glioma for its diagnosis and treatment. The BraTS challenge series (30–38) has been going on since 2012 to segment the glioma. The challenges of such a segmentation task are caused by the heterogeneous appearance and shape of gliomas. The heterogeneity of the glioma reflects its shape, modalities, and many different histological subregions, such as the peritumoral edema; the necrotic core; and the enhancing and the non-enhancing tumor core. Therefore, these series of challenges provide multi-modal MR scans to help researchers design and train algorithms to segment tumors and their subregions. The tasks of this challenge series include (i) low- and high-grade glioma segmentation (37, 38), (ii) survival prediction from pre-operative images (32, 33), and (iii) the quantification of segmentation uncertainty (30, 31). Besides the BraTS challenge series, dataset (40) is another one for the segmentation of low-grade glioma and provides T1-weight and T2-weight MR images with biopsy-proven gene status of each subject by FISH (fluorescence in-situ hybridization) [181]. Dataset (39) focuses on the processing of brain tumor and aims to design and evaluate DL-based automatic algorithms for glioblastoma segmentation and further research.
Cerebral Aneurysm Challenges. Cerebral aneurysm is life threatening because the rupture of the cerebral aneurysm is associated with high mortality and cognitive impairment in case of survival. Therefore, challenges (41–43) focus on the detection of the cerebral aneurysm for the strategy to prevent rupture. These three challenges respectively focus on detection, segmentation, and rupture risk assessment of the cerebral aneurysm.
Ischemic Stroke Lesion Datasets and Challenges. Similar to tumor segmentation, brain lesion segmentation also focuses on detecting brain abnormalities. However, the difference is that lesion segmentation deals with damaged tissues. Challenges (44–48) focus on stroke lesion segmentation because stroke is also life threatening and can disable the surviving patients. Stroke is often associated with high socioeconomic costs and disabilities. Automatic analysis algorithms help diagnose and treat stroke, since its manifestation is triggered by local thrombosis, hemodynamic factors, or embolic causes. In MR images, the infarct core can be identified with diffusion MR images, whereas the penumbra (which can be treated) can be characterized by perfusion MR images. The challenge ISLES 2015 (47) focuses on subacute ischemic stroke lesion segmentation and acute stroke outcome/penumbra estimation and provides 50 and 60 multi-modality MR scans of data for training and validation, respectively, for two subtasks (i. e., subacute ischemic stroke lesion segmentation and acute stroke outcome/penumbra estimation). The subsequent year’s challenge, ISLES 2016 (46), focuses on the segmentation of lesions and the prediction of the degree of disability. This challenge provides about 70 scans, including clinical parameters and MR modalities, such as DWI, ADC, and perfusion maps. The challenge ISLES 2017 (45) focuses on the segmentation with acute MR images, and ISLES 2018 (44), focuses on the segmentation task based on acute CT perfusion data. Moreover, dataset (48) i.e., ATLAS) focuses on the segmentation of the stroke lesion and the brain itself after stroke for further treatments.
ICH-Related Datasets. ICH is a sudden bleeding into the brain tissues or ventricles or both of them, which can be caused by traumatic brain injuries or other reasons [92]. ICH could lead to disability or even death if it is not treated on time. Normal diagnosis and treatment is based on the analysis of brain CT to localize the regions of ICH. Thus, dataset and challenge (49, 50) focus on the detection and segmentation of ICH to help medics locate the hemorrhage regions and decide on a treatment plan. Dataset (51) also provides data for the classification of normal or hemorrhage CT images.
Multiple Sclerosis Lesion Related Datasets. The multiple sclerosis lesion is another kind of lesion in the brain that is not life threatening but can cause disabilities. Datasets and challenges (52–54) are related to multiple sclerosis lesion segmentation with multi-modality MR data, such as T1w, T2w, and FLAIR (fluid-attenuated inversion recovery, a kind of MR image, “removing” signal from the CSF [5]).

3.2.2 Classification of Brain Disease.

Except for the lesion segmentation, brain disease classification also plays an essential role in healthcare. Brain-related diseases have a severe effect on patients’ health and their lives (e. g., AD [25, 67, 105, 199] and Parkinson’s disease (PD)). Therefore, effective diagnosis and early intervention can effectively reduce the health damage to patients, the effect on the social times of families, and the economical impact on society. In this section, we first introduce the datasets and challenges of AD (55,–59, 64), then introduce other diseases (65–68). Table 3 shows the relevant challenges and datasets.
Table 3.
Table 3. Summary of Datasets and Challenges for Brain Disease Classification Tasks
The datasets listed in Table 3 mainly use MRI and PT. For AD, PET is commonly used with specific contrast agents to find landmarks associated with AD. For other diseases, MRI or specific modalities are used to better detect diseases. AD has a more detailed division of cognitive impairment, whereas datasets of other diseases are binary classifications. The numbers of datasets and challenges that included or did not include multi-center data are nearly 50-50, whereas there is a dataset that did not report whether it was used or not. The challenges listed in Table 3 are required to register when accessing, whereas other datasets are opened or public under specific licenses.
Alzheimer’s Disease. AD affects a person’s behavior, cognition, memory, and daily life activities. Such progressive neurodegenerative disorder affects the normal daily life of patients because suffering from such a disease makes patients not know who they are and what they should do, which then progresses to the point until they forget everything they know. The disease takes an unbearable toll on the patient and leads to a high cost to their loved ones and to society. For example, according to Sudha et al. [9], AD became the sixth deadly cause in the United States in 2018 and costs around $200 to $300 billion.
Therefore, researchers are doing everything they can to explore the causes of AD and its treatments. Diagnosis based on medical images has become a research focus because early diagnosis and intervention have significance on the progress of this disease. Hence, many researchers work on the classification (i.e., prediction of AD using brain images). The datasets mainly include ADNI and Open Access Series of Imaging Studies (OASIS).
ADNI is a series of projects that aim to develop clinical, imaging, genetic, and biochemical biomarkers for the early detection and tracking of AD. It includes four stages: ADNI-1 (55), ADNI-GO (56), ADNI-2 (57), and ADNI-3 (58). These projects provide image data of the brain for researchers, and the modalities of images include MR (T1 and T2) and PET (FDG, PIB, Florbetapir, and AV-1451). These four stages consists of 1,400 subjects. The subjects can be categorized into normal cognition, Mild Cognitive Impairment (MCI), and AD, where MCI can be split into early mild cognitive impairment and later mild cognitive impairment.
OASIS is a series of projects aiming to provide neuroimaging data of the brain, which researchers can freely access. OASIS released three datasets: OASIS-1, OASIS-2, and OASIS-3. All of these three datasets are related to AD, but they are also used in functional areas segmentation and other tasks. OASIS-1 (59) contains 418 subjects aged from 18 to 96 years, and for subjects older than 60 years, there are 100 subjects diagnosed with AD. The dataset includes 434 MR sessions. OASIS-2 (60) contains 150 subjects, aged between 60 and 96 years, and each subject includes three or four MR sessions (T1). About 72 subjects were diagnosed as normal, whereas 51 subjects were diagnosed with AD. In addition, there are 14 subjects who were diagnosed as normal but were characterized as AD at a later visit. OASIS-3 (61) includes more than 1,000 subjects, more than 2,000 MR sessions (T1w, T2w, FLAIR, etc.), and more than 1,500 PET sessions (PIB, AV45, and FDG). The dataset includes 609 normal subjects and 489 AD subjects.
Moreover, there are many other challenges based on ADNI and OASIS or independence datasets. Challenge TADPOLE (62) is based on ADNI and aims at the prediction of the longitudinal evolution. Dataset (63) is based on OASIS and is released on Kaggle for the classification of AD. Challenge CADDementia (64) is an independent AD-related challenge to classify subjects into normal cognition, MCI, and AD.
Other Diseases. Similar to AD, other brain diseases are also important from the diagnosis and treatment perspective. However, the number of datasets and challenges of these diseases is not as large as AD. A few datasets focus on PD and Spinocerebellar Ataxia Type 2 (SCA2). Datasets ANT (65) and (66) provide images of PD with MR images and classification labels. Dataset (67) provides images and classification labels of SCA2. Dataset MTOP (68) provides images and annotations for the diagnosis of mild traumatic brain injury.

3.3 Eye-Related Datasets and Challenges

As the human’s imaging sensor, the eyes’ health is essential for human beings, and eye diseases may lead to blindness. We introduce the relevant challenges and datasets in this subsection and list them in Table 4. Many of the studies presented the size of their dataset, with approximately half of them separating it into training and testing sets. Table 4 includes a list of challenges that offered rankings for participants’ methods, with many of them being associated with MICCAI and ISBI. It is worth noting that access to these challenges often requires registration, whereas other datasets may be accessible with specific licenses.
Table 4.
Table 4. Summary of Datasets and Challenges of Eye Disease Related Tasks
Datasets according to the Modality. With regard to eye-related datasets and challenges, the main used modalities are fundus photography (69–72, 74–79, 84, 86) and OCT (73, 80, 81, 83, 85). The data provided by challenge (82) consists of surgery videos. Fundus photography can help medics evaluate health of the eye and locate retinal lesions because fundus photography clearly shows the important parts of the eye, such as the blood vessels and the optic disc. OCT is a new imaging approach that is safe for the eye and shows the retinal tissues in detail. However, it has also disadvantages—it is not suitable for diagnosing microangioma and the planning of retinal laser for photocoagulation treatment.
Datasets according to the Analysis Task. These datasets and challenges can be used for four tasks:
(1)
Classification tasks focus on classifying whether the subject has specific diseases or judging whether the subject is abnormal. Datasets and challenges (70–73, 76–78, 84) focus on predicting a single disease, whereas others (69, 75, 79–81) focus on diagnosing multiple diseases. Challenge (71) is an upgrade of challenge (72) that provided more data for participants.
(2)
Segmentation is another task, which provides more information compared to classification. Datasets and challenges focus on the segmentation of the tissues and lesions, such as disc (71, 72, 76, 79), Cup (71, 72), retinal lesion (76, 79), blood vessels (74), and retinal (83) and DME lesions (85), for further diagnosis and disease analysis.
(3)
Datasets and challenges (72, 73, 76, 78, 79, 86) focus on the detection of lesions or other landmarks. These tasks help medics locate key targets, such as areas and tissues, for effective diagnosis or provide feature details for other automated algorithms.
(4)
Unlike other tasks, the last one focuses on the annotation of the tools used for eye-related surgery (82).
Datasets according to Focused Eye Diseases. Researchers mainly focus on the following diseases:
Glaucoma (72, 75)
Closure glaucoma (73)
(Age-related) macular degeneration (75, 78, 80)
Cataract (75)
Diabetes retinopathy (75, 77, 79, 80, 81, 84, 86)
Hypertension (75)
Pathologic myopia (75, 76)
Diabetic macular edema (79, 85)
Besides these diseases, dataset (82) aims at the annotation of images and videos in surgery.

3.4 Datasets and Challenges of Other Subjects

Besides the brain’s structural analysis, the image processing, and the computer-aided diagnosis tasks, another important research focus is the human neck because it holds many essential glands and organs. This subsection discusses the datasets and challenges of the neck and teeth, covered in Sections 3.4.1 and 3.4.2, respectively. Moreover, many researchers are working on the analysis of behavior and cognition with DL-based methods. We discuss the details in Section 3.4.3. Table 5 includes datasets and challenges that cover various modalities, including CT, PET, MR, and ultrasound, with a focus on segmentation. Most of the challenges listed in Table 5 have rankings for submitted data and require registration for data access, whereas others have specific licenses. Some of these challenges are affiliated with MICCAI, whereas others are associated with AAPM, and one challenge is jointly hosted by ISBI and MICCAI.
Table 5.
Table 5. Summary of Datasets and Challenges of Head and Neck Related Diseases

3.4.1 Neck-Related Datasets.

The neck is essential for our health, as it holds many glands and organs. When these become abnormal, effective diagnosis and segmentation play an essential role in their treatments. The related image datasets and challenges are listed in Table 5.
Datasets and challenges (87–92) focus on the segmentation of glands and the lesions in relevant glands. Dataset (93) focuses on the binary classification tumor vs. normal. Challenge (94) aims at the task of thyroid gland nodules detection with ultrasound images and videos. Challenge (95) focuses on the nerves segmentation in the neck, whereas challenge (96) focuses on evaluating carotid bifurcation.

3.4.2 Cephalometric and Teeth Related Datasets.

Challenges (97, 98) focus on the diagnosis of dental x-ray images. The main tasks of these two challenges include landmark localization and caries segmentation. Challenge (97) provides around 400 cephalometric x-ray images with the annotation of landmarks by two experienced dentists. Challenge (98) provides about 120 bitewing images with experts’ annotations of different parts of the teeth.

3.4.3 Behavior and Cognition Datasets.

To understand what we see, hear, smell, and feel, our brain draws on neurons in the brain to compute and analyze the stimulations and understand what, where, why, and when questions and scenarios. Many researchers now use artificial neural networks as a research method to analyze the relationship between brain activities and stimulation. They use functional MR images to scan our brain activity, analyze the hemodynamic feedback, and identify the area of the neurons that react. Therefore, the analysis of the reactions of the brain in response to a specific stimulation is an important research focus. Researchers use DL to detect or decode the stimulation of subjects to work out the brain’s functionality. The related datasets are listed in Table 6. These datasets are all opened with specific licenses, and all of them are single-center data, which means that the publisher collected the data themselves. Due to the difficulty of collecting data, the scale of these datasets is not too large (less than 50).
Table 6.
Table 6. Summary of Datasets and Challenges Used for Behavioral and Perception Related Tasks
Some datasets (99–101) focus on classifying the stimulations or the subject’s attribution based on the subject’s functional MR images. Dataset (99) aims to identify whether the subject is a beginner or an expert in programming via the reaction of their brain to source codes. Dataset (100) focuses on diagnosing subjects with depression vs. subjects with no depression using audio stimulations and analyzing the subjects’ brain activity. Dataset (101) focuses on the influence of cannabis on the brain.
Datasets (102–109) focus on the encoding of the stimulations (i.e., brain activities’ decoding). Datasets (105, 107, 108) aim to rebuild what subjects have seen using DL-based methods from their brain activities using functional MR images. In addition, datasets (103, 109) work on the encoding of faces that subjects have seen from functional MR images with similar modalities.

3.5 Summary

Within this section, we have provided an overview of datasets and obstacles concerning the head and neck region. To present this information, we created six tables detailing each entry within the prior subsections. Table 1 specifically pertained to datasets and challenges centered around brain segmentation, registration, and generation. Table 2 listed datasets and challenges that mainly focused on the segment of brain images, and additionally, there were many other datasets and challenges focused on the tasks, such as generation and registration. The datasets and challenges, listed in Table 3, mainly focused on the classification of brain disease. Table 4 listed the datasets and challenges related to the classification, segmentation, detection, and other tasks associated with the diagnosis and surgery of eye disease. Table 5 included datasets and challenges about diseases of the head and neck, such as the segmentation of the glands or lesions in glands and cephalometric analysis. The datasets listed in Table 6 were related to behavior and perception.

4 Chest and Abdomen Related Datasets and Challenges

There are many vital organs in the chest and abdomen. For example, the heart is responsible for the blood supply, the lungs are responsible for breathing, and the kidneys are responsible for the production of urine to eliminate toxins from the body. Therefore, medical image analysis of organs in the chest and abdomen is an important research focus. Most of the tasks are computer-aided diagnosis with classification, detection, and segmentation of lesions being the most targeted tasks.
Many datasets and challenges aim to segment one or more organs in the chest and abdomen for diagnosis or treatment planning. Section 4.1 discusses the datasets and challenges related to inter- or intra-organ segmentation. Section 4.2 introduces the datasets and challenges that focus on the classification, detection, and segmentation tasks related to the diagnosis of diseases of the chest and abdominal organs. Section 4.3 describes the datasets and challenges of the chest and abdomen that are not categorized in other sections, including regression, tracking, registration, and other tasks related to the chest and abdominal organs.

4.1 Datasets for Chest and Abdominal Organ Segmentation

This subsection covers the datasets and challenges of the chest and abdominal organs that are used for anatomic segmentation tasks. The anatomic segmentation tasks include inter-organ segmentation (Section 4.1.1) and intra-organ segmentation (Section 4.1.2). Inter-organ segmentation aims to separate an organ from the background or mark the boundaries between multiple organs and the background. Intra-organ segmentation aims to segment the organ into different parts at the anatomical level.
Table 7 showcases the datasets and challenges that are dedicated to the segmentation of organs in the chest and abdominal regions. The majority of these datasets and challenges use CT and MR modalities. Information such as the number of images and division of training/testing sets were reported for most of these datasets and challenges. The challenges listed in Table 7 generally provide rankings for submitted methods, and many of them are affiliated with MICCAI, whereas a few are associated with ISBI or AAPM. Access to most of these challenges requires registration, whereas datasets and the remaining challenges are accessible under specific licenses.
Table 7.
Table 7. Summary of Datasets and Challenges for Chest and Abdominal Organ Segmentation Tasks

4.1.1 Datasets of the Chest and Abdominal Organs.

Inter-organ segmentation is a necessary information for the pre-planning of surgery and diagnosis. A well-segmented contour of the organs provides a precise mask, which helps produce accurate segmentation results for the diagnosis, treatment, and operation. This subsection introduces datasets and challenges for inter-organ segmentation of a single organ and of multiple organs.
Chest and Abdomen Datasets according to the Organ. Inter-organ segmentation is one of the most attractive topics of medical image analysis [97, 103]. It is a basic pre-processing step in diagnosis and treatment, as it provides contour of the organs. Some datasets and challenges focus on the larger organs, such as the liver and lungs, whereas others focus on multi-organ segmentation or smaller organs. This is because of challenges, such as imbalanced labels and blurred contour in low resolution, brought by multi-organs and smaller organs. The datasets and challenges that we have covered involve the following organs and parts:
Lung (110–117)
Liver (110, 111, 115, 116, 118–121)
Kidney (110, 115, 116, 118, 120, 122)
Prostate (111, 123–127)
Esophagus (113, 114, 120, 128)
Heart (111, 113, 128, 129)
Pancreas (111, 115, 120, 130)
Aorta (115, 120, 128)
Spleen (115, 118, 120)
Adrenal glands (115, 120)
Bladder (110, 115)
Gallbladder (115, 120)
Spinal cord (113, 114)
Trachea (115, 128)
Colon (111)
Breast (131)
Lymph (132)
Clavicles (133)
Stomach (120)
Chest and Abdomen Datasets according to Modality. The most commonly used image modalities for the chest and abdominal organs segmentation are MR and CT. As Table 1 shows, many datasets and challenges use MR images. MR images have higher resolution under certain conditions and have better resolution for soft body tissues and organs, such as the heart and prostate. Meanwhile, CT is the most widely used modality for organ segmentation and other tasks and diagnosis related to the chest and abdomen, such as the lung and liver, according to our research, because of its convenience, effectiveness, and low cost.
Chest and Abdomen Datasets according to Focus. The purpose of these datasets and challenges can be categorized into three groups: further analysis, benchmark, and radiotherapy. Most datasets and challenges that provide annotated organs’ contours are provided with the objective to focus on further analysis and treatments. One of the challenges of segmentation is to achieve a robust segmentation of the whole organ and separate it from the background, without omitting the lesions and tumors, and thus some test benchmarks (111, 115) are provided for researchers to evaluate their algorithms. Another challenge is the imbalance between different organs because of their sizes and shapes, and such an imbalance makes it challenging to segment small organs and provide valuable information for analysis and treatment. Dataset and challenges (113, 128) provide data for researchers to address the problem.
Single Chest and Abdominal Organ Segmentation. Single-organ contour segmentation tasks usually focus on segmentation with an anatomical purpose (112, 117, 123–125, 127) or the research of a region for subsequent tasks (119, 121, 122, 126, 130–132). The difficulty of the former task is that the lesions and tumors may affect the segmentation by separating the organ from the background, whereas the latter’s difficulty is for researchers and their algorithms to perform more precise segmentation. Challenge (133) focused on the segmentation of the clavicles.
Chest and Abdomen Multi-Organ Segmentation. Chest and abdomen multi-organ segmentation focuses distinguishing inter-organs. Some of these datasets and challenges (110, 111, 118) focus on the segmentation of multiple organs, including the relatively larger organs, which are easier to segment, and the relatively smaller organs, which can be more challenging to segment compared to the larger ones, especially when the model is handling larger and smaller organs at the same time. Similarly, some of these datasets and challenges (113, 128) focus on the “organ at risk,” which means that these organs are healthy but might be at risk because of radiation therapy. Dataset (115) aims to provide a benchmark for the segmentation algorithms. Dataset (120) focuses on multi-atlas-based methods, which are widely used in brain-related research.

4.1.2 Chest and Abdomen Intra-Organ Segmentation.

Different from inter-organ segmentation of the chest and abdominal organs, intra-organ segmentation aims to segment the organ into different parts. Just as the hand has five fingers, organs are made up of multiple parts, and a typical example is the Couinaud liver segmentation method [58]. This subsection introduces the datasets and challenges for organ segmentation. These datasets and challenges are listed in Table 7.
Heart-Related Datasets and Challenges. Most of these datasets and challenges (137–143) are related to heart segmentation. The most frequently used modalities are MR and ultrasound, and the aim is to segment the heart into the left atrium, chambers, valves, and other parts. Although MR and ultrasound can effectively image the different tissues of the heart, the heartbeat results in blurred images, which makes the segmentation task more difficult, whereas for ultrasound, the dynamic nature of its images is another challenge for the segmentation algorithm.
Other Chest and Abdomen Body Parts. Challenge (144) provides 55 CT scans and focuses on the segmentation of the lung with the labeling of its different parts: outside the lungs, the left lung, the upper lobe of the left lung, the lower lobe of the left lung, the upper lobe of the right lung, the middle lobe of the right lung, and the lower lobe of the right lung. The biggest challenge is the effect of the lung lesions and diseases, such as tuberculosis and pulmonary emphysema, on the performance of segmentation. Moreover, challenges (145, 146) focus on segmentation of the lung vessels.

4.2 Datasets for Diagnosis of Chest and Abdominal Diseases

Diseases of organs in the chest and abdomen have a significant impact on human health. Therefore, many researchers work on this problem by analyzing medical images. Several researchers have designed automatic or semi-automatic algorithms for classification, segmentation, detection, and characterization tasks to help medics diagnose these diseases. In this subsection, we describe the datasets and challenges related to diagnosis of diseases of the chest and abdomen, which are reported in Tables 8 through 10, respectively.
Table 8.
Table 8. Summary of Datasets and Challenges for Chest and Abdominal Organ Related Tasks I
Table 9.
Table 9. Summary of Datasets and Challenges for Chest and Abdominal Organ Related Tasks II
Table 10.
Table 10. Summary of Datasets and Challenges for Chest and Abdominal Organ Related Tasks III
The majority of datasets and challenges listed in Tables 8 through 10 provided information on the scale of the data, including whether they included multi-center data and whether they had rankings. Many of the challenges were associated with various conferences, such as MICCAI, ISBI, RSNA, AAPM, and ICIAR. Most of these challenges required registration to access the data, whereas other datasets and challenges were available under specific licenses. Additionally, most of them reported the division of the training/testing set.
Chest and Abdomen Datasets according to Modality. According to the datasets and challenges collected, CT is the most commonly used imaging modality for the chest and abdomen because of its suitable imaging quality and ability to clearly display tissues and lesions. Some datasets and challenges also provide CT images using contrast agents for clearer images. Besides CT imaging, there are other modalities, including MR, x-ray digital radiographs, PET, and endoscopy, among others. The MR images are used in breast-related diagnosis, cardiac-related tasks, soft tissue sarcoma detection, and ventilation imaging. Because of the organs’ size and the CT’s resolution, which is limited by the imaging exposure time and radiation dose, MR is a more suitable imaging modality for small or specific organs. PET is always used with other modalities, such as CT and MRI. The contrast agent’s density is related to the metabolism, which means that the density of radiation from the contrast agent will be high in the tumor, so PET is always used for tumor-related tasks. Endoscopy images are used for medical inspection of the stomach, intestines, and others.
Chest and Abdomen Datasets according to Classification of Diseases. The classification of diseases intends to determine whether a subject is healthy or not. A quick and early diagnosis can allow effective interventions to increase the probability of the patient recovering before the condition worsens.
The main focus of these datasets is to judge whether there is any cancer, lesion, or tumor, such as soft tissue sarcoma (147), prostate lesion (148, 149), lung cancer (150–153), and breast cancer (154).
Another focus is the classification of diseases. These diseases include mainly pneumothorax (155), cardiac diseases (156), tuberculosis (157), pneumonia (158), and COVID-19, which are discussed at the end of this subsection.
The endoscopy-related challenges (159–161) provide data with the aim to classify RGB images and videos into “normal” vs. “abnormal” categories. The abnormal category in (160) includes VA (mucosa showing villous atrophy), CH (mucosa showing crypt hypertophy), and VACH (mucosa showing both villous atrophy and crypt hypertrophy); the abnormal category in (161) includes GMP (gastric metaplasia), BAR (intestinal metaplasia), and NPL (neoplastic mucosa).
Dataset (162) focuses on classification based on diagnostic records. These datasets and challenges provide data for researchers to design AI-based algorithms to diagnose common diseases.
Chest and Abdomen Datasets for Attribute Classification. The characterization task of the tumor and lesion is also called attribute classification, which focuses on the subsequent characterization analysis of tumors and lesions following detection and segmentation tasks using automatic analysis algorithms. A typical example is the attributes classification of pulmonary and lung cancer (164–168). The datasets and challenges usually provide CT scans with the annotation of different attributes, such as lesion type, spiculation, lesion localization, margin, lobulation, calcification, and cavity. Each attribute includes two or more categories. Another focus is the characterization of breast-related lesions and tumors (169, 170).
Chest and Abdomen Datasets for Detection. In most research and clinical situations, classification is not enough. The medics and researchers usually focus on lesions related to such a disease and their localization. Further treatment evaluations and a plan and interpretability are the specific focus for medics and DL researchers. Thus, detection and segmentation are the tasks receiving a lot of attention today. The detection task aims to find a region of interest and localize its position. The regions of interest usually include the following:
Lung cancer and tumor (167, 171–174)
Pulmonary nodule (164, 168, 174–176)
Celiac-related damage (177–180)
Action and artifact of surgeon (180, 181)
Other lung lesions (182, 183)
Polyp (179, 184)
Breast cancer (169, 185, 186)
Cervical cancer (187)
Liver cancer (188)
Chest and Abdomen Datasets for Segmentation. Segmentation is a refinement of the detection task because it provides information about the location and the pixel-level labels. Pixel-level annotations can help researchers design pixel-level algorithms for accurate and effective quantification, volume calculations, and other analysis and diagnosis of tumors and lesions at the pixel level (e. g., monitoring of tumor size). According to the datasets and challenges we have collected, most of them aim at the segmentation of the tumor and lesion from CT of the following:
Heart-associated conditions (189)
Kidney tumor (122)
Liver tumor (119)
Lung cancer (190–192)
Pneumothorax (155)
Polyp (179)
Pulmonary nodule (193)
Additionally, challenge (194) centers around the segmentation and ventilation-related tasks of the lung. Meanwhile, challenges (178, 180) concentrate on the segmentation of artifacts (e.g., polyps) in endoscopic images. The discrepancies between the two challenges are the specific tasks and corresponding data. Challenge (195) encompasses a range of intricate tasks associated with medical images and offers various modalities.
COVID-19. In 2020, COVID-19 became a research focus because it caused more than 100 million infections and 2 million deaths. Different datasets and challenges focus on this devastating disease and provide data to help researchers develop DL models to detect COVID-19 via various medical image modalities.
With respect to modalities, most of these datasets and challenges use either CT or CR images, and some provide both modalities. One exception is dataset (196), which uses ultrasound images. These datasets provide image annotations labeled by radiologists.
Most of these datasets and challenges are related to classification tasks. Datasets (197–201) directly focus on diagnosing COVID-19 from normal subjects. In contrast, datasets and challenges (196, 202–205) focus on diagnosing COVID-19 from a few other similar diseases, which can also lead to lung opacity or other symptoms, such as Middle East respiratory syndrome, severe acute respiratory syndrome, and acute respiratory distress syndrome. Moreover, other datasets and challenges (206, 207) focus on the diagnosis task, with natural language processing or genomics.
Similarly, some other datasets (199, 208, 209) focus on the segmentation or detection of COVID-19-related lesions, such as ground-glass opacity, air-containing space, and pleural effusion.
Additionally, dataset (210) offers images and clinical data to support medical professionals and researchers in their diagnostic efforts and algorithm development.

4.3 Datasets for Other Chest and Abdomen Related Tasks

Besides the classification, detection, and segmentation tasks, there are several other tasks that are the current focus of research. In the following, we present the datasets and challenges related to these tasks and report them in Table 11.
Table 11.
Table 11. Summary of Datasets and Challenges of Other Medical Applications in Chest and Abdomen
The datasets and challenges presented in Table 11 primarily consist of MR, CT, and ultrasound data, with some providing information on the size of the dataset. Challenges in this table include rankings of participant methods, with many associated with MICCAI. Although most challenges require registration to access data, some datasets and challenges are openly available.
Chest and Abdomen Datasets for Regression. Similar to attributes classification, regression is another task that aims to compute or measure the target attributes from given images, but the difference is that the outputs of regression are continuous. A typical example is fetal biometric measurements (211, 212). These challenges provide ultrasound images to help researchers design algorithms to measure such attributes to estimate the gestational age and monitor the fetus’s growth. Another example is cardiac measurements (213–219). Of these, challenge (214) provided data more clinical data compared with its previous challenge (215). These datasets and challenges provide MR or ultrasound images to analyze the heart’s attributes to detect heart diseases.
Chest and Abdomen Datasets for Tracking. Tracking is a crucial task because organs and the body tend to move during imaging. The motion characteristics of organs such as the heart provide valuable information. Challenges (220, 221) focus on tracking the liver using ultrasound data to analyze surgery and treatments. The tasks of challenges (220, 221) are similar, but challenge (220) provided more data compared to the previous year. Datasets and challenges (222–224) focus on the tracking of the heart. They provide ultrasound images to track and analyze the heart.
Chest and Abdomen Datasets for Registration. Challenge (225) focuses on the CT registration of lungs and provides CT scans with and without enhancement and contrast agents. Meanwhile, challenges (226, 227) focus on the registration between different modalities of the heart and provide MR, CT, and other modalities to register images with beating hearts. Challenge (228) focuses on the registration of lungs.
Datasets for Other Chest and Abdomen Related Tasks. Challenge (229) focuses on the detection of amniotic fluid. Challenges (230, 232) focus on localizing specific landmarks, including the amniotic fluid and the heart, using ultrasound and MR images. Challenge (233) focuses on the classification of surgery videos. Dataset (234) focuses on reconstruction of the coronary artery.

4.4 Summary

In this section, we provided a summary of the datasets and challenges related to the chest and abdomen. We organized them into five tables and described them in detail. Table 7 included datasets and challenges that concentrate on the segmentation of organs in the chest and abdomen, with some focused on single- or multiple-organ segmentation, whereas others focused on intra-organ segmentation. Tables 8, 9, and 10 presented datasets and challenges associated with diagnostic tasks such as disease classification, tumor segmentation, and lesion detection. Table 11 listed other datasets and challenges that focus on tasks like regression, tracking, registration, and generation.

5 Datasets and Challenges for Pathology and Blood

Although radiography, MRI, and other imaging modalities have been used as the basis for diagnosis, pathology images are also used as a gold standard for diagnosis, particularly for tumors and lesions. Digital pathology images are generally obtained by collecting tissue samples, making slices, staining, and imaging. Therefore, pathology images are also one of the mainstream image modalities used for diagnosis.
The focus of these datasets and challenges include (i) the classification and segmentation of basic elements (e. g., cell and nucleus) in pathology images and (ii) blood-based diagnosis from images. In this section, we present datasets and challenges of pathology images (Section 5.1), and we cover the datasets and challenges of blood images in Section 5.2.

5.1 Datasets and Challenges for Pathology

Pathology images are used as one of the bases for the cancer diagnosis [29]. Pathologists and automatic algorithms analyze images based on specific features, such as cancer cells and cells under mitosis. Many organizations and researchers provide datasets and challenges that focus on microscopic pathology and at the Whole Slide Image (WSI) level. The relevant datasets and challenges are listed in Table 12.
Table 12.
Table 12. Summary of Datasets and Challenges for Pathology-Related Image Analysis
Table 12 contains datasets and challenges that primarily use H&E as the staining method. Most of the listed datasets and challenges report the size of the data and whether it includes multi-center data. The division of the training/testing set is also reported for most of them. Multi-center datasets and challenges are more prevalent than single-modality ones. The challenges provide rankings for submitted methods, and they are mostly associated with MICCAI, ISBI, or other conferences. Some of the challenges require registration to access the data, whereas others are open with specific licenses.
Imaging Datasets and Challenges. In most situations, WSI is used in pathology diagnosis. Unlike CT or MR images, the pathology image is an optical image similar to a picture taken by a camera. However, one major difference is that a pathology image is imaged by transillumination, whereas the usual photo is imaged by reflection. Another difficulty is in the size of the image. WSI is stored in a multi-resolution pyramid structure. A single multi-resolution WSI is generally achieved by capturing many small high-resolution image patches, and it might contain up to billions of pixels. Thus, WSI is used as a virtual microscope in diagnosis for clinical research, and many challenges use WSI, such as (235–240).
However, for some machine learning based algorithms, it is hard to directly process and analyze the whole WSI—for example, cell segmentation (241, 242). In some situations, the original image is too large to be analyzed directly by algorithms because of limited computing resources. In other situations, “end-to-end” DL-based algorithms are hard to use for both low- and high-resolution-level features without zoom-in and zoom-out action and intelligent agents. Therefore, pathology image patches are used in several other challenges, such as (243) for visual question answering, (244) for mitosis classification, and (241, 242) for multi-organ nucleus detection and segmentation.
Stains. Slides made from human tissues are without color and require staining. The commonly used stains include hematoxylin, eosin, and diaminobenzidine. Usually, two or more stains are used in staining the slide, and the most commonly used combinations include Hematoxylin & Eosin (H&E) and hematoxylin & diaminobenzidine (H-DAB), as listed in Table 12.
Pathology Datasets according to Disease. Pathology slides are widely used in the diagnosis of many diseases, especially cancer. Cancer cells and tissues have different shapes compared to their normal counterpart. Thus, the diagnosis via pathology is the gold standard. Many datasets and challenges, such as (241, 242, 245), do not address any specific disease but do address cell-level tasks. In addition, many datasets and challenges target specific diseases, such as breast cancer (246, 247), cancers of the digestive system (248), cervical cancer (249, 250), lung cancer (251), thyroid cancer (252), and osteosarcoma (253).
Pathology Datasets according to Task. Generally speaking, the tasks used with these datasets and challenges can be classified into two categories: microscopic tasks and WSI-level tasks. The latter targets the diagnosis of diseases based on a classification task. Expanded from the simple classification tasks, many datasets and research methodologies focus on complex tasks, such as the segmentation of tumor cell areas (241, 242, 254) and the detection of pathological features (239, 248). Microscopic tasks derive from the clinical analysis to identify cells and detect mitosis to extract key features from pathology images to support further disease diagnosis. The following subsections expand on microscopic tasks and WSI-level tasks, respectively.

5.1.1 Microscopic-Related Datasets.

In this subsection, we introduce microscopic task related datasets and challenges. These tasks focus on microscopic features extraction (e.g., nucleus features) for further diagnosis and WSI-level tasks.
Data. Unlike the WSI level, the datasets and challenges that focus on microscopic tasks usually provide small-size patch-level images with high resolution. These patches are suitable for the annotation of microscopic-level objects and resource-limited algorithms. The size of images varies depending on the image analysis tasks. For segmentation and detection of the cells and nucleus, the size of images is usually a 1,000-pixel square to contain the suitable number of cells or nuclei. For individual cell analysis tasks (e.g., mitosis determination), the size is usually of a single cell. For other tasks (e.g., patch-level classification), the size varies from dataset to dataset.
Pathology Datasets for Cell Detection and Segmentation. The cell is one of the essential parts of the pathology image. The analysis of cells is one of the most effective ways to extract pathology image features for diagnosis. Pathologists analyze the size, shape, pattern, and stained color of the cells with their knowledge and expertise to make judgments about these cells and classify them as normal or abnormal. Thus, many datasets and challenges focus on segmentation and detection of cells. The cells and nucleus can be placed neatly on the slide. However, during slide preparation, these cells could overlap or locate randomly on the slide. Aiming at such a problem, challenges (249, 250) focus on segmentation and detection of overlapping cells and nuclei. The shape and size of cells from different organs might be different and can have different recognition and analysis challenges. Therefore, challenges (241, 242) focus on multi-organ cells or nucleus segmentation.
Pathology Datasets for Patch-Level Classification. Generally, the size of WSI is too large to be able to analyze every cell and relationships between cells. DL-based methods can find essential information from the patch-level image to support the diagnosis based on feature learning. The datasets and challenges, which provide patch-level images, mainly focus on classification, segmentation, or detection tasks. Challenges (252, 255–258) focus on patch-level image classification to determine whether metastatic or different tissue is present.
Datasets for Other Pathology Tasks. Besides the detection and segmentation of cells and patch-level classification, there are other microscopic tasks. Challenge (244) focuses on mitotic detection for nuclear atypia scoring. The atypical shape, size, and internal organization of cells are related to the progression of cancer. The more advanced the cancer is, the more atypical the cell appear. Challenge (245) focuses on cell tracking, to know how cells change shapes and move as they interact with their surrounding environment. This is the key to understanding the cell migration’s mechanobiology and its multiple implications in normal tissue development and many respective diseases. Challenge (243) focuses on the visual question answering task of pathology images using AI where the model is trained to pass the examination of the pathologist. Challenge (263) focuses on the tracking of particles, which mimic moving viruses, vesicles, receptors, and microtubule tips.

5.1.2 Datasets for WSI-Level Tasks.

WSI-level pathology tasks focus on the diagnosis of cancer and pathology image processing. WSI contains all the complete information of a patient to be able to establish an accurate diagnosis. Automatic diagnosis algorithms can quickly analyze the slide. This is useful especially in developing countries where there is a lack of well-experienced pathologists. However, it is a challenge to directly analyze WSI for both pathologists and algorithms because the size of WSI can be up to \(100,\!000 \times 100,\!000\) pixels. Thus, such analysis becomes challenging, and to address this, most of the current datasets and challenges focus on classification and segmentation of biomarkers, cells, and other regions of interest. At the end of this subsection, we introduce other datasets and challenges that are related to the tasks of regression and localization of tumors and biomarkers.
Datasets for Classification of WSI. The prime goal of the examination of pathological images, especially WSI, is to diagnose cancer. Thus, how to classify WSI with large size and limited computing resources becomes a research challenge. Datasets and challenges (235, 236, 259, 260) focus on predicting cancer or evaluating WSI, such as Gleason grade or HER2 evaluation. In addition, some datasets and challenges (238, 239, 247) focus on the classification of metastasized cancer. The main difference between (239) and (238) is the scale of the dataset, and the latter provided more data for training and testing.
Datasets for Segmentation and Detection of WSI. DL-based methods are seen as a black box that processes pathology images. These methods have achieved state-of-the-art performance, but the interpretability of them remains difficult. From the pathologists’ point of view, datasets and challenges (237–239, 248, 261) focus on segmentation and detection tasks to determine the critical elements leading to a particular diagnosis, such as the cancer cell area and signet ring cell.
Datasets for Other WSI Tasks. Besides classification and detection, there are a few other tasks based on WSI. These includes the registration of pathology images (262) for data pre-processing and the localization of lymphocytes (255).

5.2 Blood-Related Datasets

Blood image analysis is the basis of the diagnosis of many diseases. In contrast to pathology images, blood samples’ images mainly contain blood cells, and blood-related datasets and challenges are aimed at blood-related cancer and cell counting. Similar to pathology images, these datasets and challenges also focus on the segmentation, detection, and classification of cells. The relevant datasets and challenges are listed in Table 13, with most of them reporting the size of data available. Only two of them mention the division of the training/testing set. About half of the challenges listed in Table 13 are linked with ISBI, and most of them require registration to access data. Other datasets and challenges are open.
Table 13.
Table 13. Summary of Datasets and Challenges of Blood-Related Image Analysis Tasks
One of the primary tasks of these datasets is the classification of cells, which focuses on identifying the different types of cells. Dataset (264) focuses on classifying red blood cells, white blood cells, platelets, and other cells. Dataset (265) focuses on the classification of malignant and non-malignant cells. Challenges (266, 267) focuses on the classification of malignant cells from normal cells in B-ALL white blood cancer microscopic images. Other datasets and challenges (268) (lymphocytes), (269, 270) (multiple myeloma segmentation), (271) (mitochondria segmentation), and (272) (malaria detection) focus on segmentation and detection of blood cells and biomarkers. Dataset (273) focuses on stain normalization for white blood cancer (B-ALL & MM) microscopic images.

5.3 Summary

In this section, we provided an overview of the datasets and challenges related to pathology and blood. We organized them into two tables to better illustrate and describe them. Table 12 listed the datasets and challenges related to pathology, including tasks such as WSI-level classification and segmentation related to cancer, cell detection or segmentation, and patch-level tasks. Table 13 listed datasets and challenges related to blood-related microscopic images. Most of the datasets and challenges listed in Table 13 reported the scale of data, and some of them provided information on the division of the training/testing set. Half of the challenges were associated with ISBI, and most of them required registration when accessing data, whereas other datasets and challenges were open.

6 Discussion

The success of AI algorithms such as DL has led to their widespread use in several fields, including medical image analysis. Researchers with different knowledge and background levels tackle image-based clinical tasks using computer vision tools to design automatic algorithms for different applications [59, 108, 137, 137, 143, 228, 230, 231]. Although AI algorithms can successfully handle many tasks, several unsolved problems and challenges hinder the development of AI-based medical image analysis.

6.1 Problems and Challenges

DL-based algorithms learn from input images of real data through gradient descent. Large-scale annotated datasets and a powerful DL model are key to the development of successful DL models. For example, the success of AlexNet [116], GoogleNet [198], and ResNet [85] is based on powerful models, which include millions of parameters. Additionally, a large-scale dataset, such as ImageNet [60], is also necessary to train the DL model to be able to tune such a large number of parameters. However, when these methods are applied to medical image analysis, many domain-specific problems and challenges start to appear. This subsection discusses some of these challenges.

6.1.1 Data Scarcity.

The biggest challenge in the development of DL models is data scarcity. Different from other areas, the scale of the medical image datasets is usually smaller due to many limitations (e.g., ethical restrictions).
The commonly used datasets for traditional computer vision are in larger scale compared to medical image datasets. For example, the handwritten digits dataset, MNIST [122], includes a training set with 60,000 examples and a testing set with 10,000 examples; the ImageNet dataset [60] includes 3 million images for training and testing; and Microsoft COCO [132] includes more than 2 million images with annotations. In contrast, many medical image datasets are smaller and only include hundreds or at most thousands of images. For example, the challenge BraTS 2020 (30) includes 400 subjects and different modalities for each subject, the challenge REFUGE (72) provides about 1,200 images of the eye, the challenge LUNA 16 (166) provides 888 CT scans, our recently published dataset of pulmonary lesions [127] provides only 694 scans, and the challenge CAMELYON 17 (238) contains more than 1,000 WSI pathology images.
There are multiple reasons for the lack of data. The main cause is due to restricted access to medical images by non-medical researchers (i.e., barriers between disciplines). The root causes of these barriers relate to the cost and difficulties of annotation and restricted access due to ethics and privacy.
Access to Data. As mentioned in Section 1, the direct cause of data scarcity is that most non-medical researchers are not allowed to access medical data directly. Although many medical data are generated worldwide every day, most non-medical researchers have no authorization to access clinical data. The easily accessible data include publicly available datasets, but these datasets are not at a large enough scale to be able to properly train a DL model.
Ethical Reasons. Ethics of medical data usage is a major bottleneck and a limitation for researchers, particularly computer scientists, because they do not have resources to collect data. It is also hard for them to access medical data. Medical data stored in databases always contains sensitive or private information, such as name, age, gender, and ID number. In some cases, the data records of medical images can be used to identify a patient. For example, if an MR scan includes the face, an intruder can identify them for a possibly evil purpose. In most countries and regions, it is illegal to distribute such data with private information without the patients’ permission, and nobody would usually consent to such distribution. Therefore, for DL researchers, it is impossible to gain authorization to access these datasets.
Before DL researchers are able to gain authorization even to desensitized data, they still need to pass ethical reviews.
Annotation. Another root cause is the difficulty to annotate medical images. Unlike other computer vision areas, the annotation of medical images requires specialized professionals and knowledge. For example, in autopilot, when annotating objects such as vehicles and pedestrians, there are no specific annotators’ requirements because most of us can easily distinguish a car or a human. However, when annotating medical images, domain-specific knowledge is essential. For example, few people without professional knowledge would be able to tell the difference between abnormal and normal tissue. However, it is impossible for a non-specialist to mark the lesion’s contour or diagnose a disease.
This difficulty cannot easily be solved even when professionals are employed to annotate data. First, the cost of annotation of medical data is huge. Once the researcher and their organization have obtained some data, they then need to spend more money to employ a few medics for its labeling. Such annotation cost is enormous, particularly when medical resources are scarce or where medical costs are high. For example, the challenge PALM (76) provides about 1,200 images with annotation, but its organizers involved only two clinical medics. Second, the physician who annotates the data is required to have a rich clinical and diagnostic experience, thus reducing the number of people who are suitable for this task even further. Third, to avoid any subjectivity, one image needs to be annotated by two or more physicians. Another problem is this: what should be done if the labels of two annotators are not the same? In many challenges, the organizer employs many junior physicians to annotate and employs a senior physician to decide if the junior physicians’ annotations are not the same. For example, in the challenge AGE (73), each data annotation is determined by the mean of four independent ophthalmologists in a group and it is then manually verified by a senior glaucoma expert.

6.1.2 Challenges of Medical Data.

The characteristics of medical images themselves pose difficulties for medical image analysis tasks.
There are many types and modalities of images used in medical image analysis. Similar to computer vision, the modalities include both 2D and 3D. However, the medical images have several other differences. Although the average scale of a medical image dataset is smaller than computer vision related field datasets, the size of each sample of data is larger on average than the one of a computer vision related field.
For 2D images, CR, WSI, and other modalities have large variances in resolution and color than the other computer vision data. Some modalities might need more bits to encode a pixel, whereas other modalities are significantly huge. For example, CAMELYON 17 (238) only includes about 1,000 pathology images, but the whole dataset is about 3 TB. Such datasets with few large samples pose a challenge for the AI algorithms, and it requires the researchers to focus on designing algorithms that can learn from limited computational resources and be useful for clinical diagnosis.
For 3D medical images such as CT and MRI, they are dense 3D data compared with sparse data, such as point cloud, in autopilot. Like the BraTS serial of challenges (30–38), many researchers face challenges in designing algorithms that can effectively learn from a multi-modal dataset.
These characteristics of medical images require well-designed algorithms with a more robust capability to fit the data well and without overfitting. However, that further leads to the need for more data and resources. It is a challenge to learn suitable features from a small sample dataset.

6.2 No Silver Bullet

The ideal scenario is to find or invent a method or an algorithm to simultaneously solve all of these encountered problems. However, there is no silver bullet. The problems and challenges related to the data and the adopted methods cannot be entirely resolved, or sometimes a problem arises as another is solved. Nevertheless, many ideas have been introduced to address the current problems, and they are introduced in this subsection.
With respect to the problems and challenges mentioned previously, researchers are working on two research directions: (i) a more effective model with less data and (ii) a more practical approach to access data. For the learning methods with small datasets, researchers use approaches such as few-shot learning and transfer learning. To access more data, researchers adopt three main approaches, namely federated learning, lifelong learning, and active learning.

6.2.1 Practical Learning from Small Samples.

Many medical image datasets have a small number of samples. For example, challenge MRBrains13 (5) only includes 20 subjects for training and testing, whereas challenge KITS 19 (122) has about 200 subjects. Therefore, many researchers struggle to find a practical approach to learn from small samples.
Few-Shot Learning and Zero-Shot Learning. Few-shot learning hits one of the critical spots of DL-based medical image analysis problems—that is, the development of DL models with fewer data [222]. Humans can effectively learn from a few samples. Therefore, different from the standard DL-based methods, humans learn to diagnose a disease from images without the need to view tens of thousands of images (i.e., from only few-shot). Meta-learning, also called learning to learn, is a technique used to address the few-shot learning problem [90]. Meta-learning can learn the meta-features from a small data size. The number of medical images in most datasets and challenges is not as large compared to the regular computer vision related datasets and challenges. Mondal et al. [155] use few-shot learning and GAN to segment medical images. The GAN is modified for semi-supervised learning with few-shot learning. Similar to few-shot learning, zero-shot learning aims at novel samples. Rezaei and Shahidi [175] cover a review of zero-shot learning from autonomous vehicles to COVID-19 diagnosis. However, zero-shot and few-shot learning also have their disadvantages, such as domain gap, overfitting, and interpretability.
Knowledge Transfer. Transfer learning is another method that can recognize and apply knowledge and skills learned from a previous task [236]. For example, both WM and GM segmentation and multi-organ segmentation are segmentation tasks [102]. However, the general neural network training is usually independent, which means that almost nobody trains a neural network with two tasks at once. However, it does not mean that these two tasks are unrelated. Besides zero-shot learning and few-shot learning, transfer learning, or say, knowledge transfer, is another method to infer knowledge from a previously learned task. Transfer learning can be applied to two similar tasks and between different domains [107]. The most significant advantage of transfer learning is that it uses rich scale datasets to pre-train the neural network and then fine-tune and transfer the network to the main task on a dataset with few samples.

6.2.2 Effective Access to More Samples.

Besides finding a practical approach to learn from small samples, many researchers have been working on active learning and federated learning (which aims to use data without access to sensitive information). This also reduces annotation costs of DL algorithms.
Federated Learning. Federated learning provides another way to access data. As discussed previously, the limitation of accessing data is led by privacy and other problems. Instead of directly sharing data, federated learning shares the model to protect privacy from being leaked. With other privacy protection methods, federated learning can effectively use the data from each independent data center or medical center.
However, there are two disadvantages of federated learning: annotation and implementation. The problem of annotation cannot be solved by sharing data but other methods [52, 155, 186]. The main challenges are the design and implementation of a method that allow only feature sharing (needed by algorithms) and not the private information, as only a few institutions have attempted federated learning so far. For example, Intel and other institutions have attempted to apply federated learning for brain tumor related tasks in their research [187]. The main challenges in their implementation include the following:
(1)
The implementation and proof of privacy protection,
(2)
The methodology for sharing and updating millions of the model’s parameters,
(3)
Preventing attacks on DL algorithms and leaks of data privacy on the Internet or computing nodes.
Natural Language Processing. Natural language processing is also a potential tool to automatically or semi-automatically annotate medical image data. It is a standard procedure for a medic to provide a diagnostic report of the patient, particularly after the medical image was taken. Therefore, such large amounts of data (image and text) are useful for medical image analysis after desensitization, and natural language processing can be used for annotation. Several natural language processing based methods (e.g., [109, 130, 214]) have been applied in medical-related research fields.
Active Learning. Active learning aims to reduce the annotation cost by indirectly using the unlabeled data to select the “best” samples to annotate. Generally, data annotation for DL requires experts to label data so that the neural network can learn from the data. Active learning does not require too many samples at the beginning of training. In other words, active learning can “help” annotators label their data. Active learning uses the knowledge learned from the labeled data to select and annotate the unlabeled data. The unlabeled data with annotation from algorithms is used to subsequently train the network over the next number of epochs. Active learning [52, 186] is used in medical image analysis in a loop of (i) the algorithm learning from the data annotated by humans, (ii) the human annotating the unlabeled data selected by the algorithm, and (ii) the algorithm adding the newly labeled data to the training set. The advantage of active learning is obvious: annotators do not need to annotate all the data they have, and additionally, the neural network can learn from data faster from such interactive progress.

7 Conclusion

In this work, we provided a comprehensive survey of the datasets and challenges for medical image analysis, collected between 2013 and 2020. The datasets and challenges were categorized into four themes: head and neck, chest and abdomen, pathology and blood, and others. We provided a summary of all the details about these themes and data. We also discussed the problems and challenges of medical image analysis and the possible solutions to these problems and challenges.

Footnote

1
The website with the git repo will be public after the article is accepted.

References

[1]
William R. Hendee and E. Russell Ritenour. 2002. Computed tomography. In Medical Imaging Physics (4th ed.). John Wiley & Sons, 251–264.
[2]
William R. Hendee and E. Russell Ritenour. 2002. Instrumentation for nuclear imaging. In Medical Imaging Physics (4th ed.). John Wiley & Sons, 197–216.
[3]
William R. Hendee and E. Russell Ritenour. 2002. Magnetic resonance imaging and spectroscopy. In Medical Imaging Physics (4th ed.). John Wiley & Sons, 367–388.
[4]
William R. Hendee and E. Russell Ritenour. 2002. Ultrasound waves. In Medical Imaging Physics (4th ed.). John Wiley & Sons, 303–316.
[5]
Robert W. Brown, Yu-Chung N. Cheng, E. Mark Haacke, Michael R. Thompson, and Ramesh Venkatesan (Eds.). 2014. Introductory signal acquisition methods. In Magnetic Resonance Imaging: Physical Principals and Sequence Design (2nd ed.). John Wiley & Sons, 113–139.
[6]
Robert W. Brown, Yu-Chung N. Cheng, E. Mark Haacke, Michael R. Thompson, and Ramesh Venkatesan (Eds.). 2014. Magnetic resonance imaging. In Magnetic Resonance Imaging: Physical Principals and Sequence Design (2nd ed.). John Wiley & Sons, 1–17. arXiv:
[7]
Robert W. Brown, Yu-Chung N. Cheng, E. Mark Haacke, Michael R. Thompson, and Ramesh Venkatesan (Eds.). 2014. Random walks, relaxation, and diffusion. In Magnetic Resonance Imaging: Physical Principals and Sequence Design (2nd ed.). John Wiley & Sons, 619–636.
[8]
Robert W. Brown, Yu-Chung N. Cheng, E. Mark Haacke, Michael R. Thompson, and Ramesh Venkatesan (Eds.). 2014. Spin density, T1, and T2 quantification methods in MR imaging. In Magnetic Resonance Imaging: Physical Principals and Sequence Design (2nd ed.). John Wiley & Sons, 637–667.
[9]
Alzheimer’s Association 2020. 2020 Alzheimer’s disease facts and figures. Alzheimer’s and Dementia 16, 3 (March 2020), 391–460.
[10]
V. Sudha, K. Priyanka, T. Suvathi Kannathal, and S. Monisha. 2020. Diabetic retinopathy detection. International Journal of Engineering and Advanced Technology 9, 4 (2020), 1022–1026.
[11]
Hugo J. W. L. Aerts, Emmanuel Rios Velazquez, Ralph T. H. Leijenaar, Chintan Parmar, Patrick Grossmann, Sara Cavalho, Johan Bussink, René Monshouwer, Benjamin Haibe-Kains, Derek Rietveld, Frank Hoebers, Michelle M. Rietbergen, C. René Leemans, Andre Dekker, John Quackenbush, Robert J. Gillies, and Philippe Lambin. 2014. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nature Communications 5, 1 (Sept. 2014), 4006.
[12]
Rupal R. Agravat and Mehul S. Raval. 2021. A survey and analysis on automated glioma brain tumor segmentation and overall patient survival prediction. Archives of Computational Methods in Engineering 28, 5 (Aug. 2021), 4117–4152.
[13]
Paul S. Aisen, Ronald C. Petersen, Michael Donohue, and Michael W. Weiner. 2015. Alzheimer’s Disease Neuroimaging Initiative 2 clinical core: Progress and plans. Alzheimer’s and Dementia 11, 7 (July 2015), 734–739.
[14]
Zeynettin Akkus, Issa Ali, Jiří Sedlář, Jay P. Agrawal, Ian F. Parney, Caterina Giannini, and Bradley J. Erickson. 2017. Predicting deletion of chromosomal arms 1p/19q in low-grade gliomas from MR images using machine intelligence. Journal of Digital Imaging 30, 4 (Aug. 2017), 469–476.
[15]
Walid Al-Dhabyani, Mohammed Gomaa, Hussien Khaled, and Aly Fahmy. 2020. Dataset of breast ultrasound images. Data in Brief 28 (2020), 104863.
[16]
Hassan Al Hajj, Mathieu Lamard, Pierre Henri Conze, Soumali Roychowdhury, Xiaowei Hu, Gabija Maršalkaitė, Odysseas Zisimopoulos, Muneer Ahmad Dedmari, Fenqiang Zhao, Jonas Prellberg, Manish Sahu, Adrian Galdran, Teresa Araújo, Duc My Vo, Chandan Panda, Navdeep Dahiya, Satoshi Kondo, Zhengbing Bian, Arash Vahdat, Jonas Bialopetravičius, Evangello Flouty, Chenhui Qiu, Sabrina Dill, Anirban Mukhopadhyay, Pedro Costa, Guilherme Aresta, Senthil Ramamurthy, Sang Woong Lee, Aurélio Campilho, Stefan Zachow, Shunren Xia, Sailesh Conjeti, Danail Stoyanov, Jogundas Armaitis, Pheng Ann Heng, William G. Macready, Béatrice Cochener, and Gwenolé Quellec. 2019. CATARACTS: Challenge on automatic tool annotation for cataRACT surgery. Medical Image Analysis 52 (Feb. 2019), 24–41.
[17]
Sharib Ali, Mariia Dmitrieva, Noha Ghatwary, Sophia Bano, Gorkem Polat, Alptekin Temizel, Adrian Krenzer, Amar Hekalo, Yun Bo Guo, Bogdan Matuszewski, Mourad Gridach, Irina Voiculescu, Vishnusai Yoganand, Arnav Chavan, Aryan Raj, Nhan T. Nguyen, Dat Q. Tran, Le Duy Huynh, Nicolas Boutry, Shahadate Rezvy, Haijian Chen, Yoon Ho Choi, Anand Subramanian, Velmurugan Balasubramanian, Xiaohong W. Gao, Hongyu Hu, Yusheng Liao, Danail Stoyanov, Christian Daul, Stefano Realdon, Renato Cannizzaro, Dominique Lamarque, Terry Tran-Nguyen, Adam Bailey, Barbara Braden, James E. East, and Jens Rittscher. 2021. Deep learning for detection and segmentation of artefact and disease instances in gastrointestinal endoscopy. Medical Image Analysis 70 (2021), 102002.
[18]
Sharib Ali, Felix Zhou, Christian Daul, Barbara Braden, Adam Bailey, Stefano Realdon, James East, Georges Wagnieres, Victor Loschenov, Enrico Grisan, Walter Blondel, and Jens Rittscher. 2019. Endoscopy artefact detection (EAD 2019) dataset. arXiv:1905.03209 (2019).
[19]
Isabel Alkhasli, Katrin Sakreida, Felix M. Mottaghy, and Ferdinand Binkofski. 2019. Modulation of fronto-striatal functional connectivity using transcranial magnetic stimulation. Frontiers in Human Neuroscience 13 (June 2019), 190.
[20]
Guilherme Aresta, Teresa Araújo, Scotty Kwok, Sai Saketh Chennamsetty, Mohammed Safwan, Varghese Alex, Bahram Marami, Marcel Prastawa, Monica Chan, Michael Donovan, Gerardo Fernandez, Jack Zeineh, Matthias Kohl, Christoph Walz, Florian Ludwig, Stefan Braunewell, Maximilian Baust, Quoc Dang Vu, Minh Nguyen Nhat To, Eal Kim, Jin Tae Kwak, Sameh Galal, Veronica Sanchez-Freire, Nadia Brancati, Maria Frucci, Daniel Riccio, Yaqi Wang, Lingling Sun, Kaiqiang Ma, Jiannan Fang, Ismael Kone, Lahsen Boulmane, Aurélio Campilho, Catarina Eloy, António Polónia, and Paulo Aguiar. 2019. BACH: Grand challenge on breast cancer histology images. Medical Image Analysis 56 (Aug. 2019), 122–139.
[21]
Samuel G. Armato, Karen Drukker, Feng Li, Lubomir Hadjiiski, Georgia D. Tourassi, Roger M. Engelmann, Maryellen L. Giger, George Redmond, Keyvan Farahani, Justin S. Kirby, and Laurence P. Clarke. 2016. LUNGx challenge for computerized lung nodule classification. Journal of Medical Imaging 3, 4 (Dec. 2016), 044506.
[22]
Samuel G. Armato, Lubomir Hadjiiski, Georgia D. Tourassi, Karen Drukker, Maryellen L. Giger, Feng Li, George Redmond, Keyvan Farahani, Justin S. Kirby, and Laurence P. Clarke. 2015. Guest editorial: LUNGx challenge for computerized lung nodule classification: Reflections and lessons learned. Journal of Medical Imaging 2, 2 (June 2015), 020103.
[23]
Samuel G. Armato, Geoffrey McLennan, Luc Bidaut, Michael F. McNitt-Gray, Charles R. Meyer, Anthony P. Reeves, Binsheng Zhao, Denise R. Aberle, Claudia I. Henschke, Eric A. Hoffman, Ella A. Kazerooni, Heber MacMahon, Edwin J. R. Van Beek, David Yankelevitz, Alberto M. Biancardi, Peyton H. Bland, Matthew S. Brown, Roger M. Engelmann, Gary E. Laderach, Daniel Max, Richard C. Pais, David P. Y. Qing, Rachael Y. Roberts, Amanda R. Smith, Adam Starkey, Poonam Batra, Philip Caligiuri, Ali Farooqi, Gregory W. Gladish, C. Matilda Jude, Reginald F. Munden, Iva Petkovska, Leslie E. Quint, Lawrence H. Schwartz, Baskaran Sundaram, Lori E. Dodd, Charles Fenimore, David Gur, Nicholas Petrick, John Freymann, Justin Kirby, Brian Hughes, Alessi Vande Casteele, Sangeeta Gupte, Maha Sallam, Michael D. Heath, Michael H. Kuhn, Ekta Dharaiya, Richard Burns, David S. Fryd, Marcos Salganicoff, Vikram Anand, Uri Shreter, Stephen Vastagh, Barbara Y. Croft, and Laurence P. Clarke. 2011. The lung image database consortium (LIDC) and image database resource initiative (IDRI): A completed reference database of lung nodules on CT scans. Medical Physics 38, 2 (Jan. 2011), 915–931.
[24]
Harish Babu Arunachalam, Rashika Mishra, Bogdan Armaselu, Ovidiu Daescu, Maria Martinez, Patrick Leavey, Dinesh Rakheja, Kevin Cederberg, Anita Sengupta, and Molly Ni’Suilleabhain. 2017. Computer aided image segmentation and classification for viable and non-viable tumor identification in osteosarcoma. In Proceedings of the Pacific Symposium on Biocomputing. 195–206.
[25]
Jong Bin Bae, Subin Lee, Wonmo Jung, Sejin Park, Weonjin Kim, Hyunwoo Oh, Ji Won Han, Grace Eun Kim, Jun Sung Kim, Jae Hyoung Kim, and Ki Woong Kim. 2020. Identification of Alzheimer’s disease using a convolutional neural network model based on T1-weighted magnetic resonance imaging. Scientific Reports 10, 1 (Dec. 2020), 22252.
[26]
Spyridon Bakas, Mauricio Reyes, Andras Jakab, Stefan Bauer, Markus Rempfler, Alessandro Crimi, Russell Takeshi Shinohara, Christoph Berger, Sung Min Ha, Martin Rozycki, Marcel Prastawa, Esther Alberts, Jana Lipkova, John Freymann, Justin Kirby, Michel Bilello, Hassan M. Fathallah-Shaykh, Roland Wiest, Jan Kirschke, Benedikt Wiestler, Rivka Colen, Aikaterini Kotrotsou, Pamela Lamontagne, Daniel Marcus, Mikhail Milchenko, Arash Nazeri, Marc Andr Weber, Abhishek Mahajan, Ujjwal Baid, Elizabeth Gerstner, Dongjin Kwon, Gagan Acharya, Manu Agarwal, Mahbubul Alam, Alberto Albiol, Antonio Albiol, Francisco J. Albiol, Varghese Alex, Nigel Allinson, Pedro H. A. Amorim, Abhijit Amrutkar, Ganesh Anand, Simon Andermatt, Tal Arbel, Pablo Arbelaez, Aaron Avery, Muneeza Azmat, B. Pranjal, Wenjia Bai, Subhashis Banerjee, Bill Barth, Thomas Batchelder, Kayhan Batmanghelich, Enzo Battistella, Andrew Beers, Mikhail Belyaev, Martin Bendszus, Eze Benson, Jose Bernal, Halandur Nagaraja Bharath, George Biros, Sotirios Bisdas, James Brown, Mariano Cabezas, Shilei Cao, Jorge M. Cardoso, Eric N. Carver, Adri Casamitjana, Laura Silvana Castillo, Marcel Cat, Philippe Cattin, Albert Cérigues, Vinicius S. Chagas, Siddhartha Chandra, Yi Ju Chang, Shiyu Chang, Ken Chang, Joseph Chazalon, Shengcong Chen, Wei Chen, Jefferson W. Chen, Zhaolin Chen, Kun Cheng, Ahana Roy Choudhury, Roger Chylla, Albert Clrigues, Steven Colleman, Ramiro German Rodriguez Colmeiro, Marc Combalia, Anthony Costa, Xiaomeng Cui, Zhenzhen Dai, Lutao Dai, Laura Alexandra Daza, Eric Deutsch, Changxing Ding, Chao Dong, Shidu Dong, Wojciech Dudzik, Zach Eaton-Rosen, Gary Egan, Guilherme Escudero, Tho Estienne, Richard Everson, Jonathan Fabrizio, Yong Fan, Longwei Fang, Xue Feng, Enzo Ferrante, Lucas Fidon, Martin Fischer, Andrew P. French, Naomi Fridman, Huan Fu, David Fuentes, Yaozong Gao, Evan Gates, David Gering, Amir Gholami, Willi Gierke, Ben Glocker, Mingming Gong, Sandra Gonzlez-Vill, T. Grosges, Yuanfang Guan, Sheng Guo, Sudeep Gupta, Woo Sup Han, Il Song Han, Konstantin Harmuth, Huiguang He, Aura Hernndez-Sabat, Evelyn Herrmann, Naveen Himthani, Winston Hsu, Cheyu Hsu, Xiaojun Hu, Xiaobin Hu, Yan Hu, Yifan Hu, Rui Hua, Teng Yi Huang, Weilin Huang, Sabine van Huffel, Quan Huo, H. V. Vivek, Khan M. Iftekharuddin, Fabian Isensee, Mobarakol Islam, Aaron S. Jackson, Sachin R. Jambawalikar, Andrew Jesson, Weijian Jian, Peter Jin, V. Jeya Maria Jose, Alain Jungo, Bernhard Kainz, Konstantinos Kamnitsas, Po Yu Kao, Ayush Karnawat, Thomas Kellermeier, Adel Kermi, Kurt Keutzer, Mohamed Tarek Khadir, Mahendra Khened, Philipp Kickingereder, Geena Kim, Nik King, Haley Knapp, Urspeter Knecht, Lisa Kohli, Deren Kong, Xiangmao Kong, Simon Koppers, Avinash Kori, Ganapathy Krishnamurthi, Egor Krivov, Piyush Kumar, Kaisar Kushibar, Dmitrii Lachinov, Tryphon Lambrou, Joon Lee, Chengen Lee, Yuehchou Lee, Matthew Chung Hai Lee, Szidonia Lefkovits, Laszlo Lefkovits, James Levitt, Tengfei Li, Hongwei Li, Wenqi Li, Hongyang Li, Xiaochuan Li, Yuexiang Li, Heng Li, Zhenye Li, Xiaoyu Li, Zeju Li, Xiao Gang Li, Wenqi Li, Zheng Shen Lin, Fengming Lin, Pietro Lio, Chang Liu, Boqiang Liu, Xiang Liu, Mingyuan Liu, Ju Liu, Luyan Liu, Xavier Lladó, Marc Moreno Lopez, Pablo Ribalta Lorenzo, Zhentai Lu, Lin Luo, Zhigang Luo, Jun Ma, Kai Ma, Thomas Mackie, Anant Madabhushi, Issam Mahmoudi, Klaus H. Maier-Hein, Pradipta Maji, C. P. Mammen, Andreas Mang, B. S. Manjunath, Michal Marcinkiewicz, Steven McDonagh, Stephen McKenna, Richard McKinley, Miriam Mehl, Sachin Mehta, Raghav Mehta, Raphael Meier, Christoph Meinel, Dorit Merhof, Craig Meyer, Robert Miller, Sushmita Mitra, Aliasgar Moiyadi, David Molina-Garcia, Miguel A. B. Monteiro, Grzegorz Mrukwa, Andriy Myronenko, Jakub Nalepa, Thuyen Ngo, Dong Nie, Holly Ning, Chen Niu, Nicholas K. Nuechterlein, Eric Oermann, Arlindo Oliveira, Diego D. C. Oliveira, Arnau Oliver, Alexander F. I. Osman, Yu Nian Ou, Sebastien Ourselin, Nikos Paragios, Moo Sung Park, Brad Paschke, J. Gregory Pauloski, Kamlesh Pawar, Nick Pawlowski, Linmin Pei, Suting Peng, Silvio M. Pereira, Julian Perez-Beteta, Victor M. Perez-Garcia, Simon Pezold, Bao Pham, Ashish Phophalia, Gemma Piella, G. N. Pillai, Marie Piraud, Maxim Pisov, Anmol Popli, Michael P. Pound, Reza Pourreza, Prateek Prasanna, Vesnakovska Pr, Tony P. Pridmore, Santi Puch, Lodie Puybareau, Buyue Qian, Xu Qiao, Martin Rajchl, Swapnil Rane, Michael Rebsamen, Hongliang Ren, Xuhua Ren, Karthik Revanuru, Mina Rezaei, Oliver Rippel, Luis Carlos Rivera, Charlotte Robert, Bruce Rosen, Daniel Rueckert, Mohammed Safwan, Mostafa Salem, Joaquim Salvi, Irina Sanchez, Irina Snchez, Heitor M. Santos, Emmett Sartor, Dawid Schellingerhout, Klaudius Scheufele, Matthew R. Scott, Artur A. Scussel, Sara Sedlar, Juan Pablo Serrano-Rubio, N. Jon Shah, Nameetha Shah, Mazhar Shaikh, B. Uma Shankar, Zeina Shboul, Haipeng Shen, Dinggang Shen, Linlin Shen, Haocheng Shen, Varun Shenoy, Feng Shi, Hyung Eun Shin, Hai Shu, Diana Sima, Matthew Sinclair, Orjan Smedby, James M. Snyder, Mohammadreza Soltaninejad, Guidong Song, Mehul Soni, Jean Stawiaski, Shashank Subramanian, Li Sun, Roger Sun, Jiawei Sun, Kay Sun, Yu Sun, Guoxia Sun, Shuang Sun, Yannick R. Suter, Laszlo Szilagyi, Sanjay Talbar, Dacheng Tao, Dacheng Tao, Zhongzhao Teng, Siddhesh Thakur, Meenakshi H. Thakur, Sameer Tharakan, Pallavi Tiwari, Guillaume Tochon, Tuan Tran, Yuhsiang M. Tsai, Kuan Lun Tseng, Tran Anh Tuan, Vadim Turlapov, Nicholas Tustison, Maria Vakalopoulou, Sergi Valverde, Rami Vanguri, Evgeny Vasiliev, Jonathan Ventura, Luis Vera, Tom Vercauteren, C. A. Verrastro, Lasitha Vidyaratne, Veronica Vilaplana, Ajeet Vivekanandan, Guotai Wang, Qian Wang, Chiatse J. Wang, Weichung Wang, Duo Wang, Ruixuan Wang, Yuanyuan Wang, Chunliang Wang, Guotai Wang, Ning Wen, Xin Wen, Leon Weninger, Wolfgang Wick, Shaocheng Wu, Qiang Wu, Yihong Wu, Yong Xia, Yanwu Xu, Xiaowen Xu, Peiyuan Xu, Tsai Ling Yang, Xiaoping Yang, Hao Yu Yang, Junlin Yang, Haojin Yang, Guang Yang, Hongdou Yao, Xujiong Ye, Changchang Yin, Brett Young-Moxon, Jinhua Yu, Xiangyu Yue, Songtao Zhang, Angela Zhang, Kun Zhang, Xuejie Zhang, Lichi Zhang, Xiaoyue Zhang, Yazhuo Zhang, Lei Zhang, Jianguo Zhang, Xiang Zhang, Tianhao Zhang, Sicheng Zhao, Yu Zhao, Xiaomei Zhao, Liang Zhao, Yefeng Zheng, Liming Zhong, Chenhong Zhou, Xiaobing Zhou, Fan Zhou, Hongtu Zhu, Jin Zhu, Ying Zhuge, Weiwei Zong, Jayashree Kalpathy-Cramer, Keyvan Farahani, Christos Davatzikos, Koen van Leemput, and Bjoern Menze. 2018. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. arXiv:1811.02629 (2018). http://arxiv.org/abs/1811.02629
[27]
Simone Balocco, Carlo Gatta, Francesco Ciompi, Andreas Wahle, Petia Radeva, Stephane Carlier, Gozde Unal, Elias Sanidas, Josepa Mauri, Xavier Carillo, Tomas Kovarnik, Ching Wei Wang, Hsiang Chou Chen, Themis P. Exarchos, Dimitrios I. Fotiadis, François Destrempes, Guy Cloutier, Oriol Pujol, Marina Alberti, E. Gerardo Mendizabal-Ruiz, Mariano Rivera, Timur Aksoy, Richard W. Downe, and Ioannis A. Kakadiaris. 2014. Standardized evaluation methodology and reference database for evaluating IVUS image segmentation. Computerized Medical Imaging and Graphics 38, 2 (March2014), 70–90.
[28]
Vivek Singh Bawa, Gurkirt Singh, Francis Kaping’a, Inna Skarga-Bandurova, Alice Leporini, Carmela Landolfo, Armando Stabile, Francesco Setti, Riccardo Muradore, Elettra Oleari, and Fabio Cuzzolin. 2020. ESAD: Endoscopic surgeon action detection dataset. arXiv:2006.07164 (2020). http://arxiv.org/abs/2006.07164
[29]
Babak Ehteshami Bejnordi, Mitko Veta, Paul Johannes Van Diest, Bram Van Ginneken, Nico Karssemeijer, Geert Litjens, Jeroen A. W. M. Van Der Laak, Meyke Hermsen, Quirine F. Manson, Maschenka Balkenhol, Oscar Geessink, Nikolaos Stathonikos, Marcory C. R. F. Van Dijk, Peter Bult, Francisco Beca, Andrew H. Beck, Dayong Wang, Aditya Khosla, Rishab Gargeya, Humayun Irshad, Aoxiao Zhong, Qi Dou, Quanzheng Li, Hao Chen, Huang Jing Lin, Pheng Ann Heng, Christian Haß, Elia Bruni, Quincy Wong, Ugur Halici, Mustafa Ümit Öner, Rengul Cetin-Atalay, Matt Berseth, Vitali Khvatkov, Alexei Vylegzhanin, Oren Kraus, Muhammad Shaban, Nasir Rajpoot, Ruqayya Awan, Korsuk Sirinukunwattana, Talha Qaiser, Yee Wah Tsang, David Tellez, Jonas Annuscheit, Peter Hufnagl, Mira Valkonen, Kimmo Kartasalo, Leena Latonen, Pekka Ruusuvuori, Kaisa Liimatainen, Shadi Albarqouni, Bharti Mungal, Ami George, Stefanie Demirci, Nassir Navab, Seiryo Watanabe, Shigeto Seno, Yoichi Takenaka, Hideo Matsuda, Hady Ahmady Phoulady, Vassili Kovalev, Alexander Kalinovsky, Vitali Liauchuk, Gloria Bueno, M. Milagro Fernandez-Carrobles, Ismael Serrano, Oscar Deniz, Daniel Racoceanu, and Rui Venâncio. 2017. Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. Journal of the American Medical Association 318, 22 (Dec. 2017), 2199–2210.
[30]
Eyal Bercovich and Marcia C. Javitt. 2018. Medical imaging: From roentgen to the digital revolution, and beyond. Rambam Maimonides Medical Journal 9, 4 (Oct.2018), e0034.
[31]
Olivier Bernard, Johan G. Bosch, Brecht Heyde, Martino Alessandrini, Daniel Barbosa, Sorina Camarasu-Pop, Frederic Cervenansky, Sébastien Valette, Oana Mirea, Michel Bernier, Pierre Marc Jodoin, Jaime Santo Domingos, Richard V. Stebbing, Kevin Keraudren, Ozan Oktay, Jose Caballero, Wei Shi, Daniel Rueckert, Fausto Milletari, Seyed Ahmad Ahmadi, Erik Smistad, Frank Lindseth, Maartje Van Stralen, Chen Wang, Örjan Smedby, Erwan Donal, Mark Monaghan, Alex Papachristidis, Marcel L. Geleijnse, Elena Galli, and Jan D’Hooge. 2016. Standardized evaluation system for left ventricular segmentation algorithms in 3D echocardiography. IEEE Transactions on Medical Imaging 35, 4 (April2016), 967–977.
[32]
Olivier Bernard, Alain Lalande, Clement Zotti, Frederick Cervenansky, Xin Yang, Pheng Ann Heng, Irem Cetin, Karim Lekadir, Oscar Camara, Miguel Angel Gonzalez Ballester, Gerard Sanroma, Sandy Napel, Steffen Petersen, Georgios Tziritas, Elias Grinias, Mahendra Khened, Varghese Alex Kollerathu, Ganapathy Krishnamurthi, Marc Michel Rohe, Xavier Pennec, Maxime Sermesant, Fabian Isensee, Paul Jager, Klaus H. Maier-Hein, Peter M. Full, Ivo Wolf, Sandy Engelhardt, Christian F. Baumgartner, Lisa M. Koch, Jelmer M. Wolterink, Ivana Isgum, Yeonggul Jang, Yoonmi Hong, Jay Patravali, Shubham Jain, Olivier Humbert, and Pierre Marc Jodoin. 2018. Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: Is the problem solved? IEEE Transactions on Medical Imaging 37, 11 (Nov. 2018), 2514–2525.
[33]
Dulari Bhatt, Chirag Patel, Hardik Talsania, Jigar Patel, Rasmika Vaghela, Sharnil Pandya, Kirit Modi, and Hemant Ghayvat. 2021. CNN variants for computer vision: History, architecture, application, challenges and future scope. Electronics 10, 20 (2021), 2470.
[34]
Patrick Bilic, Patrick Ferdinand Christa, Eugene Vorontsov, Grzegorz Chlebusr, Hao Chenm, Qi Doum, Chi Wing Fum, Xiao Hanp, Pheng Ann Hengm, Jrgen Hesserq, Samuel Kadourye, Tomasz Kopczyskiv, Miao Leo, Chunming Lio, Xiaomeng Lim, Jana Lipkova, John Lowengrubn, Hans Meiner, Jan Hendrik Moltzr, Chris Pale, Marie Pirauda, Xiaojuan Qim, Jin Qil, Markus Rempera, Karsten Rothq, Andrea Schenkr, Anjany Sekuboyinaa, Ping Zhouk, Christian Hulsemeyera, Marcel Beetza, Florian Ettlingera, Felix Gruena, Georgios Kaissisb, Fabian Lohferb, Rickmer Brarenb, Julian Holchc, Felix Hofmannc, Wieland Sommerc, Volker Heinemannc, Colin Jacobsd, Gabriel Efrain Humpire Mamanid, Bram Van Ginnekend, Gabriel Chartrande, An Tange, Michal Drozdzale, Samuel Kadourye, Avi Ben-Cohenf, Eyal Klangf, Marianne M. Amitaif, Eli Konenf, Hayit Greenspanf, Johan Moreaug, Alexandre Hostettlerg, Luc Solerg, Refael Vivantih, Adi Szeskinh, Naama Lev-Cohainh, Jacob Sosnah, Leo Joskowiczh, Ashnil Kumarw, Avinash Korex, Chunliang Wangy, Dagan Fengz, Fan Liaa, Ganapathy Krishnamurthix, Jian Heab, Jianrong Wuaa, Jinman Kimx, Jinyi Zhouac, Jun Maad, Junbo Liaa, Kevis Kokitsi Maninisae, Krishna Chaitanya Kaluvax, Lei Bix, Mahendra Khenedx, Miriam Beliverae, Qizhong Linaa, Xiaoping Yangad, Yading Yuanaf, Yinan Chenaa, Yuanqiang Liad, Yudong Qius, Yuli Wuad, and Bjoern Menzea. 2019. The liver tumor segmentation benchmark (LiTS). arXiv:1901.04056 (2019). http://arxiv.org/abs/1901.04056
[35]
B. Nicolas Bloch, Ashali Jain, and Conrade Carl Jaffe. 2015. BREAST-DIAGNOSIS. Retrieved August 23, 2023 from
[36]
B. Nicolas Bloch, Ashali Jain, and C. Carl Jaffe. 2015. PROSTATE-DIAGNOSIS. Retrieved August 23, 2023 from
[37]
N. Bloch. 2015. NC-ISBI 2013 Challenge: Automated Segmentation of Prostate Structures (ISBI-MR-Prostate-2018). Retrieved August 23, 2023 from
[38]
Hrvoje Bogunovic, Freerk Venhuizen, Sophie Klimscha, Stefanos Apostolopoulos, Alireza Bab-Hadiashar, Ulas Bagci, Mirza Faisal Beg, Loza Bekalo, Qiang Chen, Carlos Ciller, Karthik Gopinath, Amirali K. Gostar, Kiwan Jeon, Zexuan Ji, Sung Ho Kang, Dara D. Koozekanani, Donghuan Lu, Dustin Morley, Keshab K. Parhi, Hyoung Suk Park, Abdolreza Rashno, Marinko Sarunic, Saad Shaikh, Jayanthi Sivaswamy, Ruwan Tennakoon, Shivin Yadav, Sandro De Zanet, Sebastian M. Waldstein, Bianca S. Gerendas, Caroline Klaver, Clara I. Sanchez, and Ursula Schmidt-Erfurth. 2019. RETOUCH: The retinal OCT fluid detection and segmentation benchmark and challenge. IEEE Transactions on Medical Imaging 38, 8 (Aug. 2019), 1858–1874.
[39]
Dennis Bontempi, Sergio Benini, Alberto Signoroni, Michele Svanera, and Lars Muckli. 2020. CEREBRUM: A fast and fully-volumetric Convolutional Encoder-decodeR for weakly-supervised sEgmentation of BRain strUctures from out-of-the-scanner MRI. Medical Image Analysis 62 (Sept. 2020), 101688.
[40]
Peter Boord, Tara M. Madhyastha, Mary K. Askren, and Thomas J. Grabowski. 2017. Executive attention networks show altered relationship with default mode network in PD. NeuroImage: Clinical 13 (2017), 1–8.
[41]
Jannis Born, Gabriel Brändle, Manuel Cossio, Marion Disdier, Julie Goulet, Jérémie Roulin, and Nina Wiedemann. 2020. POCOVID-Net: Automatic detection of COVID-19 from a new lung ultrasound imaging dataset (POCUS). arXiv:2004.12084 (2020). http://arxiv.org/abs/2004.12084
[42]
Jannis Born, Nina Wiedemann, Gabriel Brändle, Charlotte Buhre, Bastian Rieck, and Karsten Borgwardt. 2020. Accelerating COVID-19 differential diagnosis with explainable ultrasound image analysis. arXiv:2009.06116 (2020).
[43]
Jiri Borovec, Jan Kybic, Ignacio Arganda-Carreras, Dmitry V. Sorokin, Gloria Bueno, Alexander V. Khvostikov, Spyridon Bakas, Eric I.Chao Chang, Stefan Heldmann, Kimmo Kartasalo, Leena Latonen, Johannes Lotz, Michelle Noga, Sarthak Pati, Kumaradevan Punithakumar, Pekka Ruusuvuori, Andrzej Skalski, Nazanin Tahmasebi, Masi Valkonen, Ludovic Venet, Yizhe Wang, Nick Weiss, Marek Wodzinski, Yu Xiang, Yan Xu, Yan Yan, Paul Yushkevich, Shengyu Zhao, and Arrate Munoz-Barrutia. 2020. ANHIR: Automatic non-rigid histological image registration challenge. IEEE Transactions on Medical Imaging 39, 10 (2020), 3042–3052.
[44]
M. Boulanger, Jean Claude Nunes, H. Chourak, A. Largent, S. Tahri, O. Acosta, R. De Crevoisier, C. Lafond, and A. Barateau. 2021. Deep learning methods to generate synthetic CT from MRI in radiotherapy: A literature review. Physica Medica 89 (Sept. 2021), 265–281.
[45]
Esther E. Bron, Marion Smits, Wiesje M. van der Flier, Hugo Vrenken, Frederik Barkhof, Philip Scheltens, Janne M. Papma, Rebecca M. E. Steketee, Carolina Méndez Orellana, Rozanna Meijboom, Madalena Pinto, Joana R. Meireles, Carolina Garrett, António J. Bastos-Leite, Ahmed Abdulkadir, Olaf Ronneberger, Nicola Amoroso, Roberto Bellotti, David Cárdenas-Peña, Andrés M. Álvarez-Meza, Chester V. Dolph, Khan M. Iftekharuddin, Simon F. Eskildsen, Pierrick Coupé, Vladimir S. Fonov, Katja Franke, Christian Gaser, Christian Ledig, Ricardo Guerrero, Tong Tong, Katherine R. Gray, Elaheh Moradi, Jussi Tohka, Alexandre Routier, Stanley Durrleman, Alessia Sarica, Giuseppe Di Fatta, Francesco Sensi, Andrea Chincarini, Garry M. Smith, Zhivko V. Stoyanov, Lauge Sørensen, Mads Nielsen, Sabina Tangaro, Paolo Inglese, Christian Wachinger, Martin Reuter, John C. van Swieten, Wiro J. Niessen, and Stefan Klein. 2015. Standardized evaluation of algorithms for computer-aided diagnosis of dementia based on structural MRI: The CADDementia challenge. NeuroImage 111 (May 2015), 562–579.
[46]
Mateusz Buda, Ashirbani Saha, Ruth Walsh, Sujata Ghate, Nianyi Li, Albert Świȩcicki, Joseph Y. Lo, and Maciej A. Mazurowski. 2020. Detection of masses and architectural distortions in digital breast tomosynthesis: A publicly available dataset of 5,060 patients and a deep learning model. arXiv:2001.07995 (2020). http://arxiv.org/abs/2011.07995
[47]
Gabriele Campanella, Matthew G. Hanna, Luke Geneslaw, Allen Miraflor, Vitor Werneck Krauss Silva, Klaus J. Busam, Edi Brogi, Victor E. Reuter, David S. Klimstra, and Thomas J. Fuchs. 2019. Clinical-grade computational pathology using weakly supervised deep learning on whole slide images. Nature Medicine 25, 8 (2019), 1301–1309.
[48]
Victor M. Campello, Polyxeni Gkontra, Cristian Izquierdo, Carlos Martin-Isla, Alireza Sojoudi, Peter M. Full, Klaus Maier-Hein, Yao Zhang, Zhiqiang He, Jun Ma, Mario Parreno, Alberto Albiol, Fanwei Kong, Shawn C. Shadden, Jorge Corral Acero, Vaanathi Sundaresan, Mina Saber, Mustafa Elattar, Hongwei Li, Bjoern Menze, Firas Khader, Christoph Haarburger, Cian M. Scannell, Mitko Veta, Adam Carscadden, Kumaradevan Punithakumar, Xiao Liu, Sotirios A. Tsaftaris, Xiaoqiong Huang, Xin Yang, Lei Li, Xiahai Zhuang, David Vilades, Martin L. Descalzo, Andrea Guala, Lucia La Mura, Matthias G. Friedrich, Ria Garg, Julie Lebel, Filipe Henriques, Mahir Karakas, Ersin Cavus, Steffen E. Petersen, Sergio Escala, Santi Segui, Jose F. Rodriguez-Palomares, and Karim Lekadir. 2021. Multi-centre, multi-vendor and multi-disease cardiac segmentation: The M&Ms challenge. IEEE Transactions on Medical Imaging 40, 12 (2021), 3543–3554.
[49]
Aaron Carass, Snehashis Roy, Amod Jog, Jennifer L. Cuzzocreo, Elizabeth Magrath, Adrian Gherman, Julia Button, James Nguyen, Ferran Prados, Carole H. Sudre, Manuel Jorge Cardoso, Niamh Cawley, Olga Ciccarelli, Claudia A. M. Wheeler-Kingshott, Sébastien Ourselin, Laurence Catanese, Hrishikesh Deshpande, Pierre Maurel, Olivier Commowick, Christian Barillot, Xavier Tomas-Fernandez, Simon K. Warfield, Suthirth Vaidya, Abhijith Chunduru, Ramanathan Muthuganapathy, Ganapathy Krishnamurthi, Andrew Jesson, Tal Arbel, Oskar Maier, Heinz Handels, Leonardo O. Iheme, Devrim Unay, Saurabh Jain, Diana M. Sima, Dirk Smeets, Mohsen Ghafoorian, Bram Platel, Ariel Birenbaum, Hayit Greenspan, Pierre Louis Bazin, Peter A. Calabresi, Ciprian M. Crainiceanu, Lotta M. Ellingsen, Daniel S. Reich, Jerry L. Prince, and Dzung L. Pham. 2017. Longitudinal multiple sclerosis lesion segmentation: Resource and challenge. NeuroImage 148 (March 2017), 77–102.
[50]
Carlos E. Cardenas, A. Mohamed, G. Sharp, M. Gooding, H. Veeraraghavan, and J. Yang. 2019. AAPM RT-MAC Grand Challenge 2019 (AAPM-RT-MAC). Retrieved August 23, 2023 from
[51]
Johan D. Carlin and Nikolaus Kriegeskorte. 2017. Adjudicating between face-coding models with individual-face fMRI responses. PLoS Computational Biology 13, 7 (July 2017), e1005604.
[52]
Jacob Carse and Stephen McKenna. 2019. Active learning for patch-based digital pathology using convolutional neural networks to reduce annotation costs. In Digital Pathology. Lecture Notes in Computer Science, Vol. 11435. Springer, 20–27.
[53]
Nadine Chang, John A. Pyles, Abhinav Gupta, Michael J. Tarr, and Elissa M. Aminoff. 2018. BOLD5000, a public fMRI dataset of 5000 images. arXiv:1809.01281 (2018).
[54]
Stephanie J. Chiu, Michael J. Allingham, Priyatham S. Mettu, Scott W. Cousins, Joseph A. Izatt, and Sina Farsiu. 2015. Kernel regression based segmentation of optical coherence tomography images with diabetic macular edema. Biomedical Optics Express 6, 4 (April 2015), 1172.
[55]
Muhammad E. H. Chowdhury, Tawsifur Rahman, Amith Khandakar, Rashid Mazhar, Muhammad Abdul Kadir, Zaid Bin Mahbub, Khandakar Reajul Islam, Muhammad Salman Khan, Atif Iqbal, Nasser Al Emadi, Mamun Bin Ibne Reaz, and Mohammad Tariqul Islam. 2020. Can AI help in screening viral and COVID-19 pneumonia? IEEE Access 8 (2020), 132665–132676.
[56]
Kenneth Clark, Bruce Vendt, Kirk Smith, John Freymann, Justin Kirby, Paul Koppel, Stephen Moore, Stanley Phillips, David Maffitt, Michael Pringle, Lawrence Tarbox, and Fred Prior. 2013. The cancer imaging archive (TCIA): Maintaining and operating a public information repository. Journal of Digital Imaging 26, 6 (Dec. 2013), 1045–1057.
[57]
Olivier Commowick, Audrey Istace, Michaël Kain, Baptiste Laurent, Florent Leray, Mathieu Simon, Sorina Camarasu Pop, Pascal Girard, Roxana Améli, Jean Christophe Ferré, Anne Kerbrat, Thomas Tourdias, Frédéric Cervenansky, Tristan Glatard, Jérémy Beaumont, Senan Doyle, Florence Forbes, Jesse Knight, April Khademi, Amirreza Mahbod, Chunliang Wang, Richard McKinley, Franca Wagner, John Muschelli, Elizabeth Sweeney, Eloy Roura, Xavier Lladó, Michel M. Santos, Wellington P. Santos, Abel G. Silva-Filho, Xavier Tomas-Fernandez, Hélène Urien, Isabelle Bloch, Sergi Valverde, Mariano Cabezas, Francisco Javier Vera-Olmos, Norberto Malpica, Charles Guttmann, Sandra Vukusic, Gilles Edan, Michel Dojat, Martin Styner, Simon K. Warfield, François Cotton, and Christian Barillot. 2018. Objective evaluation of multiple sclerosis lesion segmentation using a data management and processing infrastructure. Scientific Reports 8, 1 (Dec. 2018), 13650.
[58]
C. Couinaud. 1999. Liver anatomy: Portal (and suprahepatic) or biliary segmentation. Digestive Surgery 16, 6 (1999), 459–467.
[59]
Aydin Demircioğlu, Magdalena Charis Stein, Moon Sung Kim, Henrike Geske, Anton S. Quinsten, Sebastian Blex, Lale Umutlu, and Kai Nassenstein. 2021. Detecting the pulmonary trunk in CT scout views using deep learning. Scientific Reports 11, 1 (Dec. 2021), 10215.
[60]
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2010. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, Los Alamitos, CA, 248–255.
[61]
Shivang Desai, Ahmad Baghal, Thidathip Wongsurawat, Piroon Jenjaroenpun, Thomas Powell, Shaymaa Al-Shukri, Kim Gates, Phillip Farmer, Michael Rutherford, Geri Blake, Tracy Nolan, Kevin Sexton, William Bennett, Kirk Smith, Shorabuddin Syed, and Fred Prior. 2020. Chest imaging representing a COVID-19 positive rural U.S. population. Scientific Data 7 (2020), 414.
[62]
Rahul S. Desikan, Florent Ségonne, Bruce Fischl, Brian T. Quinn, Bradford C. Dickerson, Deborah Blacker, Randy L. Buckner, Anders M. Dale, R. Paul Maguire, Bradley T. Hyman, Marilyn S. Albert, and Ronald J. Killiany. 2006. An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest. NeuroImage 31, 3 (July 2006), 968–980.
[63]
Rahul Duggal, Anubha Gupta, Ritu Gupta, and Pramit Mallick. 2017. SD-Layer: Stain deconvolutional layer for CNNs in medical microscopic imaging. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2017. Lecture Notes in Computer Science, Vol. 10435. Springer, 435–443.
[64]
Rahul Duggal, Anubha Gupta, Ritu Gupta, Manya Wadhwa, and Chirag Ahuja. 2016. Overlapping cell nuclei segmentation in microscopic images using deep belief networks. In Proceedings of the ACM International Conference on Computer Vision (ICVGIP’16). Article 82, 8 pages.
[65]
Jan Egger, Christina Gsaxner, Antonio Pepe, Kelsey L. Pomykala, Frederic Jonske, Manuel Kurz, Jianning Li, and Jens Kleesiek. 2022. Medical deep learning–A systematic meta-review. Computer Methods and Programs in Biomedicine 221 (2022), 106874.
[66]
Mark Everingham, Luc Van Gool, Christopher K. I. Williams, John Winn, and Andrew Zisserman. 2010. The Pascal Visual Object Classes (VOC) challenge. International Journal of Computer Vision 88, 2 (June 2010), 303–338.
[67]
Zhonghao Fan, Johann Li, Liang Zhang, Guangming Zhu, Ping Li, Xiaoyuan Lu, Peiyi Shen, Syed Afaq Ali Shah, Mohammed Bennamoun, Tao Hua, and Wei Wei. 2021. U-net based analysis of MRI for Alzheimer’s disease diagnosis. Neural Computing and Applications 33 (April 2021), 13587–13599.
[68]
Scott H. Faro and Feroze B. Mohamed. 2010. BOLD fMRI: A guide to functional imaging for neuroscientists. Springer, New York, NY.
[69]
Evangello Flouty, Abdolrahim Kadkhodamohammadi, Imanol Luengo, Felix Fuentes-Hurtado, Hinde Taleb, Santiago Barbarisi, Gwenolé Quellec, and Danail Stoyanov. 2019. CaDIS: Cataract dataset for image segmentation. arXiv:1906.11586 (2019). http://arxiv.org/abs/1906.11586
[70]
Huazhu Fu, Fei Li, José Ignacio Orlando, Hrvoje Bogunović, Xu Sun, Jingan Liao, Yanwu Xu, Shaochong Zhang, and Xiulan Zhang. 2020. ADAM: Automatic Detection challenge on Age-related Macular degeneration. Retrieved August 23, 2023 from
[71]
Huazhu Fu, Fei Li, Xu Sun, Xingxing Cao, Jingan Liao, José Ignacio Orlando, Xing Tao, Yuexiang Li, Shihao Zhang, Mingkui Tan, Chenglang Yuan, Cheng Bian, Ruitao Xie, Jiongcheng Li, Xiaomeng Li, Jing Wang, Le Geng, Panming Li, Huaying Hao, Jiang Liu, Yan Kong, Yongyong Ren, Hrvoje Bogunović, Xiulan Zhang, and Yanwu Xu. 2020. AGE challenge: Angle closure glaucoma evaluation in anterior segment optical coherence tomography. arXiv:2005.02258 (2020). http://arxiv.org/abs/2005.02258
[72]
Debashis Ganguly, Srabonti Chakraborty, Maricel Balitanas, and Tai-Hoon Kim. 2010. Medical imaging: A review. In Security-Enriched Urban Computing and Smart Grid, Tai-Hoon Kim, Adrian Stoica, and Ruay-Shiung Chang (Eds.). Springer, Berlin, Germany, 504–516.
[73]
Marios A. Gavrielides, Lisa M. Kinnard, Kyle J. Myers, Jennifer Peregoy, William F. Pritchard, Rongping Zeng, Juan Esparza, John Karanian, and Nicholas Petrick. 2010. A resource for the assessment of lung nodule size estimation methods: Database of thoracic CT scans of an anthropomorphic phantom. Optics Express 18, 14 (July 2010), 15244.
[74]
Shiv Gehlot, Anubha Gupta, and Ritu Gupta. 2020. SDCT-AuxNet\(\theta\): DCT augmented stain deconvolutional CNN with auxiliary classifier for cancer diagnosis. Medical Image Analysis 61 (April 2020), 101661.
[75]
Olivier Gevaert, Jiajing Xu, Chuong D. Hoang, Ann N. Leung, Yue Xu, Andrew Quon, Daniel L. Rubin, Sandy Napel, and Sylvia K. Plevritis. 2012. Non-small cell lung cancer: Identifying prognostic imaging biomarkers by leveraging public gene expression microarray data—Methods and preliminary results. Radiology 264, 2 (Aug. 2012), 387–396.
[76]
Jianqiao Zhou, Xiaohong Jia, Dong Ni, Alison Noble, Ruobing Huang, Tao Tan, and Manh The Van. 2020. Thyroid nodule segmentation and classification in ultrasound images. Retrieved August 23, 2023 from
[77]
A. L. Goldberger, L. A. Amaral, L. Glass, J. M. Hausdorff, P. C. Ivanov, R. G. Mark, J. E. Mietus, G. B. Moody, C. K. Peng, and H. E. Stanley. 2000. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 101, 23 (June 2000), e215–e220.
[78]
Christina Gsaxner, Jürgen Wallner, Xiaojun Chen, Wolfgang Zemann, and Jan Egger. 2019. Facial model collection for medical augmented reality in oncologic cranio-maxillofacial surgery. Scientific Data 6, 1 (Dec. 2019), 310.
[79]
Guohua Shen, Tomoyasu Horikawa, Kei Majima, and Yukiyasu Kamitani. 2020. Deep Image Reconstruction. Retrieved August 23, 2023 from
[80]
Anubha Gupta, Pramit Mallick, Ojaswa Sharma, Ritu Gupta, and Rahul Duggal. 2018. PCSeg: Color model driven probabilistic multiphase level set based tool for plasma cell segmentation in multiple myeloma. PLoS ONE 13, 12 (2018), e0207908.
[81]
Ritu Gupta, Pramit Mallick, Rahul Duggal, Anubha Gupta, and Ojaswa Sharma. 2017. Stain color normalization and segmentation of plasma cells in microscopic images as a prelude to development of computer assisted automated disease diagnostic tool in multiple myeloma. Clinical Lymphoma Myeloma and Leukemia 17, 1 (Feb. 2017), e99.
[82]
K. Hameeteman, M. A. Zuluaga, M. Freiman, L. Joskowicz, O. Cuisenaire, L. Flórez Valencia, M. A. Gülsün, K. Krissian, J. Mille, W. C.K. Wong, M. Orkisz, H. Tek, M. Hernández Hoyos, F. Benmansour, A. C.S. Chung, S. Rozie, M. van Gils, L. van den Borne, J. Sosna, P. Berman, N. Cohen, P. C. Douek, I. Sánchez, M. Aissat, M. Schaap, C. T. Metz, G. P. Krestin, A. van der Lugt, W. J. Niessen, and T. Van Walsum. 2011. Evaluation framework for carotid bifurcation lumen segmentation and stenosis grading. Medical Image Analysis 15, 4 (Aug. 2011), 477–488.
[83]
Stephanie A. Harmon, Thomas H. Sanford, Sheng Xu, Evrim B. Turkbey, Holger Roth, Ziyue Xu, Dong Yang, Andriy Myronenko, Victoria Anderson, Amel Amalou, Maxime Blain, Michael Kassin, Dilara Long, Nicole Varble, Stephanie M. Walker, Ulas Bagci, Anna Maria Ierardi, Elvira Stellato, Guido Giovanni Plensich, Giuseppe Franceschelli, Cristiano Girlando, Giovanni Irmici, Dominic Labella, Dima Hammoud, Ashkan Malayeri, Elizabeth Jones, Ronald M. Summers, Peter L. Choyke, Daguang Xu, Mona Flores, Kaku Tamura, Hirofumi Obinata, Hitoshi Mori, Francesca Patella, Maurizio Cariati, Gianpaolo Carrafiello, Peng An, Bradford J. Wood, and Baris Turkbey. 2020. Artificial intelligence for the detection of COVID-19 pneumonia on chest CT using multinational datasets. Nature Communications 11, 1 (Dec. 2020), 4080.
[84]
Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. 2020. Mask R-CNN. IEEE Transactions on Pattern Analysis and Machine Intelligence 42, 2 (2020), 386–397.
[85]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition.770–778.
[86]
Xuehai He, Yichen Zhang, Luntian Mou, Eric Xing, and Pengtao Xie. 2020. PATHVQA: 30000+ questions for medical visual question answering. arXiv:2003.10286 (2020). http://arxiv.org/abs/2003.10286
[87]
Tobias Heimann, Brain Van Ginneken, Martin A. Styner, Yulia Arzhaeva, Volker Aurich, Christian Bauer, Andreas Beck, Christoph Becker, Reinhard Beichel, György Bekes, Fernando Bello, Gerd Binnig, Horst Bischof, Alexander Bornik, Peter M. M. Cashman, Ying Chi, Andrés Córdova, Benoit M. Dawant, Márta Fidrich, Jacob D. Furst, Daisuke Furukawa, Lars Grenacher, Joachim Hornegger, Dagmar Kainmüller, Richard I. Kitney, Hidefumi Kobatake, Hans Lamecker, Thomas Lange, Jeongjin Lee, Brian Lennon, Rui Li, Senhu Li, Hans Peter Meinzer, Gábor Németh, Daniela S. Raicu, Anne Mareike Rau, Eva M. Van Rikxoort, Mikaël Rousson, Lászlo Ruskó, Kinda A. Saddi, Günter Schmidt, Dieter Seghers, Akinobu Shimizu, Pieter Slagmolen, Erich Sorantin, Grzegorz Soza, Ruchaneewan Susomboon, Jonathan M. Waite, Andreas Wimmer, and Ivo Wolf. 2009. Comparison and evaluation of methods for liver segmentation from CT datasets. IEEE Transactions on Medical Imaging 28, 8 (2009), 1251–1265.
[88]
Nicholas Heller, Niranjan Sathianathen, Arveen Kalapara, Edward Walczak, Keenan Moore, Heather Kaluzniak, Joel Rosenberg, Paul Blake, Zachary Rengel, Makinna Oestreich, Joshua Dean, Michael Tradewell, Aneri Shah, Resha Tejpaul, Zachary Edgerton, Matthew Peterson, Shaneabbas Raza, Subodh Regmi, Nikolaos Papanikolopoulos, and Christopher Weight. 2019. The KiTS19 challenge data: 300 kidney tumor cases with clinical context, CT semantic segmentations, and surgical outcomes. arXiv:1904.00445 (2019). http://arxiv.org/abs/1904.00445
[89]
Tomoyasu Horikawa and Yukiyasu Kamitani. 2017. Generic decoding of seen and imagined objects using hierarchical visual features. Nature Communications 8, 1 (Aug. 2017), 15037.
[90]
Timothy M. Hospedales, Antreas Antoniou, Paul Micaelli, and Amos J. Storkey. 2022. Meta-learning in neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence 44 (2022), 5149–5169.
[91]
Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv:1704.04861 (2017). http://arxiv.org/abs/1704.04861
[92]
Murtadha D. Hssayeni, Muayad S. Croock, Aymen D. Salman, Hassan Falah Al-Khafaji, Zakaria A. Yahya, and Behnaz Ghoraani. 2020. Intracranial hemorrhage segmentation using a deep convolutional model. Data 5, 1 (Feb. 2020), 14.
[93]
Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian Q. Weinberger. 2017. Densely connected convolutional networks. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR’17). 2261–2269.
[94]
Yoshiharu Ikutani, Takatomi Kubo, Satoshi Nishida, Hideaki Hata, Kenichi Matsumoto, Kazushi Ikeda, and Shinji Nishimoto. 2021. Expert programmers have fine-tuned cortical representations of source code. eNeuro 8, 1 (2021), ENEURO.0405-20-2020.
[95]
Jeremy Irvin, Pranav Rajpurkar, Michael Ko, Yifan Yu, Silviana Ciurea-Ilcus, Chris Chute, Henrik Marklund, Behzad Haghgoo, Robyn Ball, Katie Shpanskaya, Jayne Seekins, David A. Mong, Safwan S. Halabi, Jesse K. Sandberg, Ricky Jones, David B. Larson, Curtis P. Langlotz, Bhavik N. Patel, Matthew P. Lungren, and Andrew Y. Ng. 2019. CheXpert: A large chest radiograph dataset with uncertainty labels and expert comparison. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence (AAAI’19), the 31st Innovative Applications of Artificial Intelligence Conference (IAAI’19), and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI’19).590–597.
[96]
Ivana Išgum, Manon J. N. L. Benders, Brian Avants, M. Jorge Cardoso, Serena J. Counsell, Elda Fischi Gomez, Laura Gui, Petra S. Huppi, Karina J. Kersbergen, Antonios Makropoulos, Andrew Melbourne, Pim Moeskops, Christian P. Mol, Maria Kuklisova-Murgasova, Daniel Rueckert, Julia A. Schnabel, Vedran Srhoj-Egekher, Jue Wu, Siying Wang, Linda S. de Vries, and Max A. Viergever. 2015. Evaluation of automatic neonatal brain segmentation algorithms: The NeoBrainS12 challenge. Medical Image Analysis 20, 1 (Feb. 2015), 135–151.
[97]
Oscar Jimenez-Del-Toro, Henning Muller, Markus Krenn, Katharina Gruenberg, Abdel Aziz Taha, Marianne Winterstein, Ivan Eggel, Antonio Foncubierta-Rodriguez, Orcun Goksel, Andras Jakab, Georgios Kontokotsios, Georg Langs, Bjoern H. Menze, Tomas Salas Fernandez, Roger Schaer, Anna Walleyo, Marc Andre Weber, Yashin Dicente Cid, Tobias Gass, Mattias Heinrich, Fucang Jia, Fredrik Kahl, Razmig Kechichian, Dominic Mai, Assaf B. Spanier, Graham Vincent, Chunliang Wang, Daniel Wyeth, and Allan Hanbury. 2016. Cloud-based evaluation of anatomical structure segmentation and landmark detection algorithms: VISCERAL anatomy benchmarks. IEEE Transactions on Medical Imaging 35, 11 (Nov. 2016), 2459–2475.
[98]
Yuan Jin, Antonio Pepe, Jianning Li, Christina Gsaxner, and Jan Egger. 2021. Deep learning and particle filter-based aortic dissection vessel tree segmentation. In Medical Imaging 2021: Biomedical Applications in Molecular, Structural, and Functional Imaging, Barjor S. Gimi and Andrzej Krol (Eds.), Vol. 11600. International Society for Optics and Photonics, SPIE, 116001W.
[99]
Alistair E. W. Johnson, Tom Pollard, Roger Mark, Seth Berkowitz, and Steven Horng. 2019. The MIMIC-CXR Database. Retrieved August 23, 2023 from
[100]
C. Daniel Johnson, Mei-Hsiu Chen, Alicia Y. Toledano, Jay P. Heiken, Abraham Dachman, Mark D. Kuo, Christine O. Menias, Betina Siewert, Jugesh I. Cheema, Richard G. Obregon, Jeff L. Fidler, Peter Zimmerman, Karen M. Horton, Kevin Coakley, Revathy B. Iyer, Amy K. Hara, Robert A. Halvorsen, Giovanna Casola, Judy Yee, Benjamin A. Herman, Lawrence J. Burgart, and Paul J. Limburg. 2008. Accuracy of CT colonography for detection of large adenomas and cancers. New England Journal of Medicine 359, 12 (Sept. 2008), 1207–1217.
[101]
Rashed Karim, Lauren Emma Blake, Jiro Inoue, Qian Tao, Shuman Jia, R. James Housden, Pranav Bhagirath, Jean Luc Duval, Marta Varela, Jonathan Behar, Loic Cadour, Rob J. van der Geest, Hubert Cochet, Maria Drangova, Maxime Sermesant, Reza Razavi, Oleg Aslanidi, Ronak Rajani, and Kawal Rhode. 2018. Algorithms for left atrial wall segmentation and thickness—Evaluation on an open-source CT and MRI image database. Medical Image Analysis 50 (Dec. 2018), 36–53.
[102]
Davood Karimi, Simon K. Warfield, and Ali Gholipour. 2021. Transfer learning in medical image segmentation: New insights from analysis of the dynamics of model parameters and learned representations. Artificial Intelligence in Medicine 116 (2021), 102078.
[103]
A. Emre Kavur, N. Sinem Gezer, Mustafa Barış, Sinem Aslan, Pierre Henri Conze, Vladimir Groza, Duc Duy Pham, Soumick Chatterjee, Philipp Ernst, Savaş Özkan, Bora Baydar, Dmitry Lachinov, Shuo Han, Josef Pauli, Fabian Isensee, Matthias Perkonigg, Rachana Sathish, Ronnie Rajan, Debdoot Sheet, Gurbandurdy Dovletov, Oliver Speck, Andreas Nürnberger, Klaus H. Maier-Hein, Gözde Bozdağı Akar, Gözde Ünal, Oğuz Dicle, and M. Alper Selver. 2021. CHAOS challenge—Combined (CT-MR) healthy abdominal organ segmentation. Medical Image Analysis 69 (Jan. 2021), 101950.
[104]
Daniel S. Kermany, Michael Goldbaum, Wenjia Cai, Carolina C. S. Valentim, Huiying Liang, Sally L. Baxter, Alex McKeown, Ge Yang, Xiaokang Wu, Fangbing Yan, Justin Dong, Made K. Prasadha, Jacqueline Pei, Magdalena Ting, Jie Zhu, Christina Li, Sierra Hewett, Jason Dong, Ian Ziyar, Alexander Shi, Runze Zhang, Lianghong Zheng, Rui Hou, William Shi, Xin Fu, Yaou Duan, Viet A. N. Huu, Cindy Wen, Edward D. Zhang, Charlotte L. Zhang, Oulan Li, Xiaobo Wang, Michael A. Singer, Xiaodong Sun, Jie Xu, Ali Tafreshi, M. Anthony Lewis, Huimin Xia, and Kang Zhang. 2018. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 172, 5 (Feb. 2018), 1122–1131.e9.
[105]
Bijen Khagi, Chung Ghiu Lee, and Goo Rak Kwon. 2019. Alzheimer’s disease classification from brain MRI based on transfer learning from CNN. In Proceedings of the 11th Biomedical Engineering International Conference (BMEiCON’18).
[106]
Salman Khan, Hossein Rahmani, Syed Afaq Ali Shah, and Mohammed Bennamoun. 2018. A Guide to Convolutional Neural Networks for Computer Vision. Synthesis Lectures on Computer Vision. Springer.
[107]
Hee E. Kim, Alejandro Cosa-Linan, Nandhini Santhanam, Mahboubeh Jannesari, Mate E. Maros, and Thomas Ganslandt. 2022. Transfer learning for medical image classification: A literature review. BMC Medical Imaging 22, 1 (April 2022), 69.
[108]
Jeoung Kun Kim, Yoo Jin Choo, Hyunkwang Shin, Gyu Sang Choi, and Min Cheol Chang. 2021. Prediction of ambulatory outcome in patients with corona radiata infarction using deep learning. Scientific Reports 11, 1 (Dec. 2021), 7989.
[109]
Yoojoong Kim, Jeong Hyeon Lee, Sunho Choi, Jeong Moon Lee, Jong Ho Kim, Junhee Seok, and Hyung Joon Joo. 2020. Validation of deep learning natural language processing algorithm for keyword extraction from pathology reports in electronic health records. Scientific Reports 10, 1 (Dec. 2020), 20265.
[110]
H. A. Kirişli, M. Schaap, C. T. Metz, A. S. Dharampal, W. B. Meijboom, S. L. Papadopoulou, A. Dedic, K. Nieman, M. A. de Graaf, M. F.L. Meijs, M. J. Cramer, A. Broersen, S. Cetin, A. Eslami, L. Flórez-Valencia, K. L. Lor, B. Matuszewski, I. Melki, B. Mohr, I. Öksüz, R. Shahzad, C. Wang, P. H. Kitslaar, G. Unal, A. Katouzian, M. Orkisz, C. M. Chen, F. Precioso, L. Najman, S. Masood, D. Ünay, L. van Vliet, R. Moreno, R. Goldenberg, E. Vuçini, G. P. Krestin, W. J. Niessen, and T. Van Walsum. 2013. Standardized evaluation framework for evaluating coronary artery stenosis detection, stenosis quantification and lumen segmentation algorithms in computed tomography angiography. Medical Image Analysis 17, 8 (Dec.2013), 859–876.
[111]
Arno Klein and Jason Tourville. 2012. 101 labeled brain images and a consistent human cortical labeling protocol. Frontiers in Neuroscience 6 (2012), 171.
[112]
Oldřich Kodym, Jianning Li, Antonio Pepe, Christina Gsaxner, Sasank Chilamkurthy, Jan Egger, and Michal Španěl. 2021. SkullBreak/SkullFix—Dataset for automatic cranial implant design and a benchmark for volumetric shape learning tasks. Data in Brief 35 (2021), 106902.
[113]
Laura Koenders, Janna Cousijn, Wilhelmina A. M. Vingerhoets, Wim Van Den Brink, Reinout W. Wiers, Carin J. Meijer, Marise W. J. Machielsen, Dick J. Veltman, Anneke E. Goudriaan, and Lieuwe De Haan. 2016. Grey matter changes associated with heavy cannabis use: A longitudinal sMRI study. PLoS ONE 11, 5 (May 2016), e0152482.
[114]
Lale Kostakoglu, Fenghai Duan, Michael O. Idowu, Paul R. Jolles, Harry D. Bear, Mark Muzi, Jean Cormack, John P. Muzi, Daniel A. Pryma, Jennifer M. Specht, Linda Hovanessian-Larsen, John Miliziano, Sharon Mallett, Anthony F. Shields, and David A. Mankoff. 2015. A phase II study of 3\(^{\prime }\)-Deoxy-3\(^{\prime }\)-18F-fluorothymidine PET in the assessment of early response of breast cancer to neoadjuvant chemotherapy: Results from ACRIN 6688. Journal of Nuclear Medicine 56, 11 (Nov. 2015), 1681–1689.
[115]
A. Krizhevsky, V. Nair, and G. Hinton. 2009. CIFAR-10 and CIFAR-100 Datasets. University of Toronto.
[116]
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2017. ImageNet classification with deep convolutional neural networks. Communications of the ACM 60 (2017), 84–90.
[117]
Hugo J. Kuijf, Adrià Casamitjana, D. Louis Collins, Mahsa Dadar, Achilleas Georgiou, Mohsen Ghafoorian, Dakai Jin, April Khademi, Jesse Knight, Hongwei Li, Xavier Lladó, J. Matthijs Biesbroek, Miguel Luna, Qaiser Mahmood, Richard Mckinley, Alireza Mehrtash, Sebastien Ourselin, Bo Yong Park, Hyunjin Park, Sang Hyun Park, Simon Pezold, Elodie Puybareau, Jeroen De Bresser, Leticia Rittner, Carole H. Sudre, Sergi Valverde, Veronica Vilaplana, Roland Wiest, Yongchao Xu, Ziyue Xu, Guodong Zeng, Jianguo Zhang, Guoyan Zheng, Rutger Heinen, Christopher Chen, Wiesje Van Der Flier, Frederik Barkhof, Max A. Viergever, Geert Jan Biessels, Simon Andermatt, Mariana Bento, Matt Berseth, Mikhail Belyaev, and M. Jorge Cardoso. 2019. Standardized assessment of automatic segmentation of white matter hyperintensities and results of the WMH segmentation challenge. IEEE Transactions on Medical Imaging 38, 11 (Nov. 2019), 2556–2568.
[118]
Neeraj Kumar, Ruchika Verma, Deepak Anand, Yanning Zhou, Omer Fahri Onder, Efstratios Tsougenis, Hao Chen, Pheng Ann Heng, Jiahui Li, Zhiqiang Hu, Yunzhi Wang, Navid Alemi Koohbanani, Mostafa Jahanifar, Neda Zamani Tajeddin, Ali Gooya, Nasir Rajpoot, Xuhua Ren, Sihang Zhou, Qian Wang, Dinggang Shen, Cheng Kun Yang, Chi Hung Weng, Wei Hsiang Yu, Chao Yuan Yeh, Shuang Yang, Shuoyu Xu, Pak Hei Yeung, Peng Sun, Amirreza Mahbod, Gerald Schaefer, Isabella Ellinger, Rupert Ecker, Orjan Smedby, Chunliang Wang, Benjamin Chidester, That Vinh Ton, Minh Triet Tran, Jian Ma, Minh N. Do, Simon Graham, Quoc Dang Vu, Jin Tae Kwak, Akshaykumar Gunda, Raviteja Chunduri, Corey Hu, Xiaoyang Zhou, Dariush Lotfi, Reza Safdari, Antanas Kascenas, Alison O’Neil, Dennis Eschweiler, Johannes Stegmaier, Yanping Cui, Baocai Yin, Kailin Chen, Xinmei Tian, Philipp Gruening, Erhardt Barth, Elad Arbel, Itay Remer, Amir Ben-Dor, Ekaterina Sirazitdinova, Matthias Kohl, Stefan Braunewell, Yuexiang Li, Xinpeng Xie, Linlin Shen, Jun Ma, Krishanu Das Baksi, Mohammad Azam Khan, Jaegul Choo, Adrian Colomer, Valery Naranjo, Linmin Pei, Khan M. Iftekharuddin, Kaushiki Roy, Debotosh Bhattacharjee, Anibal Pedraza, Maria Gloria Bueno, Sabarinathan Devanathan, Saravanan Radhakrishnan, Praveen Koduganty, Zihan Wu, Guanyu Cai, Xiaojie Liu, Yuqin Wang, and Amit Sethi. 2020. A multi-organ nucleus segmentation challenge. IEEE Transactions on Medical Imaging 39, 5 (May2020), 1380–1391.
[119]
Marjan Laal. 2013. Innovation process in medical imaging. Procedia—Social and Behavioral Sciences 81 (2013), 60–64.
[120]
Pamela LaMontagne, Tammie L. S. Benzinger, John Morris, Sarah Keefe, Russ Hornbeck, Chengjie Xiong, Elizabeth Grant, Jason Hassenstab, Krista Moulder, Andrei Vlassenko, Marcus Raichle, Carlos Cruchaga, and Daniel Marcus. 2019. OASIS-3: Longitudinal neuroimaging, clinical, and cognitive dataset for normal aging and Alzheimer disease. medRxiv 2019 (Jan. 2019), 2019.12.13.19014902.
[121]
Sarah Leclerc, Erik Smistad, Joao Pedrosa, Andreas Ostvik, Frederic Cervenansky, Florian Espinosa, Torvald Espeland, Erik Andreas Rye Berg, Pierre Marc Jodoin, Thomas Grenier, Carole Lartizien, Jan Dhooge, Lasse Lovstakken, and Olivier Bernard. 2019. Deep learning for segmentation using an open large-scale dataset in 2D echocardiography. IEEE Transactions on Medical Imaging 38, 9 (Sept. 2019), 2198–2210.
[122]
Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. 1998. Gradient-based learning applied to document recognition. Proceedings of the IEEE 86, 11 (1998), 2278–2323.
[123]
Rebecca J. Lepping, Ruth Ann Atchley, Evangelia Chrysikou, Laura E. Martin, Alicia A. Clair, Rick E. Ingram, W. Kyle Simmons, and Cary R. Savage. 2016. Neural processing of emotional musical and nonmusical stimuli in depression. PLoS ONE 11, 6 (June 2016), e0156859.
[124]
Jianning Li, Christina Gsaxner, Antonio Pepe, Ana Morais, Victor Alves, Gord von Campe, Jürgen Wallner, and Jan Egger. 2021. Synthetic skull bone defects for automatic patient-specific craniofacial implant design. Scientific Data 8, 1 (Jan. 2021), 36.
[125]
Jianning Li, Antonio Pepe, Christina Gsaxner, Gord von Campe, and Jan Egger. 2020. A baseline approach for AutoImplant: The MICCAI 2020 cranial implant design challenge. In Multimodal Learning for Clinical Decision Support and Clinical Image-Based Procedures. Lecture Notes in Computer Science, Vol. 12445. Springer, 75–84. arxiv:2006.12449
[126]
Jiahui Li, Shuang Yang, Xiaodi Huang, Qian Da, Xiaoqun Yang, Zhiqiang Hu, Qi Duan, Chaofu Wang, and Hongsheng Li. 2019. Signet ring cell detection with a semi-supervised learning framework. In Information Processing in Medical Imaging. Lecture Notes in Computer Science, Vol. 11492. Springer, 842–854.
[127]
Ping Li, Xiangwen Kong, Johann Li, Guangming Zhu, Xiaoyuan Lu, Peiyi Shen, Syed Afaq Ali Shah, Mohammed Bennamoun, and Tao Hua. 2021. A dataset of pulmonary lesions with multiple-level attributes and fine contours. Frontiers in Digital Health 2 (Feb. 2021), 609349.
[128]
Zhang Li, Zheyu Hu, Jiaolong Xu, Tao Tan, Hui Chen, Zhi Duan, Ping Liu, Jun Tang, Guoping Cai, Quchang Ouyang, Yuling Tang, Geert Litjens, and Qiang Li. 2018. Computer-aided diagnosis of lung carcinoma using deep learning—A pilot study. arXiv:1803.05471 (2018). http://arxiv.org/abs/1803.05471
[129]
D. Li, P. Wang, S. Li, T. Lu, J. HuangFu, and Y. Wang. 2020. A Large-Scale CT and PET/CT Dataset for Lung Cancer Diagnosis (Lung-PET-CT-Dx). Retrieved August 23, 2023 from
[130]
Huiying Liang, Brian Y. Tsui, Hao Ni, Carolina C. S. Valentim, Sally L. Baxter, Guangjian Liu, Wenjia Cai, Daniel S. Kermany, Xin Sun, Jiancong Chen, Liya He, Jie Zhu, Pin Tian, Hua Shao, Lianghong Zheng, Rui Hou, Sierra Hewett, Gen Li, Ping Liang, Xuan Zang, Zhiqi Zhang, Liyan Pan, Huimin Cai, Rujuan Ling, Shuhua Li, Yongwang Cui, Shusheng Tang, Hong Ye, Xiaoyan Huang, Waner He, Wenqing Liang, Qing Zhang, Jianmin Jiang, Wei Yu, Jianqun Gao, Wanxing Ou, Yingmin Deng, Qiaozhen Hou, Bei Wang, Cuichan Yao, Yan Liang, Shu Zhang, Yaou Duan, Runze Zhang, Sarah Gibson, Charlotte L. Zhang, Oulan Li, Edward D. Zhang, Gabriel Karin, Nathan Nguyen, Xiaokang Wu, Cindy Wen, Jie Xu, Wenqin Xu, Bochu Wang, Winston Wang, Jing Li, Bianca Pizzato, Caroline Bao, Daoman Xiang, Wanting He, Suiqin He, Yugui Zhou, Weldon Haw, Michael Goldbaum, Adriana Tremoulet, Chun Nan Hsu, Hannah Carter, Long Zhu, Kang Zhang, and Huimin Xia. 2019. Evaluation and accurate diagnoses of pediatric diseases using artificial intelligence. Nature Medicine 25, 3 (March 2019), 433–438.
[131]
Sook Lei Liew, Julia M. Anglin, Nick W. Banks, Matt Sondag, Kaori L. Ito, Hosung Kim, Jennifer Chan, Joyce Ito, Connie Jung, Nima Khoshab, Stephanie Lefebvre, William Nakamura, David Saldana, Allie Schmiesing, Cathy Tran, Danny Vo, Tyler Ard, Panthea Heydari, Bokkyu Kim, Lisa Aziz-Zadeh, Steven C. Cramer, Jingchun Liu, Surjo Soekadar, Jan Egil Nordvik, Lars T. Westlye, Junping Wang, Carolee Winstein, Chunshui Yu, Lei Ai, Bonhwang Koo, R. Cameron Craddock, Michael Milham, Matthew Lakich, Amy Pienta, and Alison Stroud. 2018. A large, open source dataset of stroke anatomical brain images and manual lesion segmentations. Scientific Data 5 (2018), 180011.
[132]
Tsung Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C. Lawrence Zitnick. 2014. Microsoft COCO: Common objects in context. In Computer Vision—ECCV 2014. Lecture Notes in Computer Science, Vol. 8693. Springer, 740–755.
[133]
Geert Litjens, Peter Bandi, Babak Ehteshami Bejnordi, Oscar Geessink, Maschenka Balkenhol, Peter Bult, Altuna Halilovic, Meyke Hermsen, Rob van de Loo, Rob Vogels, Quirine F. Manson, Nikolas Stathonikos, Alexi Baidoshvili, Paul van Diest, Carla Wauters, Marcory van Dijk, and Jeroen van der Laak. 2018. 1399 H&E-stained sentinel lymph node sections of breast cancer patients: The CAMELYON dataset. GigaScience 7, 6 (June 2018), giy065.
[134]
Geert Litjens, Oscar Debats, Jelle Barentsz, Nico Karssemeijer, and Henkjan Huisman. 2014. Computer-aided detection of prostate cancer in MRI. IEEE Transactions on Medical Imaging 33, 5 (May 2014), 1083–1092.
[135]
Geert Litjens, Robert Toth, Wendy van de Ven, Caroline Hoeks, Sjoerd Kerkstra, Bram van Ginneken, Graham Vincent, Gwenael Guillard, Neil Birbeck, Jindang Zhang, Robin Strand, Filip Malmberg, Yangming Ou, Christos Davatzikos, Matthias Kirschner, Florian Jung, Jing Yuan, Wu Qiu, Qinquan Gao, Philip Eddie Edwards, Bianca Maan, Ferdinand van der Heijden, Soumya Ghose, Jhimli Mitra, Jason Dowling, Dean Barratt, Henkjan Huisman, and Anant Madabhushi. 2014. Evaluation of prostate segmentation algorithms for MRI: The PROMISE12 challenge. Medical Image Analysis 18, 2 (2014), 359–373.
[136]
Justin Kirby. 2015. Prostate-3T. Retrieved August 23, 2023 from
[137]
Ning Liu, Ying Liu, Brent Logan, Zhiyuan Xu, Jian Tang, and Yanzhi Wang. 2019. Learning the dynamic treatment regimes from medical registry data through deep Q-network. Scientific Reports 9, 1 (Dec. 2019), 1495.
[138]
Pechin Lo, Bram Van Ginneken, Joseph M. Reinhardt, Tarunashree Yavarna, Pim A. De Jong, Benjamin Irving, Catalin Fetita, Margarete Ortner, Rômulo Pinho, Jan Sijbers, Marco Feuerstein, Anna Fabijanska, Christian Bauer, Reinhard Beichel, Carlos S. Mendoza, Rafael Wiemker, Jaesung Lee, Anthony P. Reeves, Silvia Born, Oliver Weinheimer, Eva M. Van Rikxoort, Juerg Tschirren, Ken Mori, Benjamin Odry, David P. Naidich, Ieneke Hartmann, Eric A. Hoffman, Mathias Prokop, Jesper H. Pedersen, and Marleen De Bruijne. 2012. Extraction of airways from CT (EXACT’09). IEEE Transactions on Medical Imaging 31, 11 (2012), 2093–2107.
[139]
Jonathan Long, Evan Shelhamer, and Trevor Darrell. 2015. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE, Los Alamitos, CA, 431–440.
[140]
Tara M. Madhyastha, Mary K. Askren, Peter Boord, and Thomas J. Grabowski. 2015. Dynamic connectivity at rest predicts attention task performance. Brain Connectivity 5, 1 (Feb. 2015), 45–59.
[141]
Gianluca Maguolo and Loris Nanni. 2020. A critic evaluation of methods for COVID-19 automatic detection from x-ray images. arXiv:2004.12823 (2020). http://arxiv.org/abs/2004.12823
[142]
Oskar Maier, Bjoern H. Menze, Janina von der Gablentz, Levin Häni, Mattias P. Heinrich, Matthias Liebrand, Stefan Winzeck, Abdul Basit, Paul Bentley, Liang Chen, Daan Christiaens, Francis Dutil, Karl Egger, Chaolu Feng, Ben Glocker, Michael Götz, Tom Haeck, Hanna Leena Halme, Mohammad Havaei, Khan M. Iftekharuddin, Pierre Marc Jodoin, Konstantinos Kamnitsas, Elias Kellner, Antti Korvenoja, Hugo Larochelle, Christian Ledig, Jia Hong Lee, Frederik Maes, Qaiser Mahmood, Klaus H. Maier-Hein, Richard McKinley, John Muschelli, Chris Pal, Linmin Pei, Janaki Raman Rangarajan, Syed M. S. Reza, David Robben, Daniel Rueckert, Eero Salli, Paul Suetens, Ching Wei Wang, Matthias Wilms, Jan S. Kirschke, Ulrike M. Krämer, Thomas F. Münte, Peter Schramm, Roland Wiest, Heinz Handels, and Mauricio Reyes. 2017. ISLES 2015—A public evaluation benchmark for ischemic stroke lesion segmentation from multispectral MRI. Medical Image Analysis 35 (Jan. 2017), 250–269.
[143]
Cui P. Mao, Fen R. Chen, Jiao H. Huo, Liang Zhang, Gui R. Zhang, Bing Zhang, and Xiao Q. Zhou. 2020. Altered resting-state functional connectivity and effective connectivity of the habenula in irritable bowel syndrome: A cross-sectional and machine learning study. Human Brain Mapping 41, 13 (Sept. 2020), 3655–3666.
[144]
Daniel S. Marcus, Anthony F. Fotenos, John G. Csernansky, John C. Morris, and Randy L. Buckner. 2010. Open Access Series of Imaging Studies: Longitudinal MRI data in nondemented and demented older adults. Journal of Cognitive Neuroscience 22, 12 (Dec. 2010), 2677–2684.
[145]
Daniel S. Marcus, Tracy H. Wang, Jamie Parker, John G. Csernansky, John C. Morris, and Randy L. Buckner. 2007. Open Access Series of Imaging Studies (OASIS): Cross-sectional MRI data in young, middle aged, nondemented, and demented older adults. Journal of Cognitive Neuroscience 19, 9 (Sept. 2007), 1498–1507.
[146]
Razvan V. Marinescu, Neil P. Oxtoby, Alexandra L. Young, Esther E. Bron, Arthur W. Toga, Michael W. Weiner, Frederik Barkhof, Nick C. Fox, Arman Eshaghi, Tina Toni, Marcin Salaterski, Veronika Lunina, Manon Ansart, Stanley Durrleman, Pascal Lu, Samuel Iddi, Dan Li, Wesley K. Thompson, Michael C. Donohue, Aviv Nahon, Yarden Levy, Dan Halbersberg, Mariya Cohen, Huiling Liao, Tengfei Li, Kaixian Yu, Hongtu Zhu, José G. Tamez-Peña, Aya Ismail, Timothy Wood, Hector Corrada Bravo, Minh Nguyen, Nanbo Sun, Jiashi Feng, B. T. Thomas Yeo, Gang Chen, Ke Qi, Shiyang Chen, Deqiang Qiu, Ionut Buciuman, Alex Kelner, Raluca Pop, Denisa Rimocea, Mostafa M. Ghazi, Mads Nielsen, Sebastien Ourselin, Lauge Sørensen, Vikram Venkatraghavan, Keli Liu, Christina Rabe, Paul Manser, Steven M. Hill, James Howlett, Zhiyue Huang, Steven Kiddle, Sach Mukherjee, Anaïs Rouanet, Bernd Taschler, Brian D. M. Tom, Simon R. White, Noel Faux, Suman Sedai, Javier de Velasco Oriol, Edgar E. V. Clemente, Karol Estrada, Leon Aksman, Andre Altmann, Cynthia M. Stonnington, Yalin Wang, Jianfeng Wu, Vivek Devadas, Clementine Fourrier, Lars Lau Raket, Aristeidis Sotiras, Guray Erus, Jimit Doshi, Christos Davatzikos, Jacob Vogel, Andrew Doyle, Angela Tam, Alex Diaz-Papkovich, Emmanuel Jammeh, Igor Koval, Paul Moore, Terry J. Lyons, John Gallacher, Jussi Tohka, Robert Ciszek, Bruno Jedynak, Kruti Pandya, Murat Bilgel, William Engels, Joseph Cole, Polina Golland, Stefan Klein, and Daniel C. Alexander. 2020. The Alzheimer’s Disease Prediction of Longitudinal Evolution (TADPOLE) Challenge: Results after 1 year follow-up. arXiv:2002.03419 (2020). http://arxiv.org/abs/2002.03419
[147]
Mario Mascalchi, Chiara Marzi, Marco Giannelli, Stefano Ciulli, Andrea Bianchi, Andrea Ginestroni, Carlo Tessa, Emanuele Nicolai, Marco Aiello, Elena Salvatore, Andrea Soricelli, and Stefano Diciotti. 2018. Histogram analysis of DTI-derived indices reveals pontocerebellar degeneration and its progression in SCA2. PLoS ONE 13, 7 (July 2018), e0200258.
[148]
Christian Matek, Simone Schwarz, Karsten Spiekermann, and Carsten Marr. 2019. Human-level recognition of blast cells in acute myeloid leukaemia with convolutional neural networks. Nature Machine Intelligence 1, 11 (Nov. 2019), 538–544.
[149]
Maryam Mehdizadeh, Cara MacNish, Di Xiao, David Alonso-Caneiro, Jason Kugelman, and Mohammed Bennamoun. 2021. Deep feature loss to denoise OCT images using deep neural networks. Journal of Biomedical Optics 26, 04 (April 2021), 046003.
[150]
Adriënne M. Mendrik, Koen L. Vincken, Hugo J. Kuijf, Marcel Breeuwer, Willem H. Bouvy, Jeroen De Bresser, Amir Alansary, Marleen De Bruijne, Aaron Carass, Ayman El-Baz, Amod Jog, Ranveer Katyal, Ali R. Khan, Fedde Van Der Lijn, Qaiser Mahmood, Ryan Mukherjee, Annegreet Van Opbroek, Sahil Paneri, Sérgio Pereira, Mikael Persson, Martin Rajchl, Duygu Sarikaya, Örjan Smedby, Carlos A. Silva, Henri A. Vrooman, Saurabh Vyas, Chunliang Wang, Liang Zhao, Geert Jan Biessels, and Max A. Viergever. 2015. MRBrainS challenge: Online evaluation framework for brain image segmentation in 3T MRI scans. Computational Intelligence and Neuroscience 2015 (2015), 1–16.
[151]
Bjoern H. Menze, Andras Jakab, Stefan Bauer, Jayashree Kalpathy-Cramer, Keyvan Farahani, Justin Kirby, Yuliya Burren, Nicole Porz, Johannes Slotboom, Roland Wiest, Levente Lanczi, Elizabeth Gerstner, Marc André Weber, Tal Arbel, Brian B. Avants, Nicholas Ayache, Patricia Buendia, D. Louis Collins, Nicolas Cordier, Jason J. Corso, Antonio Criminisi, Tilak Das, Hervé Delingette, Çağatay Demiralp, Christopher R. Durst, Michel Dojat, Senan Doyle, Joana Festa, Florence Forbes, Ezequiel Geremia, Ben Glocker, Polina Golland, Xiaotao Guo, Andac Hamamci, Khan M. Iftekharuddin, Raj Jena, Nigel M. John, Ender Konukoglu, Danial Lashkari, José António Mariz, Raphael Meier, Sérgio Pereira, Doina Precup, Stephen J. Price, Tammy Riklin Raviv, Syed M. S. Reza, Michael Ryan, Duygu Sarikaya, Lawrence Schwartz, Hoo Chang Shin, Jamie Shotton, Carlos A. Silva, Nuno Sousa, Nagesh K. Subbanna, Gabor Szekely, Thomas J. Taylor, Owen M. Thomas, Nicholas J. Tustison, Gozde Unal, Flor Vasseur, Max Wintermark, Dong Hye Ye, Liang Zhao, Binsheng Zhao, Darko Zikic, Marcel Prastawa, Mauricio Reyes, and Koen Van Leemput. 2015. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Transactions on Medical Imaging 34, 10 (Oct. 2015), 1993–2024.
[152]
Rashika Mishra, Ovidiu Daescu, Patrick Leavey, Dinesh Rakheja, and Anita Sengupta. 2017. Histopathological diagnosis for viable and non-viable tumor prediction for osteosarcoma using convolutional neural network. In Bioinformatics Research and Applications. Lecture Notes in Computer Science, Vol. 10330. Springer, 12–23.
[153]
Rashika Mishra, Ovidiu Daescu, Patrick Leavey, Dinesh Rakheja, and Anita Sengupta. 2018. Convolutional neural network for histopathological analysis of osteosarcoma. Journal of Computational Biology 25 (2018), 313–325.
[154]
Yoichi Miyawaki, Hajime Uchida, Okito Yamashita, Masa aki Sato, Yusuke Morito, Hiroki C. Tanabe, Norihiro Sadato, and Yukiyasu Kamitani. 2008. Visual image reconstruction from human brain activity using a combination of multiscale local image decoders. Neuron 60, 5 (Dec. 2008), 915–929.
[155]
Arnab Kumar Mondal, Jose Dolz, and Christian Desrosiers. 2018. Few-shot 3D multi-modal medical image segmentation using generative adversarial learning. arXiv:1810.12241 (2018). http://arxiv.org/abs/1810.12241
[156]
Susanne G. Mueller, Michael W. Weiner, Leon J. Thal, Ronald C. Petersen, Clifford Jack, William Jagust, John Q. Trojanowski, Arthur W. Toga, and Laurel Beckett. 2005. The Alzheimer’s Disease Neuroimaging Initiative. Neuroimaging Clinics of North America 15, 4 (2005), 869–877.
[157]
David Newitt and Nola Hylton. 2016. Single site breast DCE-MRI data and segmentations from patients undergoing neoadjuvant chemotherapy. Cancer Imaging Archive. Retrieved August 23, 2023 from
[158]
Meindert Niemeijer, Bram Van Ginneken, Michael J. Cree, Atsushi Mizutani, Gwénolé Quellec, Clara I. Sanchez, Bob Zhang, Roberto Hornero, Mathieu Lamard, Chisako Muramatsu, Xiangqian Wu, Guy Cazuguel, Jane You, Agustín Mayo, Qin Li, Yuji Hatanaka, Béatrice Cochener, Christian Roux, Fakhri Karray, María Garcia, Hiroshi Fujita, and Michael D. Abramoff. 2010. Retinopathy online challenge: Automatic detection of microaneurysms in digital color fundus photographs. IEEE Transactions on Medical Imaging 29, 1 (2010), 185–195.
[159]
José Ignacio Orlando, Huazhu Fu, João Barbossa Breda, Karel van Keer, Deepti R. Bathula, Andrés Diaz-Pinto, Ruogu Fang, Pheng Ann Heng, Jeyoung Kim, Joon Ho Lee, Joonseok Lee, Xiaoxiao Li, Peng Liu, Shuai Lu, Balamurali Murugesan, Valery Naranjo, Sai Samarth R. Phaye, Sharath M. Shankaranarayana, Apoorva Sikka, Jaemin Son, Anton van den Hengel, Shujun Wang, Junyan Wu, Zifeng Wu, Guanghui Xu, Yongli Xu, Pengshuai Yin, Fei Li, Xiulan Zhang, Yanwu Xu, and Hrvoje Bogunović. 2020. REFUGE Challenge: A unified framework for evaluating automated methods for glaucoma assessment from fundus photographs. Medical Image Analysis 59 (2020), 101570
[160]
P. Leavey, H. B. Arunachalam, B. Armaselu, A. Sengupta, D. Rakheja, S. Skapek, K. Cederberg, J.-P. Bach, S. Glick, M. Ni’Suilleabhain, R. Mishra, M. Martinez, R. Ziraldo, and D. Leonard.2017. American Society of Pediatric Hematology/Oncology (ASPHO) Palais des congrés de Montréal Montréal, Canada April 26–29, 2017. Pediatric Blood & Cancer 64, Suppl. 1 (2017).
[161]
Danielle F. Pace, Adrian V. Dalca, Tal Geva, Andrew J. Powell, Mehdi H. Moghari, and Polina Golland. 2015. Interactive whole-heart segmentation in congenital heart disease. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015. Lecture Notes in Computer Science, Vol. 9351. Springer, 80–88.
[162]
Yongsheng Pan, Mingxia Liu, Yong Xia, and Dinggang Shen. 2019. Neighborhood-Correction Algorithm for Cells. Springer Singapore, Singapore.
[163]
Sharnil Pandya, Aanchal Thakur, Santosh Saxena, Nandita Jassal, Chirag Patel, Kirit Modi, Pooja Shah, Rahul Joshi, Sudhanshu Gonge, Kalyani Kadam, and Prachi Kadam. 2021. A study of the recent trends of immunology: Key challenges, domains, applications, datasets, and future directions. Sensors 21, 23 (2021), 7786.
[164]
Chirag Patel, Dulari Bhatt, Urvashi Sharma, Radhika Patel, Sharnil Pandya, Kirit Modi, Nagaraj Cholli, Akash Patel, Urvi Bhatt, Muhammad Ahmed Khan, Shubhankar Majumdar, Mohd Zuhair, Khushi Patel, Syed Aziz Shah, and Hemant Ghayvat. 2022. DBGC: Dimension-based generic convolution block for object recognition. Sensors 22, 5 (24 Feb.2022), 1780.
[165]
Chirag I. Patel, Sanjay Garg, Tanish Zaveri, Asim Banerjee, and Ripal Patel. 2018. Human action recognition using fusion of features for unconstrained video sequences. Computers & Electrical Engineering 70 (2018), 284–301.
[166]
Chirag I. Patel, Dileep Labana, Sharnil Pandya, Kirit Modi, Hemant Ghayvat, and Muhammad Awais. 2020. Histogram of oriented gradient-based fusion of features for human action recognition in action video sequences. Sensors 20, 24 (2020), 7299.
[167]
Mohammad Peikari, Sherine Salama, Sharon Nofech-Mozes, and Anne L. Martel. 2017. Automatic cellularity assessment from post-treated breast surgical specimens. Cytometry Part A 91, 11 (2017), 1078–1087.
[168]
Hong Peng, Jiaohua Huo, Bo Li, Yuanyuan Cui, Hao Zhang, Liang Zhang, and Lin Ma. 2021. Predicting isocitrate dehydrogenase (IDH) mutation status in gliomas using multiparameter MRI radiomics features. Journal of Magnetic Resonance Imaging 53, 5 (May 2021), 1399–1407.
[169]
A. Gh. Podoleanu. 2012. Optical coherence tomography. Journal of Microscopy 247, 3 (2012), 209–219.
[170]
Prasanna Porwal, Samiksha Pachade, Manesh Kokare, Girish Deshmukh, Jaemin Son, Woong Bae, Lihong Liu, Jianzong Wang, Xinhui Liu, Liangxin Gao, Tian Bo Wu, Jing Xiao, Fengyan Wang, Baocai Yin, Yunzhi Wang, Gopichandh Danala, Linsheng He, Yoon Ho Choi, Yeong Chan Lee, Sang Hyuk Jung, Zhongyu Li, Xiaodan Sui, Junyan Wu, Xiaolong Li, Ting Zhou, Janos Toth, Agnes Baran, Avinash Kori, Sai Saketh Chennamsetty, Mohammed Safwan, Varghese Alex, Xingzheng Lyu, Li Cheng, Qinhao Chu, Pengcheng Li, Xin Ji, Sanyuan Zhang, Yaxin Shen, Ling Dai, Oindrila Saha, Rachana Sathish, Tânia Melo, Teresa Araújo, Balazs Harangi, Bin Sheng, Ruogu Fang, Debdoot Sheet, Andras Hajdu, Yuanjie Zheng, Ana Maria Mendonça, Shaoting Zhang, Aurélio Campilho, Bin Zheng, Dinggang Shen, Luca Giancardo, Gwenolé Quellec, and Fabrice Mériaudeau. 2020. IDRiD: Diabetic retinopathy—Segmentation and grading challenge. Medical Image Analysis 59 (Jan. 2020), 101561.
[171]
Gwenolé Quellec, Mathieu Lamard, Pierre Henri Conze, Pascale Massin, and Béatrice Cochener. 2020. Automatic detection of rare pathologies in fundus photographs using few-shot learning. Medical Image Analysis 61 (2020).
[172]
Lukas Radl, Yuan Jin, Antonio Pepe, Jianning Li, Christina Gsaxner, Fen hua Zhao, and Jan Egger. 2022. AVT: Multicenter aortic vessel tree CTA dataset collection with ground truth segmentation masks. Data in Brief 40 (2022), 107801.
[173]
Patrik F. Raudaschl, Paolo Zaffino, Gregory C. Sharp, Maria Francesca Spadea, Antong Chen, Benoit M. Dawant, Thomas Albrecht, Tobias Gass, Christoph Langguth, Marcel Luthi, Florian Jung, Oliver Knapp, Stefan Wesarg, Richard Mannion-Haworth, Mike Bowes, Annaliese Ashman, Gwenael Guillard, Alan Brett, Graham Vincent, Mauricio Orbes-Arteaga, David Cardenas-Pena, German Castellanos-Dominguez, Nava Aghdasi, Yangming Li, Angelique Berens, Kris Moe, Blake Hannaford, Rainer Schubert, and Karl D. Fritscher. 2017. Evaluation of segmentation methods on head and neck CT: Auto-segmentation challenge 2015. Medical Physics 44, 5 (May 2017), 2020–2036.
[174]
Joseph Redmon and Ali Farhadi. 2018. YOLOv3: An incremental improvement. arxiv:1804.02767 (2018). http://arxiv.org/abs/1804.02767
[175]
Mahdi Rezaei and Mahsa Shahidi. 2020. Zero-shot learning and its applications from autonomous vehicles to Covid-19 diagnosis: A review. arXiv:2004.14143 (2020).
[176]
Olaf Ronneberger, Philipp Fischer, and Thomas Brox. 2015. U-Net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015. Lecture Notes in Computer Science, Vol. 9351. Springer, 234–241.
[177]
Holger R. Roth, Le Lu, Amal Farag, Hoo Chang Shin, Jiamin Liu, Evrim B. Turkbey, and Ronald M. Summers. 2015. DeepOrgan: Multi-level deep convolutional networks for automated pancreas segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015. Lecture Notes in Computer Science, Vol. 9349. Springer, 556–564.
[178]
Holger R. Roth, Le Lu, Ari Seff, Kevin M. Cherry, Joanne Hoffman, Shijun Wang, Jiamin Liu, Evrim Turkbey, and Ronald M. Summers. 2014. A new 2.5D representation for lymph node detection using random sets of deep convolutional neural network observations. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015. Lecture Notes in Computer Science, Vol. 8673. Springer, 520–527.
[179]
Sylvia Rueda, Sana Fathima, Caroline L. Knight, Mohammad Yaqub, Aris T. Papageorghiou, Bahbibi Rahmatullah, Alessandro Foi, Matteo Maggioni, Antonietta Pepe, Jussi Tohka, Richard V. Stebbing, John E. McManigle, Anca Ciurte, Xavier Bresson, Meritxell Bach Cuadra, Changming Sun, Gennady V. Ponomarev, Mikhail S. Gelfand, Marat D. Kazanov, Ching Wei Wang, Hsiang Chou Chen, Chun Wei Peng, Chu Mei Hung, and J. Alison Noble. 2014. Evaluation and comparison of current fetal ultrasound image segmentation methods for biometric measurements: A grand challenge. IEEE Transactions on Medical Imaging 33, 4 (2014), 797–813.
[180]
Mirabela Rusu, Prabhakar Rajiah, Robert Gilkeson, Michael Yang, Christopher Donatelli, Rajat Thawani, Frank J. Jacono, Philip Linden, and Anant Madabhushi. 2017. Co-registration of pre-operative CT with ex vivo surgically excised ground glass nodules to define spatial extent of invasive adenocarcinoma on in vivo imaging: A proof-of-concept study. European Radiology 27, 10 (2017), 4209–4217.
[181]
David Scheie, Per Arne Andresen, Milada Cvancarova, Anne Signe Bø, Eirik Helseth, Kari Skullerud, and Klaus Beiske. 2006. Fluorescence in situ hybridization (FISH) on touch preparations: A reliable method for detecting loss of heterozygosity at 1p and 19q in oligodendroglial tumors. American Journal of Surgical Pathology 30, 7 (July 2006), 828–837.
[182]
K. M. Schmainda and M. Prah. 2018. Brain-Tumor-Progression. Retrieved August 23, 2023 from
[183]
Ari Seff, Le Lu, Adrian Barbu, Holger Roth, Hoo Chang Shin, and Ronald M. Summers. 2015. Leveraging mid-level semantic boundary cues for automated lymph node detection. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015. Lecture Notes in Computer Science, Vol. 9350. Springer, 53–61.
[184]
Ari Seff, Le Lu, Kevin M. Cherry, Holger R. Roth, Jiamin Liu, Shijun Wang, Joanne Hoffman, Evrim B. Turkbey, and Ronald M. Summers. 2014. 2D view aggregation for lymph node detection using a shallow hierarchy of linear classifiers. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2014. Lecture Notes in Computer Science, Vol. 8673. Springer, 544–552.
[185]
Arnaud Arindra Adiyoso Setio, Alberto Traverso, Thomas de Bel, Moira S. N. Berens, Cas van den Bogaard, Piergiorgio Cerello, Hao Chen, Qi Dou, Maria Evelina Fantacci, Bram Geurts, Robbert van der Gugten, Pheng Ann Heng, Bart Jansen, Michael M. J. de Kaste, Valentin Kotov, Jack Yu Hung Lin, Jeroen T. M. C. Manders, Alexander Sóñora-Mengana, Juan Carlos García-Naranjo, Evgenia Papavasileiou, Mathias Prokop, Marco Saletta, Cornelia M. Schaefer-Prokop, Ernst T. Scholten, Luuk Scholten, Miranda M. Snoeren, Ernesto Lopez Torres, Jef Vandemeulebroucke, Nicole Walasek, Guido C. A. Zuidhof, Bram van Ginneken, and Colin Jacobs. 2017. Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: The LUNA16 challenge. Medical Image Analysis 42 (Dec. 2017), 1–13.
[186]
Wei Shao, Liang Sun, and Daoqiang Zhang. 2018. Deep active learning for nucleus classification in pathology images. In Proceedings of the International Symposium on Biomedical Imaging. IEEE, Los Alamitos, CA, 199–202.
[187]
Micah J. Sheller, G. Anthony Reina, Brandon Edwards, Jason Martin, and Spyridon Bakas. 2019. Multi-institutional deep learning modeling without sharing patient data: A feasibility study on brain tumor segmentation. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries. Lecture Notes in Computer Science, Vol. 11383. Springer, 92–104.
[188]
Guohua Shen, Tomoyasu Horikawa, Kei Majima, and Yukiyasu Kamitani. 2019. Deep image reconstruction from human brain activity. PLoS Computational Biology 15, 1 (Jan. 2019), e1006633.
[189]
Zhao Shi, Chongchang Miao, U. Joseph Schoepf, Rock H. Savage, Danielle M. Dargis, Chengwei Pan, Xue Chai, Xiu Li Li, Shuang Xia, Xin Zhang, Yan Gu, Yonggang Zhang, Bin Hu, Wenda Xu, Changsheng Zhou, Song Luo, Hao Wang, Li Mao, Kongming Liang, Lili Wen, Longjiang Zhou, Yizhou Yu, Guang Ming Lu, and Long Jiang Zhang. 2020. A clinically applicable deep-learning model for detecting intracranial aneurysm in computed tomography angiography images. Nature Communications 11, 1 (Dec. 2020), 6090.
[190]
David Silver, Aja Huang, Chris J. Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, Sander Dieleman, Dominik Grewe, John Nham, Nal Kalchbrenner, Ilya Sutskever, Timothy Lillicrap, Madeleine Leach, Koray Kavukcuoglu, Thore Graepel, and Demis Hassabis. 2016. Mastering the game of Go with deep neural networks and tree search. Nature 529, 7587 (Jan. 2016), 484–489.
[191]
Karen Simonyan and Andrew Zisserman. 2015. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR’15): Conference Track Proceedings.
[192]
Amber L. Simpson, Michela Antonelli, Spyridon Bakas, Michel Bilello, Keyvan Farahani, Bram Van Ginneken, Annette Kopp-Schneider, Bennett A. Landman, Geert Litjens, Bjoern Menze, Olaf Ronneberger, Ronald M. Summers, Patrick Bilic, Patrick F. Christ, Richard K. G. Do, Marc Gollub, Jennifer Golia-Pernicka, Stephan H. Heckers, William R. Jarnagin, Maureen K. McHugo, Sandy Napel, Eugene Vorontsov, Lena Maier-Hein, and M. Jorge Cardoso. 2019. A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv:1902.09063 (2019). http://arxiv.org/abs/1902.09063
[193]
Roberto Souza, Oeslle Lucena, Julia Garrafa, David Gobbi, Marina Saluzzi, Simone Appenzeller, Letícia Rittner, Richard Frayne, and Roberto Lotufo. 2018. An open, multi-vendor, multi-field-strength brain MR dataset and analysis of publicly available skull stripping methods agreement. NeuroImage 170 (April 2018), 482–494.
[194]
Assaf B. Spanier and Leo Joskowicz. 2014. Rule-based ventral cavity multi-organ automatic segmentation in CT scans. In Medical Computer Vision: Algorithms for Big Data. Lecture Notes in Computer Science, Vol. 8848. Springer, 163–170.
[195]
Sergii Stirenko, Yuriy Kochura, Oleg Alienin, Oleksandr Rokovyi, Yuri Gordienko, Peng Gang, and Wei Zeng. 2018. Chest x-ray analysis of tuberculosis by deep learning with segmentation and augmentation. In Proceedings of the 2018 IEEE 38th International Conference on Electronics and Nanotechnology (ELNANO’18). 422–428.
[196]
Yue Sun, Kun Gao, Zhengwang Wu, Guannan Li, Xiaopeng Zong, Zhihao Lei, Ying Wei, Jun Ma, Xiaoping Yang, Xue Feng, Li Zhao, Trung Le Phan, Jitae Shin, Tao Zhong, Yu Zhang, Lequan Yu, Caizi Li, Ramesh Basnet, M. Omair Ahmad, M. N. S. Swamy, Wenao Ma, Qi Dou, Toan Duc Bui, Camilo Bermudez Noguera, Bennett Landman, Ian H. Gotlib, Kathryn L. Humphreys, Sarah Shultz, Longchuan Li, Sijie Niu, Weili Lin, Valerie Jewells, Dinggang Shen, Gang Li, and Li Wang. 2021. Multi-site infant brain segmentation algorithms: The iSeg-2019 challenge. IEEE Transactions on Medical Imaging 40, 5 (2021), 1363–1376.
[197]
Zaneta Swiderska-Chadaj, Hans Pinckaers, Mart van Rijthoven, Maschenka Balkenhol, Margarita Melnikova, Oscar Geessink, Quirine Manson, Mark Sherman, Antonio Polonia, Jeremy Parry, Mustapha Abubakar, Geert Litjens, Jeroen van der Laak, and Francesco Ciompi. 2019. Learning to detect lymphocytes in immunohistochemistry with deep learning. Medical Image Analysis 58 (2019), 101547.
[198]
Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. 2015. Going deeper with convolutions. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 1–9.
[199]
Ziqi Tang, Kangway V. Chuang, Charles DeCarli, Lee Way Jin, Laurel Beckett, Michael J. Keiser, and Brittany N. Dugger. 2019. Interpretable classification of Alzheimer’s disease pathologies with a convolutional neural network pipeline. Nature Communications 10, 1 (Dec. 2019), 2173.
[200]
Enzo Tartaglione, Carlo Alberto Barbano, Claudio Berzovini, Marco Calandri, and Marco Grangetto. 2020. Unveiling COVID-19 from chest x-ray with deep learning: A hurdles race with small data. International Journal of Environmental Research and Public Health 17, 18 (April 2020), 1–17.
[201]
Carlo Tessa, Nicola Toschi, Stefano Orsolini, Gaetano Valenza, Claudio Lucetti, Riccardo Barbieri, and Stefano Diciotti. 2019. Central modulation of parasympathetic outflow is impaired in de novo Parkinson’s disease patients. PLoS ONE 14, 1 (Jan. 2019), e0210324.
[202]
C. Tobon-Gomez, M. De Craene, K. McLeod, L. Tautz, W. Shi, A. Hennemuth, A. Prakosa, H. Wang, G. Carr-White, S. Kapetanakis, A. Lutz, V. Rasche, T. Schaeffter, C. Butakoff, O. Friman, T. Mansi, M. Sermesant, X. Zhuang, S. Ourselin, H. O. Peitgen, X. Pennec, R. Razavi, D. Rueckert, A. F. Frangi, and K. S. Rhode. 2013. Benchmarking framework for myocardial tracking and deformation algorithms: An open access database. Medical Image Analysis 17, 6 (Aug. 2013), 632–648.
[203]
Catalina Tobon-Gomez, Arjan J. Geers, Jochen Peters, Jürgen Weese, Karen Pinto, Rashed Karim, Mohammed Ammar, Abdelaziz Daoudi, Jan Margeta, Zulma Sandoval, Birgit Stender, Yefeng Zheng, Maria A. Zuluaga, Julian Betancur, Nicholas Ayache, Mohammed Amine Chikh, Jean Louis Dillenseger, B. Michael Kelm, Saïd Mahmoudi, Sébastien Ourselin, Alexander Schlaefer, Tobias Schaeffter, Reza Razavi, and Kawal S. Rhode. 2015. Benchmark for algorithms segmenting the left atrium from 3D CT and MRI datasets. IEEE Transactions on Medical Imaging 34, 7 (2015), 1460–1473.
[204]
Roger Trullo, Caroline Petitjean, Bernard Dubray, and Su Ruan. 2019. Multiorgan segmentation using distance-aware adversarial networks. Journal of Medical Imaging 6, 01 (2019), 1.
[205]
Vladimír Ulman, Martin Maška, Klas E. G. Magnusson, Olaf Ronneberger, Carsten Haubold, Nathalie Harder, Pavel Matula, Petr Matula, David Svoboda, Miroslav Radojevic, Ihor Smal, Karl Rohr, Joakim Jaldén, Helen M. Blau, Oleh Dzyubachyk, Boudewijn Lelieveldt, Pengdong Xiao, Yuexiang Li, Siu Yeung Cho, Alexandre C. Dufour, Jean Christophe Olivo-Marin, Constantino C. Reyes-Aldasoro, Jose A. Solis-Lemus, Robert Bensch, Thomas Brox, Johannes Stegmaier, Ralf Mikut, Steffen Wolf, Fred A. Hamprecht, Tiago Esteves, Pedro Quelhas, Ömer Demirel, Lars Malmström, Florian Jug, Pavel Tomancak, Erik Meijering, Arrate Muñoz-Barrutia, Michal Kozubek, and Carlos Ortiz-De-Solorzano. 2017. An objective comparison of cell-tracking algorithms. Nature Methods 14, 12 (Dec. 2017), 1141–1152.
[206]
M. Vallières, C. R. Freeman, S. R. Skamene, and I. El Naqa. 2015. A radiomics model from joint FDG-PET and MRI texture features for the prediction of lung metastases in soft-tissue sarcomas of the extremities. Physics in Medicine and Biology 60, 14 (July 2015), 5471–5496.
[207]
Martin Vallières, Emily Kay-Rivest, Léo Jean Perrin, Xavier Liem, Christophe Furstoss, Hugo J. W. L. Aerts, Nader Khaouam, Phuc Felix Nguyen-Tan, Chang Shu Wang, Khalil Sultanem, Jan Seuntjens, and Issam El Naqa. 2017. Radiomics strategies for risk assessment of tumour failure in head-and-neck cancer. Scientific Reports 7, 1 (Dec. 2017), 10117.
[208]
Thomas L. A. van den Heuvel, Dagmar de Bruijn, Chris L. de Korte, and Bram van Ginneken. 2018. Automated measurement of fetal head circumference using 2D ultrasound images. PLoS ONE 13, 8 (Aug. 2018), e0200412.
[209]
Bram van Ginneken, Samuel G. Armato, Bartjan de Hoop, Saskia van Amelsvoort-van de Vorst, Thomas Duindam, Meindert Niemeijer, Keelin Murphy, Arnold Schilham, Alessandra Retico, Maria Evelina Fantacci, Niccol Camarlinghi, Francesco Bagagli, Ilaria Gori, Takeshi Hara, Hiroshi Fujita, Gianfranco Gargano, Roberto Bellotti, Sabina Tangaro, Lourdes Bolaos, Francesco De Carlo, Piergiorgio Cerello, Sorin Cristian Cheran, Ernesto Lopez Torres, and Mathias Prokop. 2010. Comparing and combining algorithms for computer-aided detection of pulmonary nodules in computed tomography scans: The ANODE09 study. Medical Image Analysis 14, 6 (2010), 707–722.
[210]
Bram van Ginneken, Tobias Heimann, and Martin Styner. 2007. 3D segmentation in the clinic: A grand challenge. In Proceedings of the International Conference on Medical Image Computing and Computer Assisted Intervention.7–15. http://grand-challenge2008.bigr.nl/proceedings/pdfs/msls08/Styner.pdf
[211]
Rufin VanRullen and Leila Reddy. 2019. Reconstructing faces from fMRI patterns using deep generative neural networks. Communications Biology 2, 1 (Oct. 2019), 193.
[212]
Hadi Varmazyar, Zahra Ghareaghaji, and Saber Malekzadeh. 2020. MRI hippocampus segmentation using deep learning autoencoders.
[213]
Ruchika Verma, Neeraj Kumar, Abhijeet Patil, Nikhil Cherian Kurian, Swapnil Rane, and Amit Sethi. 2020. Multi-organ nuclei segmentation and classification challenge 2020. (Feb. 2020), 1–3.
[214]
Natalia Viani, Riley Botelle, Jack Kerwin, Lucia Yin, Rashmi Patel, Robert Stewart, and Sumithra Velupillai. 2021. A natural language processing approach for identifying temporal disease onset information from mental healthcare text. Scientific Reports 11, 1 (Dec. 2021), 757.
[215]
Oriol Vinyals, Igor Babuschkin, Wojciech M. Czarnecki, Michaël Mathieu, Andrew Dudzik, Junyoung Chung, David H. Choi, Richard Powell, Timo Ewalds, Petko Georgiev, Junhyuk Oh, Dan Horgan, Manuel Kroiss, Ivo Danihelka, Aja Huang, Laurent Sifre, Trevor Cai, John P. Agapiou, Max Jaderberg, Alexander S. Vezhnevets, Rémi Leblond, Tobias Pohlen, Valentin Dalibard, David Budden, Yury Sulsky, James Molloy, Tom L. Paine, Caglar Gulcehre, Ziyu Wang, Tobias Pfaff, Yuhuai Wu, Roman Ring, Dani Yogatama, Dario Wünsch, Katrina McKinney, Oliver Smith, Tom Schaul, Timothy Lillicrap, Koray Kavukcuoglu, Demis Hassabis, Chris Apps, and David Silver. 2019. Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature 575, 7782 (Nov. 2019), 350–354.
[216]
Bo Wang, Shuo Jin, Qingsen Yan, Haibo Xu, Chuan Luo, Lai Wei, Wei Zhao, Xuexue Hou, Wenshuo Ma, Zhengqing Xu, Zhuozhao Zheng, Wenbo Sun, Lan Lan, Wei Zhang, Xiangdong Mu, Chenxi Shi, Zhongxiao Wang, Jihae Lee, Zijian Jin, Minggui Lin, Hongbo Jin, Liang Zhang, Jun Guo, Benqi Zhao, Zhizhong Ren, Shuhao Wang, Wei Xu, Xinghuan Wang, Jianming Wang, Zheng You, and Jiahong Dong. 2021. AI-assisted CT imaging analysis for COVID-19 screening: Building and deploying a medical AI system. Applied Soft Computing 98 (Jan. 2021), 106897.
[217]
Guotai Wang, Wenqi Li, Sébastien Ourselin, and Tom Vercauteren. 2017. Automatic brain tumor segmentation using cascaded anisotropic convolutional neural networks. CoRR abs/1709.00382 (2017). http://arxiv.org/abs/1709.00382
[218]
Linda Wang, Zhong Qiu Lin, and Alexander Wong. 2020. COVID-Net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest x-ray images. Scientific Reports 10, 1 (Dec. 2020), 19549.
[219]
Li Wang, Dong Nie, Guannan Li, Élodie Puybareau, Jose Dolz, Qian Zhang, Fan Wang, Jing Xia, Zhengwang Wu, Jia Wei Chen, Kim Han Thung, Toan Duc Bui, Jitae Shin, Guodong Zeng, Guoyan Zheng, Vladimir S. Fonov, Andrew Doyle, Yongchao Xu, Pim Moeskops, Josien P. W. Pluim, Christian Desrosiers, Ismail Ben Ayed, Gerard Sanroma, Oualid M. Benkarim, Adrià Casamitjana, Verónica Vilaplana, Weili Lin, Gang Li, and Dinggang Shen. 2019. Benchmark on automatic six-month-old infant brain segmentation algorithms: The iSeg-2017 challenge. IEEE Transactions on Medical Imaging 38, 9 (Sept. 2019), 2219–2230.
[220]
Lucy Lu Wang, Kyle Lo, Yoganand Chandrasekhar, Russell Reas, Jiangjiang Yang, Darrin Eide, Kathryn Funk, Rodney Kinney, Ziyang Liu, William Merrill, Paul Mooney, Dewey Murdick, Devvret Rishi, Jerry Sheehan, Zhihong Shen, Brandon Stilson, Alex D. Wade, Kuansan Wang, Chris Wilhelm, Boya Xie, Douglas Raymond, Daniel S. Weld, Oren Etzioni, and Sebastian Kohlmeier. 2020. CORD-19: The COVID-19 open research dataset. arXiv:2004.10706 (2020). http://arxiv.org/abs/2004.10706
[221]
Xiaosong Wang, Yifan Peng, Le Lu, Zhiyong Lu, Mohammadhadi Bagheri, and Ronald M. Summers. 2017. ChestX-ray8: Hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR’17). 3462–3471.
[222]
Yaqing Wang, Quanming Yao, James T. Kwok, and Lionel M. Ni. 2020. Generalizing from a few examples: A survey on few-shot learning. ACM Computing Surveys 53, 3 (April 2020), Article 63, 34 pages.
[223]
Donglai Wei, Zudi Lin, Daniel Franco-Barranco, Nils Wendt, Xingyu Liu, Wenjie Yin, Xin Huang, Aarush Gupta, Won Dong Jang, Xueying Wang, Ignacio Arganda-Carreras, Jeff W. Lichtman, and Hanspeter Pfister. 2020. MitoEM dataset: Large-scale 3D mitochondria instance segmentation from EM images. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2020. Lecture Notes in Computer Science, Vol. 12265. Springer, 66–76.
[224]
Michael W. Weiner, Dallas P. Veitch, Paul S. Aisen, Laurel A. Beckett, Nigel J. Cairns, Robert C. Green, Danielle Harvey, Clifford R. Jack, William Jagust, John C. Morris, Ronald C. Petersen, Jennifer Salazar, Andrew J. Saykin, Leslie M. Shaw, Arthur W. Toga, and John Q. Trojanowski. 2017. The Alzheimer’s Disease NeuroImaging Initiative 3: Continued innovation for clinical trial improvement. Alzheimer’s and Dementia 13, 5 (May 2017), 561–571.
[225]
Zhaohan Xiong, Qing Xia, Zhiqiang Hu, Ning Huang, Cheng Bian, Yefeng Zheng, Sulaiman Vesal, Nishant Ravikumar, Andreas Maier, Xin Yang, Pheng-Ann Heng, Dong Ni, Caizi Li, Qianqian Tong, Weixin Si, Elodie Puybareau, Younes Khoudli, Thierry Géraud, Chen Chen, Wenjia Bai, Daniel Rueckert, Lingchao Xu, Xiahai Zhuang, Xinzhe Luo, Shuman Jia, Maxime Sermesant, Yashu Liu, Kuanquan Wang, Davide Borra, Alessandro Masci, Cristiana Corsi, Coen de Vente, Mitko Veta, Rashed Karim, Chandrakanth Jayachandran Preetha, Sandy Engelhardt, Menyun Qiao, Yuanyuan Wang, Qian Tao, Marta Nuñez-Garcia, Oscar Camara, Nicolo Savioli, Pablo Lamata, and Jichao Zhao. 2021. A global benchmark of algorithms for segmenting the left atrium from late gadolinium-enhanced cardiac magnetic resonance imaging. Medical Image Analysis 67 (Jan. 2021), 101832.
[226]
Jinzhong Yang, Harini Veeraraghavan, Samuel G. Armato, Keyvan Farahani, Justin S. Kirby, Jayashree Kalpathy-Kramer, Wouter van Elmpt, Andre Dekker, Xiao Han, Xue Feng, Paul Aljabar, Bruno Oliveira, Brent van der Heyden, Leonid Zamdborg, Dao Lam, Mark Gooding, and Gregory C. Sharp. 2018. Autosegmentation for thoracic radiation treatment planning: A grand challenge at AAPM 2017. Medical Physics 45, 10 (2018), 4568–4581.
[227]
Xiaofeng Yang, Ning Wu, Guanghui Cheng, Zhengyang Zhou, David S. Yu, Jonathan J. Beitler, Walter J. Curran, and Tian Liu. 2014. Automated segmentation of the parotid gland based on atlas registration and machine learning: A longitudinal MRI study in head-and-neck radiation therapy. International Journal of Radiation Oncology Biology Physics 90, 5 (Dec. 2014), 1225–1233.
[228]
Jason Yim, Reena Chopra, Terry Spitz, Jim Winkens, Annette Obika, Christopher Kelly, Harry Askham, Marko Lukic, Josef Huemer, Katrin Fasler, Gabriella Moraes, Clemens Meyer, Marc Wilson, Jonathan Dixon, Cian Hughes, Geraint Rees, Peng T. Khaw, Alan Karthikesalingam, Dominic King, Demis Hassabis, Mustafa Suleyman, Trevor Back, Joseph R. Ledsam, Pearse A. Keane, and Jeffrey De Fauw. 2020. Predicting conversion to wet age-related macular degeneration using deep learning. Nature Medicine 26, 6 (June 2020), 892–899.
[229]
Jure Zbontar, Florian Knoll, Anuroop Sriram, Tullie Murrell, Zhengnan Huang, Matthew J. Muckley, Aaron Defazio, Ruben Stern, Patricia Johnson, Mary Bruno, Marc Parente, Krzysztof J. Geras, Joe Katsnelson, Hersh Chandarana, Zizhao Zhang, Michal Drozdzal, Adriana Romero, Michael Rabbat, Pascal Vincent, Nafissa Yakubova, James Pinkerton, Duo Wang, Erich Owens, C. Lawrence Zitnick, Michael P. Recht, Daniel K. Sodickson, and Yvonne W. Lui. 2018. fastMRI: An open dataset and benchmarks for accelerated MRI. arXiv:1811.08839 (2018). http://arxiv.org/abs/1811.08839
[230]
Liang Zhang, Johann Li, Ping Li, Xiaoyuan Lu, Peiyi Shen, Guangming Zhu, Syed Afaq Shah, Mohammed Bennarmoun, Kun Qian, and Björn W. Schuller. 2020. MeDaS: An open-source platform as service to help break the walls between medicine and informatics. arXiv:2007.06013 (2020). http://arxiv.org/abs/2007.06013
[231]
Liang Zhang, Jiaming Zhang, Peiyi Shen, Guangming Zhu, Ping Li, Xiaoyuan Lu, Huan Zhang, Syed Afaq Shah, and Mohammed Bennamoun. 2020. Block level skip connections across cascaded V-Net for multi-organ segmentation. IEEE Transactions on Medical Imaging 39, 9 (2020), 2782–2793.
[232]
Qi Zhang, Yang Xiao, Wei Dai, Jingfeng Suo, Congzhi Wang, Jun Shi, and Hairong Zheng. 2016. Deep learning based classification of breast tumors with shear-wave elastography. Ultrasonics 72 (2016), 150–157.
[233]
S. Zhang, W. Yoshida, H. Mano, T. Yanagisawa, K. Shibata, M. Kawato, and B. Seymour. 2020. Cognitive control of sensory pain encoding in the pregenual anterior cingulate cortex. D1 - decoder construction in day 1, d2 - adaptive control in day 2. OpenNeuro. Retrieved August 23, 2023 from
[234]
Binsheng Zhao, Leonard P. James, Chaya S. Moskowitz, Pingzhen Guo, Michelle S. Ginsberg, Robert A. Lefkowitz, Yilin Qin, Gregory J. Riely, Mark G. Kris, and Lawrence H. Schwartz. 2009. Evaluating variability in tumor measurements from same-day repeat CT scans of patients with non-small cell lung cancer. Radiology 252, 1 (July 2009), 263–272.
[235]
Jinyu Zhao, Xuehai He, Xingyi Yang, Yichen Zhang, Shanghang Zhang, and Pengtao Xie. 2020. COVID-CT-Dataset: A CT image dataset about COVID-19. arXiv:2003.13865 (2020). http://arxiv.org/abs/2003.13865
[236]
Fuzhen Zhuang, Zhiyuan Qi, Keyu Duan, Dongbo Xi, Yongchun Zhu, Hengshu Zhu, Hui Xiong, and Qing He. 2020. A comprehensive survey on transfer learning. arxiv:1911.02685[cs.LG] (2020).
[237]
Xiahai Zhuang. 2019. Multivariate mixture model for myocardial segmentation combining multi-source images. IEEE Transactions on Pattern Analysis and Machine Intelligence 41, 12 (2019), 2933–2946.

Cited By

View all
  • (2024)Effect of Training Data Differences on Accuracy in MR Image Generation Using Pix2pixPix2pixを用いたMR画像生成における学習データの違いが精度に及ぼす影響Japanese Journal of Radiological Technology10.6009/jjrt.2024-1487Online publication date: 2024
  • (2024)Classification of Multi-view mammogram images using a parallel pre-trained models system2024 8th International Conference on Image and Signal Processing and their Applications (ISPA)10.1109/ISPA59904.2024.10536707(1-6)Online publication date: 21-Apr-2024
  • (2024)Bias Inheritance and its Amplification in GAN-Based Synthetic Data Augmentation for Skin Lesion Classification2024 3rd International Conference on Artificial Intelligence For Internet of Things (AIIoT)10.1109/AIIoT58432.2024.10574714(1-6)Online publication date: 3-May-2024
  • Show More Cited By

Index Terms

  1. A Systematic Collection of Medical Image Datasets for Deep Learning

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Computing Surveys
    ACM Computing Surveys  Volume 56, Issue 5
    May 2024
    1019 pages
    EISSN:1557-7341
    DOI:10.1145/3613598
    Issue’s Table of Contents
    This work is licensed under a Creative Commons Attribution-Share Alike International 4.0 License.

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 27 November 2023
    Online AM: 16 August 2023
    Accepted: 03 August 2023
    Revised: 04 May 2023
    Received: 08 August 2022
    Published in CSUR Volume 56, Issue 5

    Check for updates

    Author Tags

    1. Medical image analysis
    2. deep learning
    3. datasets
    4. challenges
    5. computer-aided diagnosis

    Qualifiers

    • Survey

    Funding Sources

    • National Natural Science Foundation of China
    • National Key R&D Program of China

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)8,282
    • Downloads (Last 6 weeks)1,011
    Reflects downloads up to 01 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Effect of Training Data Differences on Accuracy in MR Image Generation Using Pix2pixPix2pixを用いたMR画像生成における学習データの違いが精度に及ぼす影響Japanese Journal of Radiological Technology10.6009/jjrt.2024-1487Online publication date: 2024
    • (2024)Classification of Multi-view mammogram images using a parallel pre-trained models system2024 8th International Conference on Image and Signal Processing and their Applications (ISPA)10.1109/ISPA59904.2024.10536707(1-6)Online publication date: 21-Apr-2024
    • (2024)Bias Inheritance and its Amplification in GAN-Based Synthetic Data Augmentation for Skin Lesion Classification2024 3rd International Conference on Artificial Intelligence For Internet of Things (AIIoT)10.1109/AIIoT58432.2024.10574714(1-6)Online publication date: 3-May-2024
    • (2024)A systematic review of major evaluation metrics for simulator-based automatic assessment of driving after strokeHeliyon10.1016/j.heliyon.2024.e3293010:12(e32930)Online publication date: Jun-2024
    • (2024)HCT: a hybrid CNN and transformer network for hyperspectral image super-resolutionMultimedia Systems10.1007/s00530-024-01387-930:4Online publication date: 20-Jun-2024
    • (2024)Cross‐Species Segmentation of Animal Prostate Using a Human Prostate Dataset and Limited Preoperative Animal Images: A Sampled Experiment on Dog Prostate TissueInternational Journal of Imaging Systems and Technology10.1002/ima.2313834:4Online publication date: 10-Jul-2024
    • (2023)Additional look into GAN-based augmentation for deep learning COVID-19 image classificationMachine Graphics and Vision10.22630/MGV.2023.32.3.632:3/4(107-124)Online publication date: 18-Dec-2023

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Get Access

    Login options

    Full Access

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media