Singha 2021
Singha 2021
net/publication/353702181
CITATIONS READS
20 1,761
3 authors:
Rini Thakur
Indian Institute of Science Education and Research Bho…
10 PUBLICATIONS 419 CITATIONS
SEE PROFILE
All content following this page was uploaded by Rini Thakur on 18 January 2022.
Abstract
Deep learning is slowly taking over the medical image analysis field with advance-
ments in imaging tools, and growing demand for fast, accurate, and automated image
analysis. In this chapter we have extensively reviewed the field from inception to its
current state-of-the-art techniques. We begin by introducing imaging advancements
over the decades, followed by general pre-analysis processes. Further, we describe the
different kinds of deep learning models used for image analysis, after broadly clas-
sifying them into supervised and unsupervised learning. Here the chapter touches
upon the architecture and complex functionality of each layer of such a deep neu-
ral network. CNN being one of the most popular tools for image analysis is then
reviewed in-depth. We have discussed its architecture and learning along with the
mathematical framework behind it, followed by real-world applications. The second
half of this chapter is devoted to the biological applications of deep learning. We start
by reviewing advancements from sub-cellular to cellular to the tissue level, ending
with an extensive review at the organ level. We have included various deep learning
tools and techniques that people have attempted to better understand the morphol-
ogy and physiology of the human body, and its related diseases. Finally, the chapter
ends with current field-specific challenges and the ways forward to eliminate them.
Keywords: Deep learning, medical image analysis, convolutional neural
networks (CNNs), image classification, image segmentation, medical diagnosis
Sujata Dash, Subhendu Kumar Pani, S. Balamurugan and Ajith Abraham (eds.) Biomedical Data
Mining for Information Retrieval: Methodologies, Techniques and Applications, (293–350) © 2021
Scrivener Publishing LLC
293
294 Biomedical Data Mining for Information Retrieval
11.1 Introduction
Image analysis is a complex process that a human brain performs typi-
cally, at least a million times a day. The brain is a versatile machine but has
limits of speed, accuracy, and exhaustion. These natural machines’ limita-
tions compelled humans to develop tools like artificial intelligence (AI) in
the last century. Ever since this field was born, it has seen a multitude of
advancements due to ever-improving computational technology and high-
end storage with fast transmission capacity over distances. Following the
quest for training machines in an intelligent manner to let them learn from
data, introduced us to machine learning. Powered by self-evolving design,
machine learning finds excellent applications in solving many practical
problems for humans like remote sensing, agriculture, handwriting and
face recognition, disease diagnosis, traffic prediction, and many more.
With the increasing complexity of issues that we are dealing with, machine
learning is now further improved into deep learning. It is a more advanced
version of the earlier, involving more layers of complex algorithms.
Furthermore, this finds application in a yet more complex field of natural
architecture, the medicine. Here, deep learning deals with analyzing biolog-
ical systems’ images ranging from subcellular structures to irregular-sized
tumors using neural networks. The various image pre-processing steps like
image classification, registration, segmentation, denoising, decomposition,
are required to extract relevant information from the medical images. We
are going to discuss all these in great detail across this chapter.
This chapter is about deep learning and its recent advancements to aid
medical image analysis. We will discuss the variations that have evolved
across different techniques of deep learning according to challenges in spe-
cific fields, and emphasize one such extensively used tool, Convolutional
Neural Network (CNN), in medical image analysis. We would talk about
the architecture of CNN and its training and how it functions. We would
shed some light on the ways it is used for image denoising and classifi-
cation concerning medical image analysis. In the next section, we would
report various advancements in the medical image analysis field using deep
learning applications, in a hierarchical biological classification manner. We
will start by targeting small structures like organelles to bigger ones like
tissues and organs, providing a comprehensive overview. We have specially
designed this content for medical practitioners to provide them an exclusive
view of all significant techniques that have evolved in the field. This includes
a descriptive and tabular mention of level-wise novel techniques that aid in
disease recognition and treatment in earlier stages. We will include some
Deep Learning in Medical Image Analysis 295
time element in 3D; 4D imaging is gaining its popularity and, when com-
bined with methods CT, MRI, PET, and SPECT, it can be used in studying
fluid dynamics of internal human organs and generate structured mesh
models of them [5, 6]. In Table 11.1, advancements in the field of medical
imaging are given [1, 2, 8–11].
As shown above, the advancements in medical imaging modalities over
time led to significant data acquisition and an increased load on medi-
cal experts to analyze them, which made them prone to more errors. This
makes them prone to more errors. This thus led to the era of AI (Artificial
Intelligence), for relatively more straightforward tasks like ‘classification’ as
back as in 1995 [7]. Later machine learning, and then more advanced deep
learning tools emerged for denoising and other various pre-processing
steps before actual image analysis. Let’s first find our way to understand
artificial intelligence and deep learning in some detail.
approach to output with maximum reward. This method is widely used for
cognitive research and problem-solving applications. The parallel network
of ANN can be built with multiple processors, providing it with a high-
speed advantage at a little development cost. ANN is extensively applied
for image recognition and speech recognition.
Deep Learning is a highly rising domain in the Artificial Intelligence (AI)
world. Networks are capable of learning in an unsupervised manner from
unstructured or unlabeled data. It is also known as deep neural networks or
deep AI, occasionally. Deep Learning or deep AI is a humongous collection of
such artificial neural networks (ANNs). With ANN’s potential, we can auto-
matically extract significant low-level features and amalgamate them into
great defining features. These advances of ANN or deep learning have opened
up applications in the domain of healthcare and medicine. The intertwining
of the pre-existing medical imaging techniques with Deep Learning has given
some remarkable results that we are going to discuss across this chapter.
Deep Learning is broadly categorized into supervised and unsupervised
learning, based on the kind of data used for training. They have different
structures and working principles. The convolution neural network (CNN)
is a complex feed-forward network under supervised deep learning that is
extensively applied for image recognition and analysis [14]. Let us start by
first understanding processing in medical images.
Figure 11.1 Image depicting prostate cancer cell segmentation (Source—Alan Partion
John Hopkins University).
a prostate cancer cell. This image contains two different spatially oriented
cells, but only one of them is distinguishable. The ultimate motive is either
to detect or to segment, the evident cell.
Image Classification is the process of assigning image pixels into different
categories or classes according to application. The input images are subdi-
vided into classes based on features. The main difference between machine
learning (ML) and deep learning (DL) algorithms is the feature extraction
step. The traditional machine learning approach requires hand-crafted
features that are time-consuming, difficult, and require domain expertise,
whereas deep learning automatically extracts features by algorithms. Image
classification assessment metrics are specificity, sensitivity, and accuracy
[17]. Some examples of classification in medical imaging are classification
of a skin lesion for malignancy using Alex-Net CNN [18], classification of
hematoxylin and eosin-stained (H&E) breast biopsy images using CNNs
[19]. Some others can be to classify optical coherence tomography images
into age-related macular degeneration and diabetic macular edema. This
uses the Inception V3 network based on transfer learning [20], multi-label
chest X-ray classification using ResNet architecture [21].
Image Enhancement is generally performed to improve the overall visual
quality of medical images for an accurate diagnosis. MR images are cor-
rupted by rician noise, salt and pepper noise, Gaussian noise, and speckle
noise. Similarly, CT images are mostly corrupted by Gaussian noise, ran-
dom noise, electronic noise, statistical noise due to the statistical uncer-
tainty of physical measurements involved. This includes radiation dose,
software, and hardware-based artifacts, patient motion [22]. Ultrasound
images are degraded by the presence of ‘Speckle noise’, a signal-dependent
multiplicative disturbance [23]. It reduces the resolving power of image
and contrast that affects diagnostic value. PET images exhibit a higher
noise level as compared to MRI and CT scans, as it is corrupted by a mix-
ture of Gaussian and Poisson noise [24]. The image denoising techniques
302 Biomedical Data Mining for Information Retrieval
Hidden layer
Input layer
Output layer
s j (t ) = f (∑ xi (t )u ji )
y k (t ) = g (∑ si (t )v kj )
304 Biomedical Data Mining for Information Retrieval
Where f(z) is sigmoid activation function and g(z) is softmax function respectively,
1
f (z ) =
1 − e−z
e zm
g (z ) =
∑e k
zk
For initialization, s(0) vectors can be given any small value because
while processing enormous data, initialization value does not matter. The
size of the vector x depends on the size of input data and the size of the hid-
den layer s(t) is usually around 50–500. This hidden layer s(t) depends on
data size. The networks are trained in epochs where data are represented
from training corpus sequentially.
Weights are initialized to random Gaussian noise with small mean and
variance, and learning rate α is also initialized. The training is carried out
using a standardized backpropagation algorithm along with stochastic
gradient descent is used for training. If the log-likelihood of the data is
increased, training is carried out in a new epoch else learning rate is halved.
Within 10–20 epochs, the network is trained. The softmax equation makes
sure that the distribution is valid. With each training, step weight is calcu-
lated, and weights are updated using a backpropagation algorithm. In the
end, the optimization algorithm is applied depending upon the data to get
the desired output.
hn = f (w1xn + b1)
Where,
w1: Weight matrix used for encoder
f: Encoder function
b1: Biased vector for encoder
yn = g (w2hn + b2)
Where,
w2: Weight matrix used for decoder
g: Decoder function
b2: Biased vector for decoder
1
∑
N
Φ (θ) = θ, θ′arg min L( x i , y i )
n n =1
Dropout and ReLU are used to overcome the overfitting problem, which
is present when the data fed is small. This way, accuracy is maintained, and
errors are minimized.
Where,
P (hl |hl+1): Conditional probability distribution for the units of
layer l given the layer l + 1,
P (hL−1, hL): Joint probability distribution of the units in layers
L − 1 and L.
(2) Input for the next hidden layer is obtained from the previous
layer with the mean activation of P(h(1) = 1|h(0)) or samples
drawn from P(h(1)|h(0)) which is used as observation for
the second hidden layer.
(3) The second layer is also trained as an RBM; taking mean
activation or sample from the trained example.
(4) Repeat steps 2 and 3 until all the layers are trained until the
required data is met.
Where,
θ = {W1, W2} are the model parameters representing visible to visible or
hidden to hidden interference.
The probability that the model is allocated to a visible vector s is:
308 Biomedical Data Mining for Information Retrieval
P(s ;θ ) =
1
Z (θ ) ∑exp(E(s, h , h ;θ ))
h1 ,h2
1 2
P ( h1j = 1| s , h 2 ) = σ
∑W s + ∑W
i
1
ij i
m
h
2 2
jm j
P ( hm2 = 1| h1 ) = σ
∑W h
i
2 1
im i
P ( si = 1| h1 ) = σ
∑W h
i
1
ij j
Generated Fake
Data
Noise
Real Real
Data
input layer
hidden layers and discriminator
output layer generator
(d) (e)
Figure 11.3 Various neural network architectures. (a) Recurrent Neural network, (b)
Autoencoder, (c) Restricted Boltzmann Machine, (d) Deep Belief Network, (e) Generative
Adversarial Network. x, y, h represent input, output and hidden layers (Inspired by: Ker
et al. [40]).
nervous system attributes. The idea of the CNN receptive field is taken from
this. The ‘Neocognitron model’ was designed by Kunihiko Fukushima in
1980, and was motivated by Hubel and Weisel’s research. Pattern recogni-
tion was done in the neocognitron model with the help of convolution and
down-sampling layers [42]. The successful implementation of LeNet-5 by
LeCuN, which was a handwritten digit classification network, laid the foun-
dation of the basic CNN model, which is being modified and used for med-
ical image classification [14]. After that, many pioneer CNN classification
networks were designed in the ImageNet Large Scale Visual Recognition
Challenge (ILSVRC), a software design competition for an object detection
run by ImageNet annually. In 2012, the AlexNet network outperformed all
the other image classification methods and garnered the research commu-
nity’s considerable attention in the design of deep CNN’s. The other famous
image classification CNN’s are VGGNet, GoogleNet, ResNet, ZFNet [43].
Compared to other deep learning neural networks, there are several
advantages of CNN, specifically when input is image data. The hypothesis of
weight sharing between neurons in CNN leads to a reduction of the network
complexity and improves generalization as compared to the artificial neural
network. CNN’s can be trained easily by backpropagation as sparse connec-
tions are present in each convolution layer. The application-oriented size-
able medical image databases are available, which can be used for training in
graphics processing units (GPUs) for efficient and faster computation. CNN
possesses higher representation capability, which provides better image reso-
lution performance compared to traditional sparse representation. In sparse
representation, 2D structural information, i.e., information pertaining to
the pixel dependency in the local adjoining neighborhood region, is lost.
The sparse dictionaries are formed with vectorization of the image matrices
leading to two-dimensional structural information loss. On the other hand,
CNN is efficient in retaining two-dimensional structural information during
the training and testing process. The convolution operator with 2D masks
inherently includes the local neighboring image pixels. Transfer learning is
applied to CNN’s in which pre-trained networks are modified by changing
input dataset, no. of layers, modification in the loss function to design a new
network which inherits many features of the pre-trained network.
X ji = f
∑ iMj
xil −1 ∗ w lji + blj
(11.1)
Where Mj represents the selection of the input feature map, xil −1 is the output of
the previous layer i.e. feature map, w ijl the weight of the convolution kernel of the
lth layer, f is the activation function and, blj is the bias of the lth layer.
The typical CNN architecture for training a 256 × 256 input image
with a supervised learning objective function approach is given in Figure
11.4. The objective function or loss function is used for optimization for
training CNN for a particular image processing application. The different
operations taking place in CNN architecture are convolution, pooling, and
non-linear transformation using the activation function.
Convolution:
The convolution filter consists of weight kernels that detect a particular fea-
ture of the image. The convolution filter of fixed size is translated all over
the image, and output is matrix multiplication of convolution filter weights
Desired Output
Image
256×256 Obj Func. c(.)=
(RGB)
Figure 11.4 An illustration of a typical CNN architecture for pixel RGB images (includes
the training objective function).
312 Biomedical Data Mining for Information Retrieval
3 0 1 5 0 3 0 3
2 6 2 4 3 0 3 0 -3
–1 0 1
2 4 1 0 6 1 4 1
–2 0 2
3 0 1 5 0 3 0 2
2 6 2 4 3 2 3 0 –1 0 1
2 4 1 0 6 2 1 1
(b)
2 6 2 4 4 0 3 1
2 4 1 0 6 1 6 6
(a) (c)
Figure 11.5 (a) Input Image, (b) Convolutional filter and (c) Convolved output value for
the highlighted pixel.
with the corresponding part of the input image matrix. The convolution
filter slides over the entire input image as shown in Figure 11.5. The num-
ber of pixels in which convolution filter slides horizontally or vertically to
perform convolution over the whole image is called a stride. The convolu-
tion filter’s output is termed as a feature map because it extracts particular
image features like edges and artifacts. Sometimes, zero paddings are also
performed, i.e., image size increases by padding zeros around the image to
get the desired size feature map.
Pooling:
The pooling operator is applied to a feature map for spatial resolution reduc-
tion. The feature map becomes spatial invariant to input data distortion and
translation when its spatial resolution is reduced. Initially, average pooling
was used in CNN models, i.e., the average of all input pixel values of the local
neighborhood were propagated to the subsequent layer. In the recent mod-
els, maximum pooling operation is done, which propagates maximum value
within an image neighborhood to the subsequent layer. The max-pooling
operator selects the most considerable pixel value in the given receptive
field. Figure 11.6 depicts max-pooling and average pooling process on the
input image of size 4 × 4, convolution filter of size 2 × 2 with a stride of two.
The max-pooling selects the maximum pixel value in 2 × 2 sub-window,
whereas average pooling calculates each sub-sampled 2 × 2 window.
Activation Function:
The activation function transforms convoluted output by performing a
non-linear transformation. The predominantly used activation functions are
rectified linear units, tan hyperbolic function, softmax function, sigmoid.
Deep Learning in Medical Image Analysis 313
Average
15 9 11 5 10 5
11 5 8 6 5 8
8 7 3 17
2 3 7 3 Max 15 8
8 17
1
∑ ∑
P M
E(k ) = ( zij − tij ) k = 1 to Sk .
2
(11.2)
P×M j =1 i =1
Where M and P are the number of rows and columns in image, Sk is the
number of training iterations, zpm and tpm represents an actual and
desired image.
The stochastic gradient descent method is utilized in optimization, i.e.
minimization of this mean square error. The gradient descent method iter-
atively modifies the weight (w) from w(t) to w(t + 1) with the weight upda-
tion equation given as follows:
∂ E( k )
w(t + 1) = w(t ) − η (11.3)
∂ w(t )
314 Biomedical Data Mining for Information Retrieval
∑
P
Rlearn = Rθ arg arg min C ( x j , Rθ ( y j )) + g (θ) (11.4)
j =1
Softmax True
Function Label
Prediction
spatial variability, as the function of intensity and time for a given probe.
The interconnected layers of algorithm scan for logical patterns invariance
and compute the subcellular spread [57]. These scans locate any mislocal-
ized protein among cellular compartments, which can trigger the onset
of diseases such as Alzheimer’s or cancer or some cardiovascular disor-
ders [58]. One of the earliest studies by Boland and Murphy [59] used a
back-propagation neural network (BPNN) on the HeLa cell lines for the
quantitative estimation of protein localization patterns. In 2008, Li et al.
[60] adopted the first automated approach for subcellular delocalization
of proteins using Human Protein Atlas (HPAs) reference, which analyzed
over 3,000 images with 87.5% accuracy. Tahir et al. [61], used a hybrid
SVM-SubLoc method to achieve more than 99% correctness over the data-
set. Like the earlier one, numerous tools and methods have been reported
in the last decade, with ever-improving accuracy and applicability. One
recent work (Yang et al. [62]) uses the MIC_Locator model with the lat-
est, multi-labeled dataset, and frequency features to outperform others
in the field. Many such works have also used yeast as a model organism
(Pärnamaa and Parts [63]; Kraus et al. [64]).
Another application of deep learning that uses HTS devices is drug dis-
covery. Mapping and assessing interactions between drug molecules and
their target sites form an essential component in the drug discovery pro-
cess, called ligand–protein interactions. Many studies have examined sub-
cellular perturbations in response to a specific drug and compared diverse
drug varieties for optimum response [65, 66]. An interesting approach is
to use unsupervised machine learning or supervised (ML) to screen for
phenotypic changes at cellular/subcellular level in response to a particular
stimulus like a drug molecule, thus helping in extensive testing before clin-
ical use [67]. A 2018 study used pre-trained deep CNN networks to map
cellular action mechanisms in response to a drug molecule with over 95%
accuracy [68].
Karyotyping also effectively uses neural networks - A process of describ-
ing shape and number of chromosomes in a cell. Chromosomal abnormal-
ities arise either due to altered numbers or/and structure of chromosomes
and can develop in fatal genetic diseases like leukemia or even cancer. A
very recent study proposes the use of a two-stage chromosomal classifica-
tion method in which a Q-banded dataset with 119 chromosomal images
and ANN-MLP classifier is used to obtain 88% classification accuracy [69].
Before this, many such efforts in the field involving segmentation, feature
extraction, and classification of strained chromosomal images were made.
These automated techniques can aid cytogeneticists with high accuracy
and less effort.
Deep Learning in Medical Image Analysis 319
and out of focus images are headaches for medical examiners, especially
when dealing with incredible large datasets. To fix this quality control or
technically, ‘Pre-preprocessing’ is done, which is imperative to compensate
for the unevenness of data and aid in further analysis. Autoencoders are
one great example of this, which are extensively used for differentiating
noise from accurate data. One of the pioneering work was done by Su et al.
[75] using stacked denoising autoencoders (sDAE) to detect shape vari-
ations and overlapping images of cell fragments and also perform effec-
tive cellular segmentation. Following this [76], a U-Net architecture was
presented to improve the high-quality fixation of 3D isotropic resolution
in fluorescence microscopy. Wang et al. [137] went a step ahead to scale
up the resolution of full-field fluorescence 10×/0.4NA microscopic images
to 20×/0.75NA objective. This employed GANs (Generative Adversarial
Networks) for the purpose. Autoencoders were used by Janowczyk et al.
[77] for strain normalization to standardize the color distribution of the
test images that were to be processed by computer-aided diagnostic (CAD)
algorithms. CNN also acts as a savior in quality control. Yang et al. [78]
used deep neural networks for classifying blurred microscopic images with
high accuracy and precision. They knowingly trained their model on 384
defocused stained nuclei images of U2OS cells to make the system more
robust, achieving over 95% accuracy.
Classification frameworks have been effectively used in the diagnosis
of various autoimmune diseases. Gao et al. [79] used an automated CNN
based approach for classifying HEp-2 (Human Epithelial-2) cells. HEp-2
cells act as a substrate for antinuclear antibody (ANA) screening. It is a
good indication of developing autoimmune diseases like scleroderma,
autoimmune hepatitis. Indirect immunofluorescence (IIF) is the diagnos-
tic tool used to classify cells based on intensity and type of fluorescence in
the staining pattern. Ref. [80] is the state-of-the-art technology that uses
ten optimum layered ANN without data augmentation for over 96% accu-
racy. Sickle cell anemia also uses classification efficiently for its diagno-
sis. Ref. [81] is the current best model that classifies erythrocytes, among
other blood cell types, into a circular (healthy) and elongated (sickle cells).
This model achieves over 99% accuracy in the classification process using
DL algorithms. Many other erythrocyte classifications are often done to
study the morphology profile of red blood cells (RBCs). Models generally
use deep CNN to classify RBCs into different morphological classes. Since
RBCs carry oxygen throughout the body, the study of their morphology
serves as an indication of numerous potential diseases. CNN is used to
detect and classify inflammatory cells such as neutrophils and lymphocytes
with Hematoxylin and Eosin (H&E). High numbers of inflammatory cells
Deep Learning in Medical Image Analysis 321
have already attempted to build 3D cell models. While Chen et al. [94] have
developed a 5-D (x, y, z, t, λ) cell imaging model on dendritic cells (DC)
and T-cells of evolving mouse thymic cortex and lymph nodes. Model used
a two-photon laser scanning microscopy (TPLSM) technique for this. The
future improvements can be to fabricate these models more robustly by
closely simulating effects of the environment such as drugs or pathogens as
an example, for better applicability in medical diagnosis. Table 11.2 describes
some more DL-based applications at the subcellular and cellular level.
model on 700+ ear images. Furthermore, they tested it on 150+ ear images
to characterize the condition of the tympanic membrane and ear canal
with an above 90% classification accuracy and average sensitivity of 87.8%.
As a new crude field, much research is still needed with more in-depth
insights for its better development.
Table 11.3 Some miscellaneous advancements in the organ section and multi-
domain applicability tool (Inspired from: Litjens et al. [183]).
Source Modality Applications
Dermatology
Yu et al. [176] CNN Lesion segmentation and classification
using Deep residual networks
Esteva et al. [177] CNN Inception network trained 100k+ images
Haenssle et al. CNN Compares CNN performance with
[178] an international group of 50+
dermatologists, Google’s Inception v4
CNN architecture used
Fujisawa et al. CNN Efficient skin tumor classifier, accuracy
[179] 76.5%, sensitivity 96.3%
Brinker et al. [180] CNN Clinical melanoma classification, Par with
dermatologists without clinical training
Muscular-skeletal
Golan et al. [181] CNN CNN (with adversarial component) detects
Hip dysplasia structures and performs
measurements
Jamaludin et al. CNN VGG-19 CNN analyzes vertebral discs;
[182] finds lesion hot spots
Lee et al. [184] CNN Decision supporting system for more
accurate and efficient bone age
assessments (BAAs), >90% accuracy and
<2 s interpretation time
Chaudhari et al. CNN Thin-slice knee MR images from thicker
[186] input slices, 3D CNN, State-of-the-art
Liu et al. [187] CNN Anterior cruciate ligament (ACL) tear
detection within the knee joint, 175
subjects
Li [188] CNN Abnormal sites in radiographic images,
Evaluating musculoskeletal quality
(Continued)
334 Biomedical Data Mining for Information Retrieval
Table 11.3 Some miscellaneous advancements in the organ section and multi-
domain applicability tool (Inspired from: Litjens et al. [183]). (Continued)
Source Modality Applications
Zheng et al. [189] CNN Assessment of children pediatric leg length
discrepancy (LLD), 179 subjects, 96
times faster than humans.
Others
Jaumard-Hakoun RBM Tongue mechanics analysis during the
et al. [190] speech, auto-encoders with RBMs
combination is used
Quinn et al. [191] CNN Smartphone detection of malaria,
tuberculosis & parasite eggs
Lekadir et al. [192] CNN Ultrasound characterization of carotid
plaque composition
Rad et al. [193] CNN Segmentation TE in human blastocyst
image, Synthetic embryo images for 1st
time
Nardelli et al. CNN Convolutional Neural Regressor for airway/
[194] vessel morphology study, Chest CT
Multi-domain deep learning applications
Ghesu et al. [195] CNN Reinforcement learning of head & neck CT
and cardiac MR/US.
Moeskops et al. CNN Segments brain tissues, pectoral muscle &
[196] coronaries using unified network
Roth et al. [197] CNN Multi-stream detection of lymph nodes,
sclerotic lesions, and polyps
Tu et al. [185] CNN Human region-based multi-stream
(HR-MSCNN) model; Full-frame action
recognition
The NIH database contains 100,000 chest X-ray images. The Brainweb is the
most popular simulated and phantom database of MR images available for
public use. Researchers widely use the ‘Mini-mias database’ of mammograms
for breast cancer detection. Some datasets of computer-aided diagnosis com-
petitions organized by Microsoft, Kaggle are also available for the public.
Deep Learning in Medical Image Analysis 335
This chapter should be seen as one such effort to bridge that gap by
reporting advancements in DL tools and techniques across the medical
image analysis field, in recent years. We have discussed how deep learn-
ing algorithms revolutionized the medical field. We briefly described the
architecture and functioning of different supervised and unsupervised DL
models. We also discussed CNN in some good detail, owing to its popu-
larity and applicability in the field. The advancement section in the chapter
provides a comprehensive and exclusive overview of the reported develop-
ments in medical image analysis in a more systematized biological manner.
The unified aim of it is to motivate the readers, especially medical stake-
holders, to adopt these AI-based networks more effectively in diagnosis.
The challenges listed above can urge one to question the completeness of
DL in the analysis field, but we need to remember that these challenges can
be a matter of the past. This field has exploded in the last decade, and the
pace at which it is improving is just incomparable.
According to statistics, the misdiagnosis rate caused by humans in med-
ical image analysis can reach 10–30% [198]. The sole purpose of DL or
AI-based automated medical diagnosis is to reduce that error percentage
rate. Deep learning or any form of AI can never acquire a human’s place
in medical diagnosis or any other related field since the field is still evolv-
ing and is prone to multiple errors. Furthermore, this evolution itself is a
relative limit which is derived by human’s everlasting ambitions. Thus this
indicates that these AI tools should not be seen as a competitor to medical
doctors or researchers. Instead, they should be better utilized as an aid to
improve accuracy and time efficiency with less tedious procedures to follow.
This can create a new realm of more efficient diagnoses in the future.
References
1. Bradley, W.G., History of Medical Imaging. Proc. Am. Philos. Soc., 152, 3,
349–361, 2008.
2. Willmann, J.K., van Bruggen, N., Dinkelborg, L.M., Gambhir, S.S., Molecular
imaging in drug development. Nat. Rev. Drug Discovery, 7, 7, 591–607, Jul.
2008.
Deep Learning in Medical Image Analysis 337
21. Baltruschat, I.M., Nickisch, H., Grass, M., Knopp, T., Saalbach, A.,
Comparison of Deep Learning Approaches for Multi-Label Chest X-Ray
Classification. Sci. Rep., 9, 1, 1–10, Dec. 2019.
22. Diwakar, M. and Kumar, M., A review on CT image noise and its denoising.
Biomed. Signal Process. Control, 42, Elsevier Ltd, 73–88, 01-Apr-2018.
23. Hiremath, P.S., Prema, T., Badiger, S., Speckle Noise Reduction in Medical
Ultrasound Images, in: Advancements and Breakthroughs in Ultrasound
Imaging, InTech, 2013, https://doi.org/10.5772/56519
24. Teymurazyan, A., Riauka, T., Jans, H.S., Robinson, D., Properties of noise in
positron emission tomography images reconstructed with filtered-backpro-
jection and row-action maximum likelihood algorithm. J. Digit. Imaging, 26,
3, 447–456, Jun. 2013.
25. Thakur, R.S., Yadav, R.N., Gupta, L., State-of-art analysis of image denoising
methods using convolutional neural networks. IET Image Process., 13, 13,
2367–2380, 2019.
26. Zhang, K., Zuo, W., Member, S., Chen, Y., Meng, D., Beyond a Gaussian
Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE
Transactions on Image Processing: A Publication of the IEEE Signal Processing
Society, 26, 7, 3142–3155, 2017.
27. Zitova, B., Mathematical approaches for medical image registration, in:
Encyclopedia of Biomedical Engineering, vol. 1–3, pp. 21–32, Elsevier, 2019.
https://doi.org/10.1016/b978-0-12-801238-3.99990-2.
28. Malek, S., Mosleh, M., Dhillon, S.K., Milow, P., Bioimage informatics,
in: Encyclopedia of Bioinformatics and Computational Biology: ABC of
Bioinformatics, vol. 1–3, pp. 993–1010, Elsevier, 2019. https://doi.org/10.
1016/b978-0-12-809633-8.20308-7.
29. Mikolov, T. and Zweig, G., Context dependent recurrent neural network lan-
guage model, in: 2012 IEEE Workshop on Spoken Language Technology, SLT
2012—Proceedings, pp. 234–239, 2012.
30. Graves, A., Mohamed, A., Hinton, G., “Speech Recognition with Deep
Recurrent Neural Networks. ICASSP, IEEE Int. Conf. Acoust. Speech Signal
Process.—Proc, pp. 6645–6649, Mar. 2013.
31. Kalchbrenner, N. and Blunsom, P., Recurrent continuous translation mod-
els, in: EMNLP 2013—2013 Conference on Empirical Methods in Natural
Language Processing, Proceedings of the Conference, 2013.
32. Elman, J.L., Finding structure in time. Cogn. Sci., 14, 2, 179–211, Apr. 1990.
33. Hinton, G.E., Srivastava, N., Krizhevsky, A., Sutskever, I., Salakhutdinov,
R.R., Improving neural networks by preventing co-adaptation of feature detec-
tors, Jul. 2012. arXiv [cs.NE]. arXiv. http://arxiv.org/abs/1207.0580.
34. Hinton, G.E., Osindero, S., Teh, Y.W., A fast learning algorithm for deep
belief nets. Neural Comput., 18, 7, 1527–1554, May 2006.
35. Bengio, Y., Learning deep architectures for AI. Found. Trends Mach. Learn.,
2, 1, 1–27, 2009.
Deep Learning in Medical Image Analysis 339
36. Ackley, D.H., Hinton, G.E., Sejnowski, T.J., A learning algorithm for boltz-
mann machines. Cogn. Sci., 9, 1, 147–169, Jan. 1985.
37. Goodfellow, A.C., II and Bengio, Y., Deep Learning. Genet. Program.
Evolvable Mach., 19, 1–2, 305–307, Jun. 2017.
38. Schawinski, K. et al., Generative adversarial networks recover features in
astrophysical images of galaxies beyond the deconvolution limit. MNRAS,
467, 1, L110–L114, 2017.
39. Yang, Q. et al., Low-Dose CT Image Denoising Using a Generative Adversarial
Network With Wasserstein Distance and Perceptual Loss. IEEE Trans. Med.
Imaging, 37, 6, 1348–1357, Jun. 2018.
40. Ker, J., Wang, L., Rao, J., Lim, T., Deep Learning Applications in Medical
Image Analysis. IEEE Access, 6, 9375–9379, Dec. 2017.
41. Hubel, B., Wiesel, A., Hubel, D. N., Wiesel, T. N., Functional Architecture in
the Cat's Visual Cortex from the Neurophysiolojy Laboratory, Department
of Pharmacology Central Nervous System Is the Great Diversity of Its Cell
Types and Inter-Receptive Fields of a More Complex Type (Part I) and to.
Most. 160, 106–154, 1962.
42. Fukushima, K., Neocognitron: A self-organizing neural network model for
a mechanism of pattern recognition unaffected by shift in position. Biol.
Cybern., 36, 4, 193–202, 1980.
43. Rawat, W. and Wang, Z., Deep convolutional neural networks for image
classification: A comprehensive review. Neural Comput., 29, 9, MIT Press
Journals, 2352–2449, 01-Sep-2017.
44. Tivive, F.H.C. and Bouzerdoum, A., Efficient training algorithms for a class
of shunting inhibitory convolutional neural networks. IEEE Trans. Neural
Networks, 16, 3, 541–556, 2005.
45. Mccann, M.T., Jin, K.H., Unser, M., Deep CNN for Inverse Problems in
Imaging. IEEE Signal Process. Mag., 26, 9, 4509–4522, 2017.
46. Jifara, W., Jiang, F., Rho, S., Cheng, M., Liu, S., Medical image denoising using
convolutional neural network: A residual learning approach. J. Supercomput.,
75, 2, 704–718, Feb. 2019.
47. He, K., Zhang, X., Ren, S., Sun, J., Deep residual learning for image recogni-
tion, in: Proceedings of the IEEE Computer Society Conference on Computer
Vision and Pattern Recognition, December, vol. 2016, pp. 770–778, 2016.
48. Ioffe, S. and Szegedy, C., Batch normalization: Accelerating deep network
training by reducing internal covariate shift, in: 32nd International Conference
on Machine Learning, ICML 2015, vol. 1, pp. 448–456, 2015.
49. Jiang, D., Dou, W., Vosters, L., Xu, X., Sun, Y., Tan, T., Denoising of 3D mag-
netic resonance images with multi-channel residual learning of convolu-
tional neural network. Jpn. J. Radiol., 36, 9, 566–574, 2018.
50. Tripathi, P.C. and Bag, S., CNN-DMRI: A Convolutional Neural Network
for Denoising of Magnetic Resonance Images. Pattern Recognit. Lett., 135,
57–63, 2020.
340 Biomedical Data Mining for Information Retrieval
51. Gong, K., Guan, J., Liu, C.-C., Qi, J., PET Image Denoising Using a Deep
Neural Network Through Fine Tuning. IEEE Trans. Radiat. Plasma Med. Sci.,
3, 2, 153–161, 2019.
52. Kadimesetty, V.S., Gutta, S., Ganapathy, S., Yalavarthy, P.K., Convolutional
Neural Network-Based Robust Denoising of Low-Dose Computed
Tomography Perfusion Maps. IEEE Trans. Radiat. Plasma Med. Sci., 3, 2,
137–152, 2019.
53. Pavoni, M., Chang, Y., Park, S., Smedby, O., Convolutional neural net-
work-based image enhancement for x-ray percutaneous coronary interven-
tion. J. Med. Imaging, 5, 02, 1, Jun. 2018.
54. Kumar, K. and Rao, A.C.S., Breast cancer classification of image using con-
volutional neural network, in: Proceedings of the 4th IEEE International
Conference on Recent Advances in Information Technology, RAIT 2018, pp.
1–6, 2018.
55. Diaz-Pinto, A., Morales, S., Naranjo, V., Köhler, T., Mossi, J.M., Navea, A.,
CNNs for automatic glaucoma assessment using fundus images: An exten-
sive validation. Biomed. Eng. Online, 18, 1, 29, Mar. 2019.
56. Carpenter, A.E., Jones, T.R., Lamprecht, M.R., Clarke, C., Kang, I,H., Friman,
O., Guertin, D.A. et al., CellProfiler: Image analysis software for identifying
and quantifying cell phenotypes. Genome Biol., 7, 10, R100, 2006.
57. Handfield, L.F., Strome, B., Chong, Y.T., Moses, A.M., Local statistics allow
quantification of cell-to-cell variability from high-throughput microscope
images. Bioinformatics, 31, 6, 940–947, 2015.
58. Hung, M.-C. and Link, W., Protein localization in disease and therapy. J. Cell
Sci., 124, 20, 3381–3392, Oct. 2011.
59. Boland, M.V. and Murphy, R.F., A neural network classifier capable of recog-
nizing the patterns of all major subcellular structures in fluorescence micro-
scope images of HeLa cells. Bioinformatics, 17, 12, 1213–1223, 2001.
60. Li, J., Newberg, J.Y., Uhlén, M., Lundberg, E., Murphy, R.F., Automated
Analysis and Reannotation of Subcellular Locations in Confocal Images
from the Human Protein Atlas. PLoS One, 7, 11, e50514, Nov. 2012.
61. Tahir, M., Khan, A., Majid, A., Protein subcellular localization of fluores-
cence imagery using spatial and transform domain features. Bioinformatics
28, 1, 91–97, 2012.
62. Yang, F., Liu, Y., Wang, Y., Yin, Z., Yang, Z., MIC_Locator: A novel image-
based protein subcellular location multi-label prediction model based on
multi-scale monogenic signal representation and intensity encoding strat-
egy. BMC Bioinf., 20, 1, 522, Oct. 2019.
63. Pärnamaa, T. and Parts, L., Accurate Classification of Protein Subcellular
Localization from High-Throughput Microscopy Images Using Deep
Learning. G3: Genes|Genomes|Genetics, 7, 5, 1385–1392, May 2017.
64. Kraus, O.Z. et al., Automated analysis of high-content microscopy data with
deep learning. Mol. Syst. Biol., 13, 4, 924, Apr. 2017.
Deep Learning in Medical Image Analysis 341
65. Scheeder, C., Heigwer, F., Boutros, M., Machine learning and image-based
profiling in drug discovery. Curr. Opin. Syst. Biol., 10, 43–52, Aug. 2018.
66. Ohnuki, S., Oka, S., Nogami, S., Ohya, Y., High-Content, Image-Based
Screening for Drug Targets in Yeast. PLoS One, 5, 4, e10177, Apr. 2010.
67. Wawer, M.J. et al., Toward performance-diverse small-molecule libraries for
cell-based phenotypic screening using multiplexed high-dimensional profil-
ing. Proc. Natl. Acad. Sci. U. S. A., 111, 30, 10911–10916, 2014.
68. Kensert, A., Harrison, P.J., Spjuth, O., Transfer Learning with Deep
Convolutional Neural Networks for Classifying Cellular Morphological
Changes. SLAS Discovery, 24, 4, 466–475, Apr. 2019.
69. Hernández-Mier, Y., Nuño-Maganda, M.A., Polanco-Martagón, S., del R.
García-Chávez, M., Machine Learning Classifiers Evaluation for Automatic
Karyogram Generation from G-Banded Metaphase Images. Appl. Sci., 10, 8,
2758, Apr. 2020.
70. Alexe, G. et al., High expression of lymphocyte-associated genes in node-neg-
ative HER2+ breast cancers correlates with lower recurrence rates. Cancer
Res., 67, 22, 10669–10676, Nov. 2007.
71. Xu, J. et al., Convolutional neural network initialized active contour model
with adaptive ellipse fitting for nuclear segmentation on breast histopatho-
logical images. J. Med. Imaging, 6, 01, 1, Feb. 2019.
72. Sirinukunwattana, K., Raza, S.E.A., Tsang, Y.W., Snead, D.R.J., Cree,
I.A., Rajpoot, N.M., Locality Sensitive Deep Learning for Detection and
Classification of Nuclei in Routine Colon Cancer Histology Images. IEEE
Trans. Med. Imaging, 35, 5, 1196–1206, May 2016.
73. Ounkomol, C., Seshamani, S., Maleckar, M.M., Collman, F., Johnson, G.R.,
Label-free prediction of three-dimensional fluorescence images from trans-
mitted-light microscopy. Nat. Methods, 15, 11, 917–920, Nov. 2018.
74. Liimatainen, K., Valkonen, M., Latonen, L., Ruusuvuori, P., Cell organelle
classification with fully convolutional neural networks, 2018.
75. Su, H., Xing, F., Kong, X., Xie, Y., Zhang, S., Yang, L., Robust cell detection
and segmentation in histopathological images using sparse reconstruction
and stacked denoising autoencoders, in: Lecture Notes in Computer Science
(including subseries Lecture Notes in Artificial Intelligence and Lecture Notes
in Bioinformatics), vol. 9351, pp. 383–390, 2015.
76. Weigert, M. et al., Content-aware image restoration: pushing the limits of
fluorescence microscopy. Nat. Methods, 15, 12, 1090–1097, Dec. 2018.
77. Janowczyk, A., Basavanhally, A., Madabhushi, A., Stain Normalization using
Sparse AutoEncoders (StaNoSA): Application to digital pathology. Comput.
Med. Imaging Graph., 57, 50–61, Apr. 2017.
78. Yang, S.J. et al., Assessing microscope image focus quality with deep learn-
ing. BMC Bioinf., 19, 1, 77, Mar. 2018.
79. Gao, Z., Wang, L., Zhou, L., Zhang, J., HEp-2 Cell Image Classification with
Deep Convolutional Neural Networks. IEEE J. Biomed. Health Inform., 21, 2,
416–428, Mar. 2017.
342 Biomedical Data Mining for Information Retrieval
93. Zhong, Q., Busetto, A.G., Fededa, J.P., Buhmann, J.M., Gerlich, D.W.,
Unsupervised modeling of cell morphology dynamics for time-lapse micros-
copy. Nat. Methods, 9, 7, 711–713, May 2012.
94. Chen, Y., Ladi, E., Herzmark, P., Robey, E., Roysam, B., Automated {5-D}
analysis of cell migration and interaction in the thymic cortex from time-
lapse sequences of {3-D} multi-channel multi-photon images. J. Immunol.
Methods, 340, 1, 65–80, Jan. 2009.
95. Pawlowski, N., Caicedo, J.C., Singh, S., Carpenter, A.E., Automating mor-
phological profiling with generic deep convolutional networks. BioRxiv,
2016. https://www.biorxiv.org/content/10.1101/085118v1.abstract.
96. Paul, A., Dey, A., Mukherjee, D.P., Sivaswamy, J., Tourani, V., Regenerative
Random Forest with Automatic Feature Selection to Detect Mitosis in
Histopathological Breast Cancer Images, in: Medical Image Computing and
{Computer-Assisted} Intervention—{MICCAI} 2015, pp. 94–102, 2015.
97. Mao, Y. and Yin, Z., A Hierarchical Convolutional Neural Network for
Mitosis Detection in {Phase-Contrast} Microscopy Images, in: Medical
Image Computing and {Computer-Assisted} Intervention—{MICCAI} 2016,
pp. 685–692, 2016.
98. Balkenhol, M., Karssemeijer, N., Litjens, G.J.S., van der Laak, J., Ciompi, F.,
Tellez, D., {H&E} stain augmentation improves generalization of convolu-
tional networks for histopathological mitosis detection. Medical Imaging
2018: Digital Pathology, 2018.
99. Ouyang, W., Aristov, A., Lelek, M., Hao, X., Zimmer, C., Deep learning mas-
sively accelerates super-resolution localization microscopy. Nat. Biotechnol.,
36, 5, 460–468, Jun. 2018.
100. Li, C., Wang, X., Liu, W., Latecki, L.J., Wang, B., Huang, J., Weakly supervised
mitosis detection in breast histopathology images using concentric loss. Med.
Image Anal., 53, 165–178, Apr. 2019.
101. Rajaraman, S., Jaeger, S., Antani, S.K., Performance evaluation of deep neural
ensembles toward malaria parasite detection in thin-blood smear images.
PeerJ, 7, e6977, May 2019.
102. Shirazi, S.H., Umar, A.I., Naz, S., Razzak, M.I., Efficient leukocyte segmenta
tion and recognition in peripheral blood image. Technol. Healthcare, 24, 3,
335–347, May 2016.
103. Van Valen, D.A. et al., Deep Learning Automates the Quantitative Analysis
of Individual Cells in {Live-Cell} Imaging Experiments. PloS Comput. Biol.,
12, 11, e1005177, Nov. 2016.
104. Mao, Y., Yin, Z., Schober, J.M., Iteratively training classifiers for circulat-
ing tumor cell detection, in: 2015 {IEEE} 12th International Symposium on
Biomedical Imaging ({ISBI}), pp. 190–194, 2015.
105. Ciompi, F. et al., The importance of stain normalization in colorectal tissue
classification with convolutional networks, in: 2017 {IEEE} 14th International
Symposium on Biomedical Imaging ({ISBI} 2017), pp. 160–163, 2017.
344 Biomedical Data Mining for Information Retrieval
121. Bernal, J. et al., Deep convolutional neural networks for brain image analysis
on magnetic resonance imaging: A review. Artif. Intell. Med., 95, 64–81, Apr.
2019.
122. Juntu, J., Sijbers, J., Van Dyck, D., Gielen, J., Bias Field Correction for {MRI}
Images, in: Computer Recognition Systems, pp. 543–551, 2005.
123. Pereira, S., Pinto, A., Alves, V., Silva, C.A., Brain Tumor Segmentation Using
Convolutional Neural Networks in {MRI} Images. IEEE Trans. Med. Imaging,
35, 5, 1240–1251, 2016.
124. Kamnitsas, K. et al., Efficient multi-scale {3D} {CNN} with fully connected
{CRF} for accurate brain lesion segmentation. Med. Image Anal., 36, 61–78,
Feb. 2017.
125. Li, W., Wang, G., Fidon, L., Ourselin, S., Cardoso, M.J., Vercauteren, T., On
the Compactness, Efficiency, and Representation of {3D} Convolutional
Networks: Brain Parcellation as a Pretext Task, in: Information Processing in
Medical Imaging, pp. 348–360, 2017.
126. Zhang, W. et al., Deep convolutional neural networks for multi-modality iso-
intense infant brain image segmentation. Neuroimage, 108, 214–224, Mar.
2015.
127. Dou, Q. et al., Automatic Detection of Cerebral Microbleeds From {MR}
Images via {3D} Convolutional Neural Networks. IEEE Trans. Med. Imaging,
35, 5, 1182–1195, May 2016.
128. Brosch, T., Tang, L.Y.W., Yoo, Y., Li, D.K.B., Traboulsee, A., Tam, R., Deep
{3D} Convolutional Encoder Networks With Shortcuts for Multiscale
Feature Integration Applied to Multiple Sclerosis Lesion Segmentation. IEEE
Trans. Med. Imaging, 35, 5, 1229–1239, May 2016.
129. de Brebisson, A. and Montana, G., Deep neural networks for anatomical
brain segmentation, in: Proceedings of the {IEEE} Conference on Computer
Vision and Pattern Recognition Workshops, pp. 20–28, 2015.
130. Valverde, S. et al., Improving automated multiple sclerosis lesion seg-
mentation with a cascaded {3D} convolutional neural network approach.
Neuroimage, 155, 159–168, Jul. 2017.
131. Lyksborg, M., Puonti, O., Agn, M., Larsen, R., An Ensemble of {2D}
Convolutional Neural Networks for Tumor Segmentation, in: Image Analysis,
pp. 201–211, 2015.
132. Ebrahimighahnavieh, M.A., Luo, S., Chiong, R., Deep learning to detect
Alzheimer’s disease from neuroimaging: A systematic literature review.
Comput. Methods Programs Biomed., 187, 105242, Apr. 2020.
133. Suk, H.-I., Lee, S.-W., Shen, D., Alzheimer’s Disease Neuroimaging Initiative,
Latent feature representation with stacked auto-encoder for {AD/MCI} diag-
nosis. Brain Struct. Funct., 220, 2, 841–859, Mar. 2015.
134. Hosseini-Asl, E., Gimel, F.G., El-Baz, A., Alzheimer’s Disease Diagnostics
by Adaptation of 3D Convolutional Network. In 2016 IEEE International
Conference on Image Processing (ICIP), 2016. https://doi.org/10.1109/
icip.2016.7532332.
346 Biomedical Data Mining for Information Retrieval
135. Suk, H.-I. and Shen, D., Deep ensemble sparse regression network for
Alzheimer’s disease diagnosis, in: International Workshop on Machine
Learning in Medical Imaging, pp. 113–121, 2016.
136. Basaia, S. et al., Automated classification of Alzheimer’s disease and mild
cognitive impairment using a single {MRI} and deep neural networks.
NeuroImage Clin., 21, 101645, 2019.
137. Wang, H. et al., Ensemble of {3D} densely connected convolutional net-
work for diagnosis of mild cognitive impairment and Alzheimer’s disease.
Neurocomputing, 333, 145–156, 2019.
138. Fraiwan, L., Khnouf, R., Mashagbeh, A.R., Parkinson’s disease hand tremor
detection system for mobile application. J. Med. Eng. Technol., 40, 3, 127–
134, 2016.
139. Jane, Y.N., Nancy Jane, Y., Khanna Nehemiah, H., Arputharaj, K., A
Q-backpropagated time delay neural network for diagnosing severity of gait
disturbances in Parkinson’s disease. J. Biomed. Inf., 60, 169–176, 2016.
140. Quan, J., Xu, L., Xu, R., Tong, T., Su, J., DaTscan SPECT Image Classification
for Parkinson’s Disease, 2019. arXiv [eess.IV]. arXiv. http://arxiv.org/
abs/1909.04142.
141. Abràmoff, M.D. et al., Improved Automated Detection of Diabetic
Retinopathy on a Publicly Available Dataset Through Integration of Deep
Learning. Invest. Ophthalmol. Vis. Sci., 57, 13, 5200–5206, Oct. 2016.
142. Gulshan, V. et al., Development and Validation of a Deep Learning Algorithm
for Detection of Diabetic Retinopathy in Retinal Fundus Photographs.
JAMA, 316, 22, 2402–2410, Dec. 2016.
143. Abràmoff, M.D., Lavin, P.T., Birch, M., Shah, N., Folk, J.C., Pivotal trial of an
autonomous {AI-based} diagnostic system for detection of diabetic retinop-
athy in primary care offices. NPJ Digit Med., 1, 39, Aug. 2018.
144. Ting, D.S.W. et al., Development and Validation of a Deep Learning System
for Diabetic Retinopathy and Related Eye Diseases Using Retinal Images From
Multiethnic Populations With Diabetes. JAMA, 318, 22, 2211–2223, 2017.
145. Grassmann, F. et al., A Deep Learning Algorithm for Prediction of {Age-
Related} Eye Disease Study Severity Scale for {Age-Related} Macular
Degeneration from Color Fundus Photography. Ophthalmology, 125, 9,
1410–1420, Sep. 2018.
146. Chen, X., Xu, Y., Yan, S., Wong, D.W.K., Wong, T.Y., Liu, J., Automatic Feature
Learning for Glaucoma Detection Based on Deep Learning, in: Medical
Image Computing and {Computer-Assisted} Intervention—{MICCAI} 2015,
pp. 669–677, 2015.
147. Puthren, P.R., Agrawal, A., Padma, U., Automated Glaucoma Detection
Using Global Statistical Parameters of Retina Fundus Images, in: Proceedings
of the International Conference on {ISMAC} in Computational Vision and
{Bio-Engineering} 2018 ({ISMAC-CVB}), pp. 377–388, 2019.
Deep Learning in Medical Image Analysis 347
148. Devasia, T., Jacob, K.P., Thomas, T., Automatic Early Stage Glaucoma
Detection Using Cascade Correlation Neural Network, in: Smart Intelligent
Computing and Applications, pp. 659–669, 2019.
149. Viscaino, M., Maass, J.C., Delano, P.H., Torrente, M., Stott, C., Auat Cheein,
F., Computer-aided diagnosis of external and middle ear conditions: A
machine learning approach. PLoS One, 15, 3, e0229226, Mar. 2020.
150. Nishio, M. et al., Computer-aided diagnosis of lung nodule classification
between benign nodule, primary lung cancer, and metastatic lung cancer at
different image size using deep convolutional neural network with transfer
learning. PLoS One, 13, 7, e0200721, Jul. 2018.
151. Huang, S., Lee, F., Miao, R., Si, Q., Lu, C., Chen, Q., A deep convolutional
neural network architecture for interstitial lung disease pattern classification.
Med. Biol. Eng. Comput., 58, 4, 725–737, Apr. 2020.
152. Bermejo-Peláez, D., Ash, S.Y., Washko, G.R., San José Estépar, R., Ledesma-
Carbayo, M.J., Classification of Interstitial Lung Abnormality Patterns with
an Ensemble of Deep Convolutional Neural Networks. Sci. Rep., 10, 1, 338,
Jan. 2020.
153. Chiang, T.-C., Huang, Y.-S., Chen, R.-T., Huang, C.-S., Chang, R.-F., Tumor
Detection in Automated Breast Ultrasound Using {3-D} {CNN} and Prioritized
Candidate Aggregation. IEEE Trans. Med. Imaging, 38, 1, 240–249, Jan. 2019.
154. Sahiner, B. et al., Classification of mass and normal breast tissue: a convolu-
tion neural network classifier with spatial domain and texture images. IEEE
Trans. Med. Imaging, 15, 5, 598–610, 1996.
155. Sun, W., Tseng, T.-L.B., Zhang, J., Qian, W., Enhancing deep convolutional
neural network scheme for breast cancer diagnosis with unlabeled data.
Comput. Med. Imaging Graph., 57, 4–9, Apr. 2017.
156. Hwang, S. and Kim, H.-E., {Self-Transfer} Learning for Weakly Supervised
Lesion Localization, in: Medical Image Computing and {Computer-Assisted}
Intervention—{MICCAI} 2016, pp. 239–246, 2016.
157. Schlemper, J., Caballero, J., Hajnal, J.V., Price, A.N., Rueckert, D., A Deep
Cascade of Convolutional Neural Networks for Dynamic {MR} Image
Reconstruction. IEEE Trans. Med. Imaging, 37, 2, 491–503, Feb. 2018.
158. Chen, H. et al., Low-dose {CT} via convolutional neural network. Biomed.
Opt. Express, 8, 2, 679–694, Feb. 2017.
159. Sadda, P. and Qarni, T., {Real-Time} Medical Video Denoising with Deep
Learning: Application to Angiography. Int. J. Appl. Inf. Syst., 12, 13, 22–28,
May 2018.
160. Bai, W. et al., Automated cardiovascular magnetic resonance image analysis
with fully convolutional networks. J. Cardiovasc. Magn. Reson., 20, 1, 65, Sep.
2018.
161. Bello, G.A. et al., Deep learning cardiac motion analysis for human survival
prediction. Nat. Mach. Intell., 1, 95–104, Feb. 2019.
162. Taylor, C.A., Fonte, T.A., Min, J.K., Computational fluid dynamics applied to
cardiac computed tomography for noninvasive quantification of fractional
348 Biomedical Data Mining for Information Retrieval
flow reserve: Scientific basis. J. Am. Coll. Cardiol., 61, 22, 2233–2241, Jun.
2013.
163. Shah, S.J. et al., Phenomapping for novel classification of heart failure with
preserved ejection fraction. Circulation, 131, 3, 269–279, Jan. 2015.
164. Orlando, N., Gillies, D.J., Gyacskov, I., Romagnoli, C., D’Souza, D., Fenster,
A., Automatic prostate segmentation using deep learning on clinically
diverse {3D} transrectal ultrasound images. Med. Phys., Mar. 2020. https://
doi.org/10.1002/mp.14134.
165. Roth, H.R., Farag, A., Lu, L., Turkbey, E.B., Summers, R.M., Deep convo-
lutional networks for pancreas segmentation in {CT} imaging, in: Medical
Imaging 2015: Image Processing, vol. 9413, p. 94131G, 2015.
166. Thong, W., Kadoury, S., Piché, N., Pal, C.J., Convolutional networks for
kidney segmentation in contrast-enhanced {CT} scans. Comput. Methods
Biomech. Biomed. Eng. Imaging Vis., 6, 3, 277–282, May 2018.
167. Park, S. et al., Annotated normal {CT} data of the abdomen for deep learn-
ing: Challenges and strategies for implementation. Diagn. Interv. Imaging,
101, 1, 35–44, Jan. 2020.
168. Zhang, R. et al., Automatic Detection and Classification of Colorectal Polyps
by Transferring {Low-Level} {CNN} Features From Nonmedical Domain.
IEEE J. Biomed. Health Inform., 21, 1, 41–47, Jan. 2017.
169. Liu, J. et al., Colitis detection on computed tomography using regional con-
volutional neural networks, in: 2016 {IEEE} 13th International Symposium on
Biomedical Imaging ({ISBI}), pp. 863–866, 2016.
170. Tachibana, R., Näppi, J.J., Hironakaa, T., Kim, S.H., Yoshida, H., Deep learn-
ing for electronic cleansing in dual-energy {CT} colonography, in: Medical
Imaging 2016: {Computer-Aided} Diagnosis, vol. 9785, p. 97851M, 2016.
171. Vivanti, R., Ephrat, A., Joskowicz, L., Karaaslan, O., Lev-Cohain, N., Sosna,
J., Automatic liver tumor segmentation in follow-up {CT} studies using
convolutional neural networks, in: Proc. {Patch-Based} Methods in Medical
Image Processing Workshop, vol. 2, 2015.
172. Zhu, Y. et al., {MRI-based} prostate cancer detection with high-level repre-
sentation and hierarchical classification. Med. Phys., 44, 3, 1028–1039, Mar.
2017.
173. Pei, M., Wu, X., Guo, Y., Fujita, H., Small bowel motility assessment based
on fully convolutional networks and long short-term memory. Knowl.-Based
Syst., 121, 163–172, Apr. 2017.
174. Wimmer, G., Vécsei, A., Uhl, A., {CNN} transfer learning for the automated
diagnosis of celiac disease, in: 2016 Sixth International Conference on Image
Processing Theory, Tools and Applications ({IPTA}), pp. 1–6, 2016.
175. Zhu, R., Zhang, R., Xue, D., Lesion detection of endoscopy images based on
convolutional neural network features, in: 2015 8th International Congress on
Image and Signal Processing ({CISP}), pp. 372–376, 2015.
Deep Learning in Medical Image Analysis 349
176. Yu, L., Chen, H., Dou, Q., Qin, J., Heng, P.-A., Automated Melanoma
Recognition in Dermoscopy Images via Very Deep Residual Networks. IEEE
Trans. Med. Imaging, 36, 4, 994–1004, Apr. 2017.
177. Esteva, A. et al., Dermatologist-level classification of skin cancer with deep
neural networks. Nature, 542, 7639, 115–118, Feb. 2017.
178. Haenssle, H.A. et al., Man against machine: Diagnostic performance of a deep
learning convolutional neural network for dermoscopic melanoma recognition
in comparison to 58 dermatologists. Ann. Oncol., 29, 8, 1836–1842, Aug. 2018.
179. Fujisawa, Y. et al., Deep-learning-based, computer-aided classifier developed
with a small dataset of clinical images surpasses board-certified dermatolo-
gists in skin tumour diagnosis. Br. J. Dermatol., 180, 2, 373–381, Feb. 2019.
180. Brinker, T.J. et al., A convolutional neural network trained with dermoscopic
images performed on par with 145 dermatologists in a clinical melanoma
image classification task. Eur. J. Cancer, 111, 148–154, Apr. 2019.
181. Golan, D., Donner, Y., Mansi, C., Jaremko, J., Ramachandran, M., Others,
Fully automating Graf ’s method for {DDH} diagnosis using deep convo-
lutional neural networks, in: Deep Learning and Data Labeling for Medical
Applications, 2016. https://doi.org/10.1007/978-3-319-46976-8_14.
182. Jamaludin, A., Kadir, T., Zisserman, A., {SpineNet}: Automatically
Pinpointing Classification Evidence in Spinal {MRIs}, in: Medical Image
Computing and {Computer-Assisted} Intervention—{MICCAI} 2016, pp. 166–
175, 2016.
183. Litjens, G., Kooi, T., Bejnordi, B.E., Setio, A.A.A., Ciompi, F., Ghafoorian,
M. et al., A survey on deep learning in medical image analysis. Med. Image
Anal., 42, 60–88, 2017.
184. Lee, H. et al., Fully Automated Deep Learning System for Bone Age
Assessment. J. Digit. Imaging, 30, 4, 427–441, Aug. 2017.
185. Tu, Z., Xie, W., Qin, Q., Poppe, R., Veltkamp, R.C., Li, B. et al., Multi-stream
CNN: Learning representations based on human-related regions for action
recognition. Pattern Recognit., 79, 32–43, 2018.
186. Chaudhari, A.S. et al., Super-resolution musculoskeletal {MRI} using deep
learning. Magn. Reson. Med., 80, 5, 2139–2154, Nov. 2018.
187. Liu, F. et al., Fully Automated Diagnosis of Anterior Cruciate Ligament Tears
on Knee {MR} Images by Using Deep Learning. Radiol. Artif. Intell., 1, 3,
180091, May 2019.
188. Li, S., Deep adversarial model for musculoskeletal quality evaluation. Inf.
Process. Manage., 57, 1, 102146, Jan. 2020.
189. Zheng, Q., Shellikeri, S., Huang, H., Hwang, M., Sze, R.W., Deep Learning
Measurement of Leg Length Discrepancy in Children Based on Radiographs.
Radiology, 2020. https://doi.org/10.1148/radiol.2020192003.
190. Jaumard-Hakoun, A., Xu, K., Roussel-Ragot, P., Dreyfus, G., Denby, B.,
Tongue contour extraction from ultrasound images based on deep neural net-
work, May 2016. arXiv [cs.CV]. arXiv. http://arxiv.org/abs/1605.05912.
350 Biomedical Data Mining for Information Retrieval
191. Quinn, J.A., Nakasi, R., Mugagga, P.K.B., Byanyima, P., Lubega, W., Andama,
A., Deep Convolutional Neural Networks for {Microscopy-Based} Point of Care
Diagnostics, in: Machine Learning for Healthcare Conference, pp. 271–281, 2016.
192. Lekadir, K. et al., A Convolutional Neural Network for Automatic
Characterization of Plaque Composition in Carotid Ultrasound. IEEE J.
Biomed. Health Inf., 21, 1, 48–55, Jan. 2017.
193. Rad, R.M., Saeedi, P., Au, J., Havelock, J., Trophectoderm segmentation
in human embryo images via inceptioned {U-Net}. Med. Image Anal., 62,
101612, Feb. 2020.
194. Nardelli, P., Ross, J.C., San José Estépar, R., Generative-based airway and ves-
sel morphology quantification on chest {CT} images. Med. Image Anal., 63,
101691, Mar. 2020.
195. Ghesu, F.C., Georgescu, B., Mansi, T., Neumann, D., Hornegger, J.,
Comaniciu, D., An Artificial Agent for Anatomical Landmark Detection
in Medical Images, in: Medical Image Computing and {Computer-Assisted}
Intervention—{MICCAI} 2016, pp. 229–237, 2016.
196. Moeskops, P. et al., Deep Learning for Multi-task Medical Image
Segmentation in Multiple Modalities, in: Medical Image Computing and
{Computer-Assisted} Intervention—{MICCAI} 2016, pp. 478–486, 2016.
197. Roth, H.R. et al., Improving {Computer-Aided} Detection Using
Convolutional Neural Networks and Random View Aggregation. IEEE
Trans. Med. Imaging, 35, 5, 1170–1181, May 2016.
198. Kerlikowske, K. et al., Performance of screening mammography among
women with and without a first-degree relative with breast cancer. Ann.
Intern. Med., 133, 11, 855–863, Dec. 2000.