Professional Documents
Culture Documents
net
Current Medical Imaging, 2021, 17, 1-00 1
REVIEW ARTICLE
1
Department of Computer Science, Jamia Millia Islamia, New Delhi, India; 2Department of Computer Science, Jamia
Millia Islamia, New Delhi, India
Abstract: Background: Interpretation of medical images for the diagnosis and treatment of com-
plex diseases from high-dimensional and heterogeneous data remains a key challenge in transform-
ing healthcare. In the last few years, both supervised and unsupervised deep learning achieved
promising results in the area of medical image analysis. Several reviews on supervised deep learn-
ing are published, but hardly any rigorous review on unsupervised deep learning for medical image
analysis is available.
Objectives: The objective of this review is to systematically present various unsupervised deep
ARTICLE HISTORY
learning models, tools, and benchmark datasets applied to medical image analysis. Some of the dis-
Received: May 30, 2020 cussed models are autoencoders and their variants, restricted Boltzmann machines (RBM), deep be-
Revised: November 17, 2020 lief networks (DBN), deep Boltzmann machine (DBM), and generative adversarial network
Accepted: December 16, 2020
(GAN). Future research opportunities and challenges of unsupervised deep learning techniques for
DOI: medical image analysis are also discussed.
10.2174/1573405617666210127154257
Conclusion: Currently, interpretation of medical images for diagnostic purposes is usually per-
formed by human experts that may be replaced by computer-aided diagnosis due to advancement
in machine learning techniques, including deep learning, and the availability of cheap computing in-
frastructure through cloud computing. Both supervised and unsupervised machine learning ap-
proaches are widely applied in medical image analysis, each of them having certain pros and cons.
Since human supervisions are not always available or are inadequate or biased, therefore, unsuper-
vised learning algorithms give a big hope with lots of advantages for biomedical image analysis.
Keywords: Unsupervised learning, medical image analysis, autoencoders, restricted boltzmann machine, deep belief network.
the algorithms is limited in three ways: (i) A huge manual ef- (2)
fort to create labels, (ii) To prepare biases to check the algo-
rithms’ supervised functioning, and (iii) Reduce the scalabili-
ty of the target function at hand. (3)
To intelligently solve these issues, an unsupervised ma- where Sf and Sg are encoder and decoder activation func-
chine learning algorithm can be used. Unsupervised ma- tion (normally sigmoid, hyperbolic tangent, or an identity
chine learning algorithms not only derive insights directly function), while the parameters of the model θ = {W, b, W',
from the data and group the data based on their patterns but d} where W and W’ are encoder-decoder weight matrices,
also use these insights for data-driven decision-making. Al- and b and d are encoder and decoder bias vector respective-
so, unsupervised models are more robust in the sense that ly. Moreover, regularization or sparse constraints may be ap-
they act as a base for several different complex tasks where plied to boost the discovery process. In such a case, the hid-
these can be utilized as the holy grail of learning and classifi- den layer(s) has the same output as the input layer, and no
cation. In fact, classification is not the only task that is em- non-linearity is added. The model would simply learn an
ployed, but other tasks such as compression, dimensionality identity function. (Fig. 2 and 3 illustrate the basic structure
reduction, denoising, super-resolution, and some degree of of AE.
decision making are also performed. Therefore, it is more
useful to construct a model without having any knowledge
about the tasks. In a nutshell, we can think of unsupervised
learning as a preparation (pre-processing) step for super-
vised learning tasks, where the latter may allow better gener-
alization of a classifier [2].
Unsupervised pre-training gained popularity because of
its simplicity and has been highly addressed by the medical
imaging community. Moreover, autoencoders, along with
their several variants and other unsupervised methods, are re-
ported, which are being extensively applied in medical im-
age analysis.
3.1. Autoencoders and its Variants Fig. (1). Unsupervised deep learning models. (A higher resolution /
colour version of this figure is available in the electronic copy of
3.1.1. Autoencoders and Stacked Autoencoder the article).
Autoencoders (AEs) [3] are simple unsupervised learn-
ing models consisting of a single-layer neural network that Stacked autoencoders (SAEs) are constructed by organiz-
transforms the input into a latent or compressed representa- ing AEs on top of one another, also known as deep AEs.
tion by minimizing the reconstruction errors between input They consist of multiple AEs stacked into multiple layers
and output values of the network. By constraining the dimen- where the output of each layer is wired to the inputs of the
sion of latent representation (maybe from different input (s), successive layers (Fig. 2(B)). To obtain good parameters,
it is possible to discover relevant patterns from the data. Au- SAE uses greedy layer-wise training. The benefit of this
toencoder framework defines a feature to extract function type of autoencoder is that it can enjoy the benefits of a deep
with specific parameters [4]. Basically, AEs are trained with network having greater expressive power. Furthermore, it
a specific function. fθ is called encoder and h = fθ (x) is a fea- usually captures a useful hierarchical grouping of the input
ture vector or representation from input x. Another parame- [5].
terized function gθ is called a decoder which produces input
space back from feature space. In short, basic AEs are 3.1.2. Denoising autoencoder
trained to minimize reconstruction error in finding a value of Denoising autoencoder (DAE) is another variant of the
parameter θ given by: auto-encoder. Denoising is important and is encapsulated as
a training criterion for efficient and robust learning and to ex-
(1) tract useful features [6]. They also prevent the model from
learning a trivial solution, where Litjens et al. [7] model is
This minimization is optionally followed by a non-linear- trained to reconstruct a clean input from the corrupted ver-
ity (most commonly used for encoder and decoder) as given sion from noise or another corruption which is done by corr-
by:
A Tour of Unsupervised Deep Learning Current Medical Imaging, 2021, Vol. 17, No. 00 3
Fig. (2 A-D). A collage of networks of autoencoders and its variants. (A higher resolution / colour version of this figure is available in the
electronic copy of the article).
Fig. (3). A collage of networks in (A) Convolutional autoencoder, (B) Variational autoencoder, and (C) Contractive autoencoders. (A higher
resolution / colour version of this figure is available in the electronic copy of the article).
4 Current Medical Imaging, 2021, Vol. 17, No. 00 Raza and Singh
upting the initial input x into by using a stochastic func- The property of the penalty function is that KL (P || j) =
tion . The corrupted input is then mapped to 0, if Pj = j, otherwise, it increases gradually as j diverse
a hidden representation and recon- for P.
struction z = gθ (y). A schematic representation of DAE is
shown in Fig. (2(C)). Parameters θ and θ' are initialized ran- 3.1.4. Convolutional Autoencoder
domly and trained using stochastic gradient descent to mini- As current research witness, the stacked AE is the most
mize average reconstruction error. The denoising autoen- popular and widely used network model in deep unsuper-
coders continue minimizing the same reconstruction loss be- vised architecture [7], which requires layer-wise pre-train-
tween clean X and reconstruction from Y. This continues ing. When layers go deeper during the pre-training process
maximizing a lower bound on the mutual information be- making fully connected layers, the entire process becomes
tween input x and representation y, and the difference is ob- very tedious and time-consuming. Li et al. [13] proposed the
tained by applying mapping fθ to a corrupted input. Hence, first trial to train convolutional directly in an end-to-end
such learning is better than the identity, and it extracts fea- manner without pre-training. Guo et al. [14] suggested that
tures useful for denoising. Stack denoising autoencoder (S- convolutional autoencoder (CAE) is beneficial to learn fea-
DAE) is a deep network utilizing the power of DAE [6, 8] tures for images and preserving the local structure of data
and RBMs [9, 10]. and to avoid distortion of feature space. A typical architec-
ture of CAE is depicted in Fig. (3).
3.1.3. Sparse Autoencoder
The limitation of autoencoders to have only small num- 3.1.5. Variational Autoencoder
bers of hidden units can be overcome by adding a sparse con- Another variant of autoencoder, called variational autoen-
straint, where a large number of hidden units, more than coder (VAE), was introduced as a generative model [15]. A
one, can be introduced. The aim of sparse autoencoder typical architecture of VAE is given in Fig. 3. They utilize
(SAE) is to make a large number of neurons have a low aver- the strategy of deriving a lower bond estimator from the di-
age output so that neurons are inactive most of the time. rected graphical models with a continuous distribution of la-
Sparsest can be achieved by introducing a loss function dur- tent variables. The generative parameter θ in the decoder
ing the training or by simple manual zeroing few strongest (generative model) assists the learning process of the varia-
hidden unit activations. A schematic representation of SAE tional parameter, ϕ as an encoder in the variational approxi-
is shown in Fig. (2(D)). mation model. VAEs apply the variation approach to latent
If the activation function of the hidden neurons is aj, the representation, learning as additional loss component train-
ing estimators, known as Stochastic Gradient Variational
average activation function of each hidden neuron j will then
Bayes (SGVB) and Autoencoding Variational Bayes
be given by:
(AEVB) [15]. It optimizes the parameter ϕ and θ for the
probabilistic encoder qØ (Z, X), which is an approximation
to the generative model pθ (Z, X) where z is the latent vari-
(4) able and x is a continuous or discrete variable. It aims to
maximize the probability of each x in the training data set un-
der the entire generative process. However, an alternative
The objective of sparse constraints is to minimize Pj so configuration of generative latent variable modeling rises to
that Pj = P, where P is a sparse constraint very close to 0, give deep generative models (DGMs) instead of the existing
such as 0.05. To enforce sparse constraints, a penalty term is assumption of symmetric Gaussian posterior [16].
added to the cost function, which penalizes j, de-weighting
Recently, Ilse and collaborators [114] extended the VAE
significantly from P. The penalty term is the Kullback-Lei- framework, named domain invariant variational autoencoder
bler (KL) divergence between Bernoulli random variables, (DIVA), to tackle the problem of domain generalization,
which can be calculated as [11, 12]. such as learning representation from a set of the previously
unseen domain. Considering a perfectly disentangled latent
space, it is hypothesized that there exists a latent subspace
(5) that is invariant to change, called domain invariant, d. The
DIVA considers three independent latent subspaces such as
domain (zd), class (zy), and residual variations (zx). As zd and
where N2 is the number of neurons in the hidden layers, zy are marginally independent, the model would learn repre-
and index j is summing over the hidden units in the network. sentations zy, which are invariant with respect to the domain
d. Finally, these three latent variants are used by a single de-
coder pθ (x|zd, zx, zy) to reconstruct x. A detailed discussion
(6) on DIVA can be found in [114].
A Tour of Unsupervised Deep Learning Current Medical Imaging, 2021, Vol. 17, No. 00 5
Table 2. Applications of autoencoders and their variants for medical image analysis.
SAE were used by Suk and Shen [24] and Suk et al. [25] en input vector x can take the latent feature representation h
to extract low-level, latent features from multimodal imag- and vice-versa. These are generative models that learn a
ing datasets of MCI, AD, and HC (healthy control subjects). probability distribution over the given input space and gener-
MRI, FDG-PET, and biological marker data from CSF were ates a new data point [45]. An illustration of a typical RBM
pre-trained on SAE followed by fusion of extracted latent is shown in Fig. (4(A)). In fact, RBMs are restricted ver-
features with multi-kernel SVM to yield the diagnosis. Su- sions of Boltzmann machines where neurons must form an
pervised fine-tuning of the whole network using the pre-- arrangement of bipartite graphs. Due to this restriction, pairs
trained parameters at the starting point has led to the consid- of nodes belonging to each of the visible and hidden nodes
erable accuracy of the output. Another variant of SAE, name- have a symmetric connection between them, and nodes
ly Deep Autoencoders (DAE), when used along with the within a group have no internal connections. This restriction
Hidden Markov Model (HMM), renders the diagnosis of makes RBM a more efficient training algorithm than the gen-
MCI based on rs-fMRI data as input [26]. Apart from its use eral case of the Boltzmann machine. Hinton et al. (2010)
in classification, prediction, and diagnosis of various brain [45] proposed a practical guide to train RBMs.
disorders, several studies use SAE for lesion classification in RBMs have been utilized in various aspects of medical
prostate cancer, texture, and density of breast cancer, detec- image analysis as classification, segmentation, or detection
tion of visceral organ’s location from DCE MRI images, and of objects from radio images or pathological data. Yoo et al.
so on [5, 27, 28]. (2014) [46] proposed a method for segmentation of multiple
Besides the autoencoders and their several variants men- sclerosis (MS) lesion in multi-channel 3D MRI images.
tioned in Table 1, there are many other variants of autoen- Huang et al. 2016 [47] applied the RBM for Blind Source
coders, although not prominently visible. However, they are Separation (BSS) for fMRI data instead of Independent Com-
recently proposed and their applications in various aspects ponent Analysis (ICA) to detect latent sources that differenti-
of medical image analysis are shown. For instance, Super- ate internal or functional interaction in the special brain re-
vised Switching Autoencoders (SSAs) proposed for gion. Cai et al. (2016) [48] forwarded a novel architecture
Alzheimer’s disease classification from the single-slice im- called Transformed Deep Convolutional Network (TDCN),
age and disease regional analysis [116], Spatiotemporal At- a multi-output Convolutional Restricted Boltzmann Machine
tention Autoencoder (STAAE) for attention deficit hyperac- (CRBM) implemented to fuse the learning features from dif-
tivity disorder (ADHD) classification [117], Disentangled ferent modalities, e.g., CT and MRI in an unsupervised man-
Autoencoder for cross-stain feature extraction [118], 3D resi- ner which improved the recognition of vertebra pattern of
dual autoencoder (3D ResAE) to model deep representations different species. Jaumard-Hakoun et al. (2016) [49] trained
of fMRI [119], and Dense Residual Convolutional Auto En- a deep autoencoder in two phases. In the first phase, Transla-
coder (DRCAE) for retinal blood vessels segmentation tional RBM (tRBM) learned the relationship between inputs
[120]. (US and contour) data and, in the second phase, it recon-
structed the contour of the tongue motion during a speech
3.2. Restricted Boltzmann Machines from US image. However, RBMs and their variant were
widely applied for classification problems. Cao et al. (2015)
Restricted Boltzmann Machines (RBMs) are a variant of [50] solved the classification of imbalanced data in breast
Markov Random Field (MRF), constituting single layer undi- mammogram based oversampling and semi-supervised learn-
rected graphical model with an input layer or visible layer x ing model for false-positive reduction. Zhang et al. (2016)
= (x1, x2...... xN) and a hidden layer h = {h1, h2, ….HM}. The [51] achieved a classification AUC of 93.40% to differenti-
connection between nodes/units is bidirectional, so each giv- ate between benign and malignant breast tumors. Van
A Tour of Unsupervised Deep Learning Current Medical Imaging, 2021, Vol. 17, No. 00 7
Fig. (4). Various unsupervised learning models (A) Restricted Boltzman Machine, (B) Deep Belief Network, (C) Deep Boltzman Machine,
(D) Generative Adversarial Network. (A higher resolution / colour version of this figure is available in the electronic copy of the article).
Tulder and de Bruijne (2016) [52] proposed the Convolution- When multiple RBMs are stacked hierarchically, an undi-
al RBM, which performed good in both training data and rected generative model is formed by the top two layers, and
lung texture classification. Mathews et al. (2018) [53] de- a directed generative model is formed by lower layers. (Fig.
monstrated single-lead ECG signals classification for detec- 4(B)) illustrates the structure of DBN. Hinton et al. [57] for-
tion of ventricular and supraventricular heartbeats. Pereira et warded an algorithm “wake-sleep” for unsupervised neural
al. (2017) [54] highlighted a combined RBM-Random forest networks. Further, Hinton et al. (2006a) [10] applied a fast
classifier which jointly correlated between features imaging learning algorithm based on layer-wise training procedure,
data for brain tumor segmentation and penumbra estimation. where lower layers learn low-level features and subsequent-
A brief account for the application of RBMs in medical im- ly higher layers learn high-level features.
age analysis is shown in Table 3.
DBN has received a lot of attention in the field of com-
3.3. Deep Belief Networks puter vision for success in object recognition. Still, it was
too expensive, like training 3D images due to large training
Deep Belief Networks (DBN) is a kind of deep neural parameters, so the limitation could not be overcome. Recent-
network proposed by Bengio (2009) [56]. It is a greedy lay- ly a new variant of DBN was introduced, namely, convolu-
er-wise unsupervised learning algorithm with several layers tional DBN (ConvDBNs) [58]. Other work [59] [60] demon-
of hidden variables [8]. Layer-wise unsupervised training strated a DBN architecture that learns low-dimensional mani-
helps the optimization and provides weight initialization for fold of brain 3D-MRI that detects different variations in mor-
better generalization. In fact, DBN is a hybrid single proba- phological changes and correlates with Alzheimer’s disease
bilistic generative model, like a typical RBM. To construct a parameters and Lesion Distribution in Multiple Sclerosis.
deep architecture like SAEs where AEs layers are replaced Plis et al. (2014) [61] and Pinaya et al. (2016) [62] evaluat-
by RBMs, DBN has one lowest visible layer V, representing ed deep learning methods to characterize brain networks in
the state of input data vector and a series of hidden layers h1, neurocognitive disorders like Huntington’s disease and vari-
h2, h3, . . . hL. The following function in DBN represents the ous brain region in Schizophrenia using DBN and super-
joint distribution of visible unit V, hidden layers hl (l = 1, vised fine-tuning whereas Ortiz et al. (2016) [63] classified
2…. L): Mild Cognitive Impairment (MCI) and AD for simpler diag-
nosis and treatment based on Automated Anatomical Label-
ing (AAL) which provides grey matter (GM) image from
each brain region. Carneiro et al. (2012) [64], Carneiro and
(8) Nascimento (2013) [65] built rigid and non-rigid derivatives
which are based on training classifier to automatic segmenta- tion from the research communities in the area of medical
tion of the left ventricle of the heart from ultrasound se- image analysis. Still, with the power of well-built modeling
quence data at the same time. Ngo et al. (2017) [66] intro- abilities of DBM, even with a small number of training data,
duced a combination of DBN and level set model to yield au- DBM might get its due place in the future. Suk et al. (2014)
tomated segmentation of the left ventricle from cardiac cine [74] proposed a high-level latent and combined feature repre-
MRI. Although DBN also trains for the classification of be- sentation and fusion of multimodal images from MRI and
nign and malignant lesions from prostate ultrasonography, PET using DBM, which compared the binary classification
images from other work [67] [68] classify Autism spectrum of Alzheimer's disease (AD) and have Mild Cognitive Im-
disorders (ASDs) in children using rs-fMRI and sMRI data pairment (MCI) disease. Cao et al. (2014) [75] developed a
based on Random Neural Network clustering. Li et al. deep DBM for next-generation medical image revival sys-
(2014) [85] developed a risk analysis framework for osteo- tem instead of content-based image retrieval (CBIR), which
porosis prediction based on heterogeneous electronic health learned joint density from the multimodal learning model to
records (EHRs). To improve the classification accuracy, dis- get relative missing modalities. Recently, Wu et al. (2018)
crete wavelet transform (DWT) was combined with a DBN [76] developed a frame-by-frame heart segmentation by em-
and applied to detect Glioblastoma tumors from MRI [122]. ploying a three-layered DBM using cine MRI images to find
Summary of DBN application for medical image analysis local and global features of heart contours. Summary of the
presented in Table 4. DBM applications for medical image analysis is shown in
Table 5.
3.4. Deep Boltzmann Machine
3.5. Generative Adversarial Network (GAN)
Deep Boltzmann Machine (DBM) is a robust deep learn-
ing model proposed by Salakhutdinov et al. (2009) [69] and Generative Adversarial Network (GAN) [77] is one of
Salakhutdinov et al. (2012) [70]. They stacked multiple the recent promising techniques for building flexible deep
RBMs in a hierarchal manner to handle ambiguous input ro- generative unsupervised architecture. Goodfellow et al.
bustly. (Fig. 4(C)) represents the architecture of DBM as a (2014) [77] proposed two models generative model G and
composite model of RBMs, which clearly shows how DBM Discriminative model D, where G capture data distribution
differs from DBN. Unlike DBN, DBM forms an undirected (pg) over real data t, and D estimates the probability of a
generative model combining information from both lower
and upper layers, which improves the representation power sample coming from training data (m) not from G. Dur-
of DBMs. Training of layer-wise greedy algorithm for DBM ing iterations, the backpropagation generator and the discrim-
[71, 72] is calculated by modifying the procedure of DBN. inator compete with each other. In the training procedure,
the probability of D is maximized. This framework func-
Salakhutdinov et al. (2015) [71] and Dinggang et al. tions like a mini-max two-player game. The value function
(2017) [73] presented a three-layer DBM where the DBM V (G, D) establishes following two-player mini-max game is
learns the parameters θ = {w1, w2} the values of the neighbor given by:
layer(s) and probabilities of visible and hidden units are com-
puted using a logistic sigmoidal function. The derivative of
log-likelihood of the observation (V) with respect to the mod- (10)
el parameter (θ) is computed as,
where D(t) represents the probability of t from data m
(9) and pdata is the distribution of real-world data. This model
seems to be stable and improved as pg = pdata. A typical archi-
where Edata[.] denotes data-dependent obtained from visi- tecture of GAN is depicted in Fig. (4(D)). In fact, these two
ble units and Emodel[.] denotes data-independence obtained adversaries, Generator and Discriminator, continuously bat-
from the model. tle during the processing of training. GAN has been applied
to generate samples of photorealistic images to visualize
In the case of preceding deep networks (e.g., RBM or new designs.
DBN), deep Boltzmann machine has not gained much atten-
The recent applications of GAN produce promising re- ROI or analysis [88-99]. Some of the further applications of
sults in many diverse fields of medical image analysis. Hu et cGAN include brain tumor segmentation from MRI [100],
al. (2017) [78] employed conditional generative adversarial tissue segmentation from Ultrasound [101], retinal image
networks (cGAN) to simulate ultrasound images at given 3D synthesis [98], and effectively generate high-quality and real-
spatial locations, while Bi et al. (2017) [79] designed Multi- istic coronavirus diseases 2019 (COVID-19) CT images
-channel GAN (M-GAN) to synthesize PET data which im- [130]. Zhang et al. (2020) [127] proposed a task-driven
prove the AUC of PET-based CAD system. Bi et al. (2018) GAN (TD-GAN) in order to achieve both synthesis and pars-
[80] used an improved GAN, which uses dual-path adver- ing simultaneously for unseen real X-ray images without
sarial learning for Fully Convolutional Network (FCN) any annotations requirement from the X-ray image domain.
based image segmentation of the region of interest (ROI) The TD-GAN model provides a promising average dice of
without any medical image specification. Iqbal and Ali 86%, which achieves a similar level of accuracy as super-
(2018) [81] and Canas et al. (2018) [82] proposed MI-GAN, vised training. Few other applications of GAN include a re-
which generates precise segmented images for the applica- current GAN (RNN-GAN) to mitigate data imbalance
tion of supervised learning of retinal images, retaining patho- problems in medical image semantic segmentation [128], a
logical quality. Wang et al. (2018) [83-87] proposed 3D-con- GAN with dual discrimination (DD-GAN) to improve the
ditional GAN synthesizing high-quality brain region PET im- recognition in skin lesion segmentation [129], and so on.
age from low-dose PET image for treatment of MCI disease.
In several applications of GAN and its variants, medical im- There are several recent reviews on GAN and its
age generation or augmentation of various modalities such variants in medical image analysis, including medical image
as CT, MRI, X-ray, and PET was carried out to maximize generation [102], mammogram analysis [123], and medical
the size of the training dataset with realistic, high-quality im- imaging and analysis [124-126]. A summary of the GAN ap-
ages and preserving all deciding features in synthetic images plications for medical image analysis is presented in Table
which largely impact on the performance of segmenting 6.
S. Tools/ Packages
Models/Methods Description Language/Technology URL
No. Name
Deep learning APIs for Java hav-
1 deeplearning4j Autoencoders ing an implementation of several Java https://deeplearning4j.org/
deep learning techniques
A scientific computing framework
with good support for machine
learning algorithms that puts
2 unsupundertorch7 Autoencoder, etc. GPUs first. Unsup package pro- Lua https://github.com/torch/torch7
vides few unsupervised learning al-
gorithms such as autoencoders,
clustering, etc.
MIT licensed deep learning frame-
work that runs on CPU or GPUs
https://github.com/andersbll/deeppy
3 DeepPy Autoencoders and implements autoencoders, in Python
http://andersbll.github.io/deeppy-website/
addition to other supervised learn-
ing algorithms
Build a stacked autoencoder in the
R environment for pre-training of
https://rdrr.io/cran/SAENET/man/SAENET.-
4 SAENET.train Stacked autoencoder feed-forward NN and dimension R package
train.html
reduction of features
Domain invariant
Implementation of DIVA, an ex- Python, Scikit-image, https://github.com/AMLab-Amsterdam/DIVA
8 DIVA variational autoen-
tension of variational autoencoder Scikit-learn s
coder (DIVA)
Deep belief network pre-train in
an unsupervised manner with https://rdrr.io/github/TimoMatzen/RBM/sr-
9 Dbn DBN R package
stacks of RBM, which in return c/R/DBN.R
fine-tuned DBN
Restricted Boltzmann machine,
10 darch DBN, RBM deep belief network implementa- R package https://github.com/maddin79/darch
tion
DBN, RBM, deep Implementation of RBM, DBN, https://cran.r-project.org/web/packages/deep-
11 deepnet R package
autoencoders deep-stacked autoencoders net/
DBN and other deep learning im-
12 Vulpes DBN Visual Studio https://github.com/fsprojects/Vulpes
plementation in F#
RBM and DBM are implemented
13 pydbm DBM/ RBM in Python for pre-learning or di- Python https://pypi.org/project/pydbm/
mension reduction
Simple RBM implementation in https://github.com/echen/restricted-boltzman-
14 RBM RBM Python
Python n-machines
RBM and its Implementation of RBM and its
15 xRBM Python https://github.com/omimo/xRBM
variants variants in Tensorflow
Unsupervised representation learn-
16 DCGAN.torch GAN ing using Deep Convolutional Lua https://github.com/soumith/dcgan.torch
GAN
A Tour of Unsupervised Deep Learning Current Medical Imaging, 2021, Vol. 17, No. 00 13
S. Tools/ Packages
Models/Methods Description Language/Technology URL
No. Name
Conditional Adversarial Networks
17 pix2pix GAN for Image-to-image translation syn- Linux Shell Script https://github.com/phillipi/pix2pix
thesizing from the image
Energy-based GAN equivalent to
probabilistic GANs produces high- https://github.com/eriklindernoren/PyTorch-
18 ebgan GAN Python
-resolution images GAN/tree/master/implementations/ebgan
[Abbreviations: ADNI: Alzheimer’s Disease Neuroimaging Initiative; ABIDE: Autism Brain Imaging Data Exchange; DICOM: Digital Imaging and Communications in Medicine;
BCDR: Breast Cancer Digital Repository; CIVM: Center for in vivo Microscopy; DDSM: Digital Database for Screening Mammography; DRIVE: Digital Retinal Images for Vessel
Extraction; IDA: Image & Data Archive; ISDIS: International Society for Digital Imaging of the Skin; NBIA: National Biomedical Imaging Archive; OASIS: Open Access Series of
Imaging Studies; TCGA: The Cancer Genome Atlas; TCIA: The Cancer ImagingArchive].
S.
Data Set Modalities Medical Condition Accessibility URL
No.
1 ABIDE MRI Autism spectrum disorder Open access http://fcon_1000.projects.nitrc.org/indi/abide/
2 ADNI MRI Alzheimer’s disease Paid http://adni.loni.usc.edu/data-samples/access-data/
3 BCDR Mammography Breast cancer Open access https://bcdr.eu/
Histology of the Embryonic
4 CIVM 3D-MRM Limited access http://www.civm.duhs.duke.edu/devatlas/
and Neonatal Mouse
5 DDSM Mammography Breast cancer Open access http://marathon.csee.usf.edu/Mammography/Database.html
A huge database of various
6 DermNet Photo dermatology Limited access http://www.dermnet.com/
skin diseases
A variety of medical images,
7 DICOM MRI, CT, etc. Open access https://www.dicomlibrary.com
videos, and signal files
Retinal blood vessel segmenta-
2D color images of
8 DRIVE tion to study diabetic retino- Open access http://www.isi.uu.nl/Research/Databases/DRIVE/download.php
the retina
pathy
An online resource for neuros-
9 IDA Open access https://ida.loni.usc.edu/
cience images
Dermoscopy,
10 ISDIS telemedicine, spec- Skin disease Paid https://isdis.org/
troscopy, etc.
Online database of medical im-
Variety of imaging
11 MedPix ages, teaching cases, and clini- Open access https://medpix.nlm.nih.gov
data
cal topics
A database of the National Can-
cer Institute proving medical Limited/ open ac-
12 NBIA CT, PET, MRI, etc. https://imaging.nci.nih.gov/
images of various conditions cess
and anatomical sites
Normal aging or mild to moder-
13 OASIS MRI and PET Open access http://www.oasis-brains.org/
ate Alzheimer's Disease
Collection of MRI, A multimodal image archive Limited/ open ac-
14 TCIA http://www.cancerimagingarchive.net/
CT, etc. for various types of cancer cess
Histopathology slide images
Histopathology slide
15 TCGA from sample portions of vari- Open https://cancergenome.nih.gov/
images
ous types of cancers
Prostate MR Image Segmenta-
16 PROMISE12 MRI Open https://promise12.grand-challenge.org/
tion
able in various programming language environments, includ- clustering algorithm will only work if the images fit into nat-
ing R, Python, Lua, Java, Matlab, and Linux shell script. In urals groups.
the last few years, several image datasets have been made
(iv) Not a common choice for medical image analysis:
available to the research communities which are used as the
Unsupervised learning is not a common choice for medical
benchmark datasets. Some of the tools/packages and medi-
image analysis. However, from the literature, it is revealed
cal image benchmark datasets are listed in Table 8 and 9, re-
that these (autoencoders and their variants, DBN, RBM,
spectively.
etc.) are mostly used to learn the hierarchy level of features
6. DISCUSSION, OPPORTUNITIES, AND CHAL- for classification tasks. It is expected that unsupervised learn-
LENGES ing will play a pivotal role in solving complex medical imag-
ing problems that are not only scalable to a large amount of
Medical imaging and diagnostic techniques are the most unlabeled data but also suitable for performing unsupervised
widely used approaches for the early detection, diagnosis, and supervised learning tasks simultaneously (Yi et al.,
and treatment of complex diseases. After significant ad- 2018).
vancements in machine learning and deep learning (both su-
pervised and unsupervised), there is a paradigm shift from (v) Development of patient-specific anatomical and
the manual interpretation of medical images by human ex- organ model: Anatomical skeletons play a crucial role in un-
perts such as radiologists and physicians to automated analy- derstanding diseases and pathology. Patient-specific anatomi-
sis and interpretation, called computer-assisted diagnosis cal model is frequently used for surgery and interventions.
(CAD). As unsupervised learning algorithms can derive in- They help to plan procedures, perform measurements for de-
sights directly from data, we can use them for data-driven de- vice surging, and predict the outcome of post-surgery com-
cision making. They are more robust, and hence, can be util- plexities. Hence, the algorithm needs to be developed to con-
ized as the holy grail of learning and classification struct a patient-specific anatomical and organ model from
problems. Furthermore, these models are also utilized for medical images.
other important tasks, including compression, dimensionali-
ty reduction, denoising, super-resolution, and some degree (vi) Heterogeneous image data: In the last two to three
of decision making. decades, more emphasis was given to well-defined medical
image analysis applications, where developed algorithms
Unsupervised learning and CAD are both in their infan- were validated on well-defined types of images with wel-
cy. Researchers and practitioners have many opportunities l-defined acquisition protocol. The algorithms are required,
in this area. Some of them are: (i) allows exploratory analy- which can work on more heterogeneous data.
sis of data, (ii)allows preprocessing for the supervised algo-
rithm when it is used to generate a new representation of da- (vii) Semantic segmentation of images: Semantic seg-
ta which ensures learning accuracy and reduces memory mentation is the task of complete scene understanding, lead-
time overheads, (iii) the recent development of cloud com- ing to knowledge inference from imagery. Scene understand-
puting, GPU-based computing, parallel computing and its ing is a core of computer vision problems with several appli-
cheaper cost allows big data processing, image analysis, and cations, including human-computer interaction, self-driving
execution of complex deep learning algorithms very easily, vehicles, virtual reality, and medical image analysis. For ins-
(iv) provides seamless and secured smart health care system tance, MRI prostate segmentation using deep supervised
at reasonable bandwidth speed [105-108]. Ghoneim et al. CNN and boundary-weighted neural network-based segmen-
contributed a system for medical image forgery detection us- tation was carried out earlier [108, 110]. In another study, a
ing support vector machine-based classifier [103]. bidirectional convolution targets all slices of 3D volume da-
ta instead of an individual slice [104]. The semantic segmen-
Some of the challenges and research directions are: tation of medical images with acceptable accuracy is still
(i) Difficult to evaluate whether the algorithm has challenging.
learned anything useful: Due to the lack of labels in unsu- (viii) Medical video transmission: Enabling 3D video
pervised learning, it is nearly impossible to quantify its accu- in recently adopted telemedicine and U-healthcare applica-
racy. For instance, how can we assess whether the K-means tions result in more natural viewing conditions and better di-
algorithm found the right clusters? In this direction, there is agnosis. Also, remote surgery can be benefited from 3D
a need to develop algorithms that can give an objective per- video because of the additional dimensions of depth. Howev-
formance measure in unsupervised learning. er, it is crucial to transmitting data-hungry 3D medical video
(ii) Difficult to select the right algorithm and hard- streams in real-time through limited bandwidth channels.
ware: Selection of the right algorithm for a particular type Hence, efficient encoding and decoding techniques for 3D
of medical image analysis is not a trivial task because the video data transmission are required.
performances of the algorithm are highly dependent on the (ix) Need extensive inter-organizational collabora-
types of data. Similarly, hardware requirement also varies tions: Inter-professional and inter-organizational collabora-
from problem to problem. tion is important for better functioning of the healthcare sys-
(iii) Will unsupervised learning work for me?: It is a tem, eliminating some of the pitfalls such as limited re-
frequently asked question, but its answer depends on the sources, lack of expertise, aging populations, and combating
problem at hand. In the image segmentation problem, the chronic diseases (Karam et al., 2017). Medical image-based
A Tour of Unsupervised Deep Learning Current Medical Imaging, 2021, Vol. 17, No. 00 15
CAD needs extensive inter-organizational collaborations Unsupervised learning algorithms derive insights direct-
among doctors, radiologists, medical image analysts, and ly from data and use them for data-driven decision making.
computational data analysts. Unsupervised models are more robust and they can be util-
ized as the holy grail of learning and classification
(x) Need to capitalize on the big medical imaging mar-
problems. These models are also used for other tasks includ-
ket: According to the IHS Market report (https://technolo-
ing compression, dimensionality reduction, denoising, su-
gy.ihs.com.), the medical imaging market has a total global
per-resolution, and some degree of decision making. There-
revenue of $21.2 billion in 2016, which is forecasted to
fore, it is better to construct a model without knowing what
touch $24.0 billion by 2020. According to WHO, the global tasks will be at hand and whether or not we would use repre-
population will rise from 12% to 22% from 2015 to 2050. sentation (or model). All in all, we can think of unsuper-
Population aging leads to an increased rate of chronic diseas- vised learning as a preparation (pre-processing) step for su-
es globally, and hence there is a need to capitalize on a big pervised learning tasks, where unsupervised learning of rep-
medical imaging market worldwide. resentation may allow better generalization of a classifier.
(xi) Black-box and its acceptance by health professio-
nals: Machine learning algorithms are boons which solve CONSENT FOR PUBLICATION
the problems earlier thought to be unsolvable. However, it Not applicable.
suffers from being “black-box”, i.e., how output arrives
from the model is very complicated to interpret. Particularly, FUNDING
deep learning models are almost non-interpretable but are
still used for complex medical image analysis. Hence, its ac- None.
ceptance by health professionals is still questionable.
CONFLICT OF INTEREST
(xii) Will technology replace radiologists? For the pro-
The authors declare no conflicts of interest, financial or
cessing of medical images, deep learning algorithms help to
otherwise.
select and extract important features and construct new ones,
leading to a new representation of images not seen before. ACKNOWLEDGEMENTS
For the image interpretation side, deep learning helps to iden-
tify, classify, and quantify disease patterns. It also allows the None.
measurement of predictive targets, makes predictive models,
and so on. So, will technology “replace radiologists”, or mi- REFERENCES
grate to “virtual radiologist assistant” in the near future? [1] Wani N, Raza K. Multiple Kernel-Learning Approach for Medical
Hence, the following slogan is quite relevant in this context: Image Analysis. 2018.
“Embrace it, it will make you stronger; reject it, it may make http://dx.doi.org/10.1016/B978-0-12-813087-2.00002-6
[2] Jabeen A, Ahmad N, Raza K. Machine Learning-Based State-of-
you irrelevant”. the-Art Methods for the Classification of RNA-Seq Data. Classifi-
In a nutshell, unsupervised learning is an open topic cation in BioApps 2018; 6: 133-72.
http://dx.doi.org/10.1007/978-3-319-65981-7_6
where researchers can make contributions by developing a [3] Bourlard H, Kamp Y. Auto-association by multilayer perceptrons
new unsupervised method to train the network (e.g., solve a and singular value decomposition. Biol Cybern 1988; 59(4-5):
puzzle, generate image patterns, image patch comparison, 291-4.
http://dx.doi.org/10.1007/BF00332918 PMID: 3196773
etc.). Re-thinking of creating a great unsupervised feature [4] Bengio Y, Courville A, Vincent P. Representation learning: a re-
representation (e.g., what is the object and what is the back- view and new perspectives. IEEE Trans Pattern Anal Mach Intell
ground?) is nearly analogous to the human visual system. 2013; 35(8): 1798-828.
http://dx.doi.org/10.1109/TPAMI.2013.50 PMID: 23787338
CONCLUSION [5] Shin HC, Orton MR, Collins DJ, Doran SJ, Leach MO. Stacked
autoencoders for unsupervised feature learning and multiple organ
Medical imaging is one of the important techniques for detection in a pilot study using 4D patient data. IEEE Trans Pat-
tern Anal Mach Intell 2013; 35(8): 1930-43.
the early detection, diagnosis, and treatment of complex dis- http://dx.doi.org/10.1109/TPAMI.2012.277 PMID: 23787345
eases. Interpretation of medical images is usually performed [6] Vincent P, Larochelle H, Lajoie I. Stacked denoising autoen-
by human experts such as radiologists and physicians. After coders: learning useful representations in a deep network with a lo-
the success of machine learning techniques, including deep cal denoising criterion. J Mach Learn Res 2010; 11: 3371-408.
[7] Litjens G, Kooi T, Bejnordi BE, et al. A survey on deep learning
learning and the availability of cheap computing infrastruc- in medical image analysis. Med Image Anal 2017; 42: 60-88.
ture through cloud computing, there has been a paradigm http://dx.doi.org/10.1016/j.media.2017.07.005 PMID: 28778026
shift in the field of computer-assisted diagnosis (CAD). [8] Bengio Y, Lamblin P, Popovici D, et al. Greedy layer-wise train-
Both supervised and unsupervised machine learning ap- ing of deep networks. Adv Neural Inf Process Syst 2006; 19:
153-60.
proaches are widely applied in medical image analysis, each [9] Hinton GE, Salakhutdinov RR. Reducing the dimensionality of da-
of them with their pros and cons. Due to the fact that human ta with neural networks. Science 2006; 313(5786): 504-7.
supervisions are not always available or can be inadequate http://dx.doi.org/10.1126/science.1127647 PMID: 16873662
or biased, therefore, unsupervised learning algorithms, in- [10] Hinton GE, Osindero S, Teh YW. A fast learning algorithm for
deep belief nets. Neural Comput 2006; 18(7): 1527-54.
cluding their deep architecture, provide hope for the future http://dx.doi.org/10.1162/neco.2006.18.7.1527 PMID: 16764513
with lots of advantages. [11] Ng A. Sparse autoencoder lecture notes 2013.web.stanford.edu/-
16 Current Medical Imaging, 2021, Vol. 17, No. 00 Raza and Singh
of breast tumors with shear-wave elastography. Ultrasonics 2016; 2018; 31(6): 895-903.
72: 150-7. http://dx.doi.org/10.1007/s10278-018-0093-8 PMID: 29736781
http://dx.doi.org/10.1016/j.ultras.2016.08.004 PMID: 27529139 [69] Salakhutdinov R, Hinton G. Deep Boltzmann machines.Artificial
[52] van Tulder G, de Bruijne M. Combining generative and discrimi- Intelligence and Statistics. PMLR 2009; pp. 448-55.
native representation learning for lung CT analysis with convolu- [70] Salakhutdinov R, Hinton G. An efficient learning procedure for
tional restricted boltzmann machines. IEEE Trans Med Imaging deep Boltzmann machines. Neural Comput 2012; 24(8):
2016; 35(5): 1262-72. 1967-2006.
http://dx.doi.org/10.1109/TMI.2016.2526687 PMID: 26886968 http://dx.doi.org/10.1162/NECO_a_00311 PMID: 22509963
[53] Mathews SM, Kambhamettu C, Barner KE. A novel application [71] Salakhutdinov R. Learning deep generative models. Annu Rev
of deep learning for single-lead ECG classification. Comput Biol Stat Appl 2015; 2: 361-85.
Med 2018; 99: 53-62. http://dx.doi.org/10.1146/annurev-statistics-010814-020120
http://dx.doi.org/10.1016/j.compbiomed.2018.05.013 PMID: [72] Goodfellow I, Mirza M, Courville A, Bengio Y. Multi-prediction
29886261 deep Boltzmann machines. Advances in Neural Information Pro-
[54] Pereira S, Meier R, McKinley R, et al. Enhancing interpretability cessing Systems 2013; 548-56.
of automatically extracted machine learning features: application [73] Dinggang S, Wu G. SukHeung-Il. Deep Learning in Medical Im-
to a RBM-Random Forest system on brain lesion segmentation. age Analysis. Annu Rev Biomed Eng 2017; •••: 19.
Med Image Anal 2018; 44: 228-44. [74] Suk H-I, Lee S-W, Shen D. Alzheimer’s Disease Neuroimaging
http://dx.doi.org/10.1016/j.media.2017.12.009 PMID: 29289703 Initiative. Hierarchical feature representation and multimodal fu-
[55] Nahid A-A, Mikaelian A, Kong Y. Histopathological breast-im- sion with deep learning for AD/MCI diagnosis. Neuroimage 2014;
age classification with restricted Boltzmann machine along with 101: 569-82.
backpropagation. Biomed Res (Aligarh) 2018; 29(10): 2068-77. http://dx.doi.org/10.1016/j.neuroimage.2014.06.077 PMID:
[56] Bengio Y. Learning Deep Architectures for AI. Found Trends 25042445
Mach Learn 2019; 2(1): 1-127. [75] Cao Y, Steffey S, He J, et al. Medical image retrieval: A multimo-
http://dx.doi.org/10.1561/2200000006 dal approach. Cancer Inform 2015; 13 (Suppl. 3): 125-36.
[57] Hinton GE, Dayan P, Frey BJ, Neal RM. The “wake-sleep” algo- PMID: 26309389
rithm for unsupervised neural networks. Science 1995; 268(5214): [76] Wu J, Ruan S, Mazur TR, et al. Heart motion tracking on cine
1158-61. MRI based on a deep Boltzmann machine-driven level set
http://dx.doi.org/10.1126/science.7761831 PMID: 7761831 method. Proceedings of IEEE Int Symp Biomed Imaging 2018;
[58] Lee H, Grosse R, Ranganath R, et al. Unsupervised learning of hi- 1153-6.
erarchicalrepresentations with convolutional deep belief networks. http://dx.doi.org/10.1109/ISBI.2018.8363775
Commun ACM 2011; 54(10): 95-103. [77] Goodfellow JP-A, Mirza M, Xu B, Warde-Farley D. Generative
http://dx.doi.org/10.1145/2001269.2001295 adversarial nets. Advances in Neural Information Processing Sys-
[59] Brosch T, Tam R. Manifold learning of brain MRIs by deep learn- tems 2014; 2672-80.
ing. Lect Notes Comput Sci 2013; 8150: 633-40. [78] Hu Y, Gibson E, Lee L-L, et al. Freehand ultrasound image simu-
http://dx.doi.org/10.1007/978-3-642-40763-5_78 lation with spatially-conditioned generative adversarial networks.
[60] Brosch T, Yoo Y, Li DKB, Traboulsee A, Tam R. Modeling the Lect Notes Comput Sci 2017; 10555: 105-15.
variability in brain morphology and lesion distribution in multiple http://dx.doi.org/10.1007/978-3-319-67564-0_11
sclerosis by deep learning. Lect Notes Comput Sci 2014; 8674: [79] Bi L, Kim J, Kumar A, et al. Synthesis of positron emission to-
462-9. mography (PET) images via multi-channel generative adversarial
http://dx.doi.org/10.1007/978-3-319-10470-6_58 networks (GANs). Lect Notes Comput Sci 2017; 10555: 43-51.
[61] Plis SM, Hjelm DR, Salakhutdinov R, et al. Deep learning for neu- http://dx.doi.org/10.1007/978-3-319-67564-0_5
roimaging: a validation study. Front Neurosci 2014; 8: 229. [80] Bi L, Feng D, Kim J. Dual-Path Adversarial Learning for Fully
http://dx.doi.org/10.3389/fnins.2014.00229 PMID: 25191215 Convolutional Network (FCN)-Based Medical Image Segmenta-
[62] Pinaya WHL, Gadelha A, Doyle OM, et al. Using deep belief net- tion. Vis Comput 2018; 34(6-8): 1043-52.
work modelling to characterize differences in brain morphometry http://dx.doi.org/10.1007/s00371-018-1519-5
in schizophrenia. Sci Rep 2016; 6: 38897. [81] Iqbal T, Ali H. Generative Adversarial Network for Medical Im-
http://dx.doi.org/10.1038/srep38897 PMID: 27941946 ages (MI-GAN). J Med Syst 2018; 42(11): 231.
[63] Ortiz A, Munilla J, Górriz JM, Ramírez J. Ensembles of deep http://dx.doi.org/10.1007/s10916-018-1072-9 PMID: 30315368
learning architectures for the early diagnosis of the Alzheimer’s [82] Canas K, Liu X, Ubiera B, et al. Scalable biomedical image synth-
disease. Int J Neural Syst 2016; 26(7) esis with GAN. ACM International Conference Proceeding Series.
http://dx.doi.org/10.1142/S0129065716500258 PMID: 27478060 [83] Jeyaraj P, Nadar ERS. Deep Boltzmann Machine Algorithm for
[64] Carneiro G, Nascimento JC, Freitas A. The segmentation of the Accurate Medical Image Analysis for Classification of Cancerous
left ventricle of the heart from ultrasound data using deep learning Region. Cognitive Computation and Systems 2019; 1(3): 85-90.
architectures and derivative-based search methods. IEEE Trans Im- http://dx.doi.org/10.1049/ccs.2019.0004
age Process 2012; 21(3): 968-82. [84] Lu N, Li T, Ren X, Miao H. A Deep Learning Scheme for Motor
http://dx.doi.org/10.1109/TIP.2011.2169273 PMID: 21947526 Imagery Classification based on Restricted Boltzmann Machines.
[65] Carneiro G, Nascimento JC. Combining multiple dynamic models IEEE Trans Neural Syst Rehabil Eng 2017; 25(6): 566-76.
and deep learning architectures for tracking the left ventricle endo- http://dx.doi.org/10.1109/TNSRE.2016.2601240 PMID:
cardium in ultrasound data. IEEE Trans Pattern Anal Mach Intell 27542114
2013; 35(11): 2592-607. [85] Li H, Li X, Ramanathan M, Zhang A. Identifying informative risk
http://dx.doi.org/10.1109/TPAMI.2013.96 PMID: 24051722 factors and predicting bone disease progression via deep belief net-
[66] Ngo TA, Lu Z, Carneiro G. Combining deep learning and level set works. Methods 2014; 69(3): 257-65.
for the automated segmentation of the left ventricle of the heart http://dx.doi.org/10.1016/j.ymeth.2014.06.011 PMID: 24979059
from cardiac cine magnetic resonance. Med Image Anal 2017; 35: [86] Mardani M, Gong E, Cheng JY, et al. Deep generative adversarial
159-71. neural networks for compressive sensing MRI. IEEE Trans Med
http://dx.doi.org/10.1016/j.media.2016.05.009 PMID: 27423113 Imaging 2019; 38(1): 167-79.
[67] Azizi S, Imani F, Ghavidel S, et al. Detection of prostate cancer http://dx.doi.org/10.1109/TMI.2018.2858752 PMID: 30040634
using temporal sequences of ultrasound data: a large clinical feasi- [87] Wang Y, Yu B, Wang L, et al. 3D conditional generative adver-
bility study. Int J CARS 2016; 11(6): 947-56. sarial networks for high-quality PET image estimation at low
http://dx.doi.org/10.1007/s11548-016-1395-2 PMID: 27059021 dose. Neuroimage 2018; 174: 550-62.
[68] Akhavan Aghdam M, Sharifi A, Pedram MM. Combination of rs- http://dx.doi.org/10.1016/j.neuroimage.2018.03.045 PMID:
fMRI and sMRI Data to Discriminate Autism Spectrum Disorders 29571715
in Young Children Using Deep Belief Network. J Digit Imaging [88] Liu Z, Bicer T, Kettimuthu R, Gursoy D, De Carlo F, Foster I. To-
18 Current Medical Imaging, 2021, Vol. 17, No. 00 Raza and Singh
http://dx.doi.org/10.2174/1573405614666191115102318 PMID: [127] Zhang Y, Miao S, Mansi T, Liao R. Unsupervised X-ray image
33059556 segmentation with task driven generative adversarial networks.
[124] Wolterink JM, Kamnitsas K, Ledig C, Išgum I. Generative adver- Med Image Anal 2020; 62
sarial networks and adversarial methods in biomedical image anal- http://dx.doi.org/10.1016/j.media.2020.101664 PMID: 32120268
ysis. arXiv preprint 2018. [128] Rezaei M, Yang H, Meinel C. Recurrent generative adversarial
network for learning imbalanced medical image semantic segmen-
[125] Yi X, Walia E, Babyn P. Generative adversarial network in medi- tation. Multimedia Tools Appl 2020; 79(21): 15329-48.
cal imaging: A review. Med Image Anal 2019; 58 http://dx.doi.org/10.1007/s11042-019-7305-1
http://dx.doi.org/10.1016/j.media.2019.101552 PMID: 31521965 [129] Lei B, Xia Z, Jiang F, et al. Skin lesion segmentation via genera-
[126] Kazeminia S, Baur C, Kuijper A, et al. GANs for medical image tive adversarial networks with dual discriminators. Med Image
analysis. Artif Intell Med 2020; 9 Anal 2020; 64
http://dx.doi.org/10.1016/j.artmed.2020.101938 http://dx.doi.org/10.1016/j.media.2020.101716 PMID: 32492581