You are on page 1of 12

254 IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 41, NO.

2, FEBRUARY 2022

Hybrid Variation-Aware Network for


Angle-Closure Assessment in AS-OCT
Jinkui Hao , Fei Li , Huaying Hao, Huazhu Fu , Senior Member, IEEE, Yanwu Xu , Risa Higashita,
Xiulan Zhang, Jiang Liu , Senior Member, IEEE, and Yitian Zhao

Abstract — Automatic angle-closure assessment in Ante- the 3D iris surface from an AS-OCT sequence, and obtain
rior Segment OCT (AS-OCT) images is an important task the geometrical characteristics necessary to provide global
for the screening and diagnosis of glaucoma, and the most shape information. 2D AS-OCT slices and 3D iris rep-
recent computer-aided models focus on a binary classifica- resentations are then fed into our HV-Net to extract
tion of anterior chamber angles (ACA) in AS-OCT, i.e., open- cross-sectional appearance features and iris morphological
angle and angle-closure. In order to assist clinicians who features, respectively. To achieve similar results to those
seek better to understand the development of the spec- of dynamic gonioscopy examination, which is the current
trum of glaucoma types, a more discriminating three-class gold standard for diagnostic angle assessment, the paired
classification scheme was suggested, i.e., the classifica- AS-OCT images acquired in dark and light illumination
tion of ACA was expended to include open-, appositional- conditions are used to obtain an accurate characterization
and synechial angles. However, appositional and synechial of configurational changes in ACAs and iris shapes, using
angles display similar appearances in an AS-OCT image, a Variation-aware Block. In addition, an annealing loss
which makes classification models struggle to differentiate function was introduced to optimize our model, so as to
angle-closure subtypes based on static AS-OCT images. encourage the sub-networks to map the inputs into the more
In order to tackle this issue, we propose a 2D-3D Hybrid conducive spaces to extract dark-to-light variation repre-
Variation-aware Network (HV-Net) for open-appositional- sentations, while retaining the discriminative power of the
synechial ACA classification from AS-OCT imagery. Specifi- learned features. The proposed model is evaluated across
cally, taking into account clinical priors, we first reconstruct 1584 paired AS-OCT samples, and it has demonstrated its
superiority in classifying open-, appositional- and synechial
angles.
Manuscript received July 16, 2021; revised August 29, 2021; accepted
September 1, 2021. Date of publication September 6, 2021; date Index Terms — Anterior chamber angles, glaucoma,
of current version February 2, 2022. This work was supported in AS-OCT.
part by Zhejiang Provincial Natural Science Foundation under Grant
LZ19F010001, in part by the Youth Innovation Promotion Association I. I NTRODUCTION
Chinese Academy of Sciences (CAS) under Grant 2021298, and in part
by Ningbo 2025 S&T Megaprojects under Grant 2019B10033 and Grant
2019B10061. (Jinkui Hao and Fei Li contributed equally to this work.)
(Corresponding authors: Xiulan Zhang; Jiang Liu; Yitian Zhao.)
G LAUCOMA is a major cause of irreversible blindness
worldwide. The current gold standard for glaucoma diag-
nostic assessment is gonioscopic examination of the anterior
Jinkui Hao is with Cixi Institute of Biomedical Engineering, Ningbo segment angle (ACA). Depending on the configuration of ACA
Institute of Materials Technology and Engineering, Chinese Academy
of Sciences, Ningbo 315201, China, and also with the University after biometric examination, glaucoma can be classified into
of Chinese Academy of Sciences, Beijing 100864, China (e-mail: either of two subcategories: primary angle-closure glaucoma
haojinkui@nimte.ac.cn). (PACG) and primary open-angle glaucoma (POAG). As the
Fei Li and Xiulan Zhang are with the State Key Laboratory of
Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen Univer- most severe stage of primary angle-closure disease, PACG is
sity, Guangzhou 510275, China, and also with the Glaucoma Arti- the leading cause of blindness in Asian populations [1].
ficial Intelligence Diagnosis and Imaging Analysis Joint Research Epidemiological studies [2] have found that primary angle-
Lab, Guangzhou and Ningbo, China (e-mail: lifei.aletheus@gmail.com;
zhangxl2@mail.sysu.edu.cn). closure disease can be further divided into different sub-types:
Huaying Hao, Yanwu Xu, and Yitian Zhao are with Cixi Institute of primary angle-closure suspect (PACS), primary angle-closure
Biomedical Engineering, Ningbo Institute of Materials Technology and (PAC), and PACG.
Engineering, Chinese Academy of Sciences, Ningbo 315201, China,
and also with the Glaucoma Artificial Intelligence Diagnosis and Imaging PACS denotes an ACA with appositional angle-closure but
Analysis Joint Research Lab, Guangzhou and Ningbo, China (e-mail: in which permanent peripheral anterior synechial contact has
haohuaying@nimte.ac.cn; ywxu@ieee.org; yitian.zhao@nimte.ac.cn). not yet developed, as shown in Fig. 1 (B). PAC/PACG denotes
Huazhu Fu is with the Institute of High Performance Computing,
Agency for Science, Technology and Research, Singapore 117684 the appositional closure with elevated intraocular pressure
(e-mail: hzfu@ieee.org). (IOP), or the presence of peripheral anterior synechial contact
Risa Higashita is with Tomey Corporation, Nagoya 451-0051, Japan (occludable ACA), as shown in Fig. 1 (C). In addition, PACG
(e-mail: k-chen@tomey.co.jp).
Jiang Liu is with the Department of Computer Science and Engineering, displays optical neuropathy. The identification of appositional-
Southern University of Science and Technology, Shenzhen 518055, and synechial-angle closure is vital to guide clinicians in
China, and also with the Glaucoma Artificial Intelligence Diagnosis and developing personalized treatment plans: for example, laser
Imaging Analysis Joint Research Lab, Guangzhou and Ningbo, China
(e-mail: liuj@sustech.edu.cn). peripheral iridotomy might be used prophylactically for PACS,
Digital Object Identifier 10.1109/TMI.2021.3110602 so as to relieve angle closure [3]. In addition, this further

1558-254X © 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
HAO et al.: HV-NET FOR ANGLE-CLOSURE ASSESSMENT IN AS-OCT 255

2D slices of AS-OCT are of limited use in distinguishing


sub-types of angle-closure, probably because the 2D image
provides only a single cross-sectional slice view of the entire
anterior segment and, therefore, other slices are not taken into
account in determining the angle status [5]. On the other
hand, the appositional and synechial angle display similar
appearances in an AS-OCT image. Due to the lack of capa-
bilities identical to those of dynamic gonioscopy, it is difficult
to differentiate these two angle-closure subtypes from static
AS-OCT images. Fig. 1 shows the different types of ACAs.
Recent studies [6], [7] have established that the iris plays a
key role in the pathophysiology of the angle-closure disease.
For example, quantitative iris parameters are independently
associated with appositional ACA, and an anteriorly-bowed
iris may be associated with the degree of progression of
angle-closure [8]. Moreover, morphological changes in the iris
Fig. 1. (A)-(C) Image samples of open-appositional-synechial ACA surface are an important sign revealing and enabling the under-
grading. (D) Visual demonstration of manual annotations by gonioscopy standing of the pathogenesis of angle-closure glaucoma [9].
for one AS-OCT volume. ACA angles in the 12-1 o’clock region (red)
viewed with synechiae, and 3-4 o’clock (green) viewed with appositional Iris convexity is steeper in open-angle and appositional-
ACAs, and a sequence of AS-OCT images of a randomly-selected angle eyes, while those with a synechial-angle are flatter.
15◦ radial area. Iridotrabecular contact (ITC, as shown in (B) and (C)), Iris in eyes with angle-closure may have less stretch and
defined as the iris appearing to touch the anterior chamber angle at
the posterior pigmented trabecular meshwork, is a common presence of form a more convex structures with pupil dilation, increas-
angle-closure. ing the risk of apposition to the trabecular meshwork [10].
In addition, prior studies [11], [12] have demonstrated that
dynamic dark-light change of the anterior chamber angle
grading (PACS and PAC/PACG) will facilitate understanding under shirting conditions of illumination from weak to strong
of the progression of the spectrum of angle-closure glaucoma can be imaged and analyzed with AS-OCT, dark-to-light
types. changes of iris configurations and angle width are involved in
In clinical practice, the current clinical standard used for angle narrowing.
angle assessment is the observation of the ACA with the aid of Based on the above clinical priors, i.e., 1) iris shape is a
gonioscopy. Gonioscopic angle closure is defined as inability significant sign on understanding and revealing of the patho-
to visualize the pigmented trabecular meshwork (TM). Oph- genesis of different glaucoma types [8], [9]; 2) angle width and
thalmologists perform gonioscopy by quadrant or by clock iris configurations vary widely depending on illumination, and
hour (by clock hour in our work): ophthalmologists make an dynamic changes in iris configuration may be associated with
annotation for every 15◦ radial segment of the eye, while the risk and mechanism of angle-closure [11], [12], we aim to
moving the gonioscope counterclockwise, which accurately use 3D iris geometrical characteristics and dark-light change
describes the angle status in this 15◦ radial area. Fig. 1 (D) captured by AS-OCT to promote the distinction between the
shows an example of manual annotations. However, direct con- appositional and synechial ACA. 3D representation usually is
tact between the gonioscope and the eyes during examination more suitable than 2D slice for supporting global information
can be uncomfortable for the patient and may distort the ACA, analysis, as it provides neighborhood features. So we aim
leading to incorrect diagnostic results. Furthermore, gonio- to extract 3D characteristics by using paired-images that
scopic examination is time-consuming, subjective, and may are obtained under bright and dark illumination conditions
cause inter-observer and intra-observer variability. To this end, (Fig. 2). In this work, we introduce a hybrid deep learning
development of an automated approach is highly desirable. model that uses both the 2D and 3D representations to provide
Over the last decade, several methods have been proposed complementary information. It is worth noting that this is
for the classification of open-angle and angle-closure from a extension of our previous work [23], which only utilized
AS-OCT, including the recent Angle closure Glaucoma Evalu- 3D iris representation for the binary classification of open-
ation Challenge (AGE)1 [4]. AS-OCT, as a non-contact, non- angle and angle-closure. The main contributions of this work
invasive technique, can be used for cross-sectional views of include:
anterior segment structures(e.g, the iris and the ACA). Never- 1) We believe that this work to be the first attempt to classify
theless, almost all of the existing methods focus mainly on the anterior chamber angle into open, appositional and synechial
binary classification system of open-angle and angle-closure, subtypes by utilizing 3D iris representations. We designed a
while a more discriminative three-class approach (i.e., open-, 2D-3D Hybrid Variation-aware Network (HV-Net), and by
appositional- and synechial angles) has rarely been explored. using both 2D slices and 3D iris representations to obtain
On the one hand, due to the lack of spatial information, 2D cross-sectional appearance and 3D iris geometrical char-
acteristics, so as to assist the classification of the anterior
1 https://age.grand-challenge.org/ chamber angle into three subtypes.

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
256 IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 41, NO. 2, FEBRUARY 2022

discriminative feature representation extracted by deep learn-


ing technology. For example, Fu et al. [17] developed a
multi-context deep learning network for diagnosis of angle-
closure glaucoma, where imagery from different regions and
at a variety of scales is processed through parallel network
streams. Li et al. [20] first localized the scleral spur (SS)
points, and an ACA region centered on the detected SS was
then cropped as the input of the classification model with
a view to more accurate performance. Ferreira et al. [21]
employed multilevel convolutional neural networks based on
transfer learning to classify open and angle-closure glaucomas,
and achieved an AUC of 0.972. Xu et al. [28] utilized deep
learning classifiers to detect angle closure and primary angle-
Fig. 2. Example AS-OCT images acquired under dark and light closure disease (PACD). Amil et al. [18] extracted features
illumination conditions. The appositional angle under dark illumination in an unsupervised manner to order and classify AS-OCT
may open a particular angle when under light illumination conditions,
while the synechial angle will not due to progression to permanent
images. Fu et al. [19] proposed a Multi-Level Deep Network
contact. architecture using multiple parallel sub-networks to learn how
to construct multi-level representations.
2) We constructed a paired AS-OCT dataset obtained under The above methods all focus solely on analyzing structural
bright and dark illumination conditions. Here, a Variation- properties identified from 2D slices. In order to utilize the
aware Block (VAB) and an annealing loss were introduced to information of adjacent slices, Hao et al. [22] proposed a
obtain an accurate characterization of configurational changes sequence multi-scale aggregation network for classification of
in angle and iris shape during illumination changes. open-, appositional-, and synechial-ACAs, in which multiple
2D slices are used to generate AS-OCT sequences as the input
II. R ELATED W ORKS of the model, with the intention of acquiring temporal dynamic
In this section, we briefly review previous work investi- features. Williams et al. [29] presented a method to produce
gating AS-OCT imaging as a means to identify open-angle 3D maps of the anterior segment, so as to further assist
and angle-closure glaucoma, including both conventional and the diagnosis of angle-closure glaucoma using 3D features.
deep-learning-based methods. A summary of the existing methods for angle assessment using
Console et al. [24] developed an algorithm and measure- AS-OCT images is given in Table I.
ment method for the anterior chamber angle in AS-OCT
images. This method requires the user to input the locations III. P ROPOSED M ETHOD
of two scleral spurs: this made it impossible to analyze large- We introduce our method in detail in this section: iris
scale datasets automatically. Ni et al. [15] assessed angle- 3D reconstruction and quantification, the architecture of our
closure glaucoma by quantifying angle structure configurations classification network and a specific loss function.
via AS-OCT images, such as mean iris curvature and the
trapezoidal area of the iridocorneal angle. Tian et al. [13]
proposed a method to detect Schwalbe’s line and measure the A. 3D Iris Surface Reconstruction and Quantification
anterior chamber angle in AS-OCT images. Xu et al. [14] AS-OCT automatically scans 360◦ areas of the eye to
extracted the visual features of the ACA region from the obtain multiple consecutive radial slices that can be used to
AS-OCT image. These were then utilized to classify angle- reconstruct a 3D mesh of the iris surface, as shown in Fig. 3.
closure glaucoma. Fu et al. [16] introduced a data-driven Briefly, we first apply a segmentation network to obtain the
method to segment the cornea and iris prior to measuring initial boundary of the iris from successive AS-OCT slices.
the clinical parameters used to screen glaucoma via AS-OCT Then we apply a guided optimization method with Poisson-
images. disk sampling to reconstruct the 3D iris surface. Finally,
These conventional methods employ specific clinical several surface-based features are extracted to be used for
measurements, or handcrafted visual features (e.g., anterior subsequent angle-closure glaucoma detection. Fig. 3 illustrates
chamber area [16], iris curvature [15], or histograms of ori- the pipeline of this section.
ented gradients (HOG) [14]) to construct classifiers, which 1) Surface Reconstruction: A U-shaped network with
may yield sub-optimal diagnosis due to lack of essential wavelet refinement blocks, which can better identify the
information. Recently, convolutional neural networks (CNN)- iris boundaries in ITC regions, is first utilized to extract
based methods have been proposed to extract high-level an initial iris boundary. The automatic iris segmentation is
features in a data-driven manner for medical image clas- implemented by a well-trained model: see [23] for more
sification [25], and medical image segmentation [26], [27]. details. The detected upper boundaries are then used to gen-
For anterior chamber angles classification, several recent erate a point cloud of the iris, however, these point clouds
works have achieved promising performance, as a result of are sparse and non-uniform, resulting in a generated mesh

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
HAO et al.: HV-NET FOR ANGLE-CLOSURE ASSESSMENT IN AS-OCT 257

TABLE I
S UMMARY OF THE E XISTING M ETHODS FOR A NGLE A SSESSMENT U SING AS-OCT I MAGES

Fig. 3. Flowchart of the reconstruction and quantification of an iris surface in 3D. For (A) an AS-OCT sector within a radial region of 15◦ , a segmentation
network first extracts the boundaries of the iris (B). Then the segmentation results are converted into a 3D point cloud and a mesh is generated (C).
Afterwards, a constrained Poisson-disk sampling step is used to refine the coarse point cloud and mesh (D) to achieve a more accurate surface.
Finally, several surface measurements (E) are calculated for the subsequent diagnostic purposes.

that is coarse and lacking in local detail. As shown by the 2) Feature Extraction: The study has shown [9] that some
representative patch shown in Fig. 4 (A), the distortion of the quantitative indicators related to the iris (e.g., iris curvature),
triangular mesh lead to misrepresentation of the iris surface. are independently correlated with the severity of progres-
Additionally, due to the presence of iris frill, the geometric sion of angle-closure. Therefore, we calculated the several
variation is more dramatic in some regions than in others, curvature-related measurements after reconstructing the iris
resulting in a higher density of point cloud than in smooth surface for subsequent diagnostic classification of closed-angle
areas, which leads to a low-quality surface mesh, as shown disease, which including: principal curvatures (maximum and
in Fig. 3 (C). minimum), Gaussian curvature, mean curvature, and shape
To address this problem, we employ a constrained Poisson index [31].
disk sampling algorithm [30] for coarse meshes to refine the The shape index E of each point may be defined as
surface. This approach generates a more uniform and dense 2 C2 + C1
point cloud, while ensuring that objects of a given size will E = arctan , (1)
π C2 − C1
be distributed according to the sampling scheme and will not
overlap. Specifically, we use an adaptive radius r in practice to where C1 and C2 are the maximum and minimum curvatures
obtain a more accurate and at the same time as uniform point of a point, respectively, and E ∈ [−1, 1]. In contrast to the
cloud representation as possible: if the maximum curvature curvature, shape index is not affected by shape scaling and can
of a point is greater than the global average, r is set to r1 : yield a simple measure of local shape because it can represent
otherwise, it is specified as r2 . In our work, based on empirical flat, concave and convex regions [31].
tests, we set r1 and r2 to 6 and 10, respectively. Fig. 3 (D)
and Fig. 4 (B) show the point cloud and mesh after algorithm B. Network Architecture for ACA Assessment
optimization, which are more effective in characterizing geo- As shown in Fig. 5, our proposed Hybrid Variation-aware
metrical details. Network (HV-Net) consists of three components: 1) 2D feature

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
258 IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 41, NO. 2, FEBRUARY 2022

Fig. 4. Overview of 3D iris reconstruction and surface measurement process. (A) The original point cloud and mesh generated by our 2D iris
segmentation model. (B) The optimized point cloud and mesh. (C) Visualization of different surface measurements.

extraction sub-networks; 2) 3D feature extraction sub- A memory cell Ct is used to store the historical information.
networks; and 3) Variation-aware Block. Briefly, paired Wx∗ and Whi are 2D convolution kernels of the input and
AS-OCT representations in a 15◦ radial area are fed into hidden state. Bi-ConvLSTM makes decisions on the current
the 2D and 3D sub-networks (Section III-B.1 and III-B.2) input by processing the input data into forward and backward
to extract 2D cross-sectional appearance features and 3D iris dependencies, and is able to enhance the predictive power of
morphological features, respectively. In order to capture con- the model by analyzing the input from both directions [35].
figurational change in angle and iris shape, dark and light In the present work, the ConvLSTM module has one layer
features are integrated and processed by the Variation-aware with a kernel size of 3, and the number of hidden units is 64.
Block. The justification for using both a CNN and a ConvLSTM
The proposed HV-Net is detailed as follows. is that the CNN is good at learning local representations
1) 2D Sub-Networks: The inputs of 2D sub-networks are from spatial data, but lacks the ability to learn sequential
AS-OCT sequences with a size of n ∗ n ∗ T , n and T denoting correlations, while the ConvLSTM can be specialized for
the size of an AS-OCT slice and the number of the slice within sequential modeling, but is unable to extract features in
a 15◦ radial area, respectively. We first use a conventional a parallel manner. Their combination ensures that features
neural network (CNN) encoder to extract feature maps at from different slices are integrated efficiently and discrimina-
the slice level. The encoder can be any fully-convolutional tive cross-sectional appearance features are obtained. In our
network such as a ResNet [32]-based or VGG [33]-based implementation, the 2D sub-networks share weights across
network. We employed ResNet-34 [32], which retains the first light and dark branches to limit the number of learnable
four blocks as a encoder module. Then, a bi-directional convo- parameters.
lutional long short-term memory (Bi-ConvLSTM) module [34] 2) 3D Sub-Networks: The inputs of the 3D sub-networks
is employed to process the feature sequence from the encoder. are 3D iris representations and their corresponding measure-
Then a global average pooling layer is used to fuse these ments, as obtained in Section III-A. As in the case of the
features. The ConvLSTM can model temporal dependencies 2D sub-networks, dark and light branches share the weights
while preserving spatial information and contains less redun- in order to reduce the number of learnable parameters. Specif-
dancy for spatial data compared with fully-connected LSTM ically, to process the 3D point cloud representations of the
(FC-LSTM). ConvLSTM is defined as follows: iris and corresponding measurements, we apply the multi-
= σ (W layer perceptrons, followed by a max-pooling as a symmetric
it  xi ∗ Xt + Whi ∗ Ht −1 + Wci ◦ Ct −1 + bi ) , function. This module seems simple; however, it has been
ft = σ Wx f ∗ Xt + Wh f ∗ Ht −1 + Wc f ◦ Ct −1 + b f ,
proven to have interesting properties, and can achieve strong
Ct = ft ◦ Ct −1 + i t ◦ tanh (Wxc ∗ Xt + Whc ∗ Ht −1 + bc ) ,
performance in point cloud applications [36]. For convenience,
ot = σ (Wxo ∗ Xt + Who ∗ Ht −1 + Wco ◦ Ct + bo ) ,
we adopt the same architecture configuration with [36] as a
Ht = ot ◦ tanh (Ct ) , (2)
feature extractor, which consists of multi-layer perceptrons and
where ‘∗’ denotes the convolution operator, ‘◦’ denotes the affine transformation networks, and then aggregates features
Hadamard product, and σ is the sigmoid activation function. by max pooling. It should be noted that, this component can
Xt is the input tensor, and Ht is the hidden state tensor. i t , f t be replaced by any existing state-of-the-art pointcloud-based
and ot denote the input, forget and output gates, respectively. feature extractor.

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
HAO et al.: HV-NET FOR ANGLE-CLOSURE ASSESSMENT IN AS-OCT 259

Fig. 5. Illustration of our Hybrid Variation-aware Network (HV-Net), which includes three components: 1) 2D feature extraction sub-networks,
2) 3D feature extraction sub-networks, and 3) Variation-aware block. MLP: Multi-layer perceptron. BN: Batch Normalization. FC: Fully connected
layer. GAP: Global average pooling.

3) Variation-Aware Block: As introduced in Section I, appo- into space where the features best satisfies our hypothesis,
sitional and synechial angle display similar appearances in without affecting the discriminative power of the extracted
static AS-OCT images. To achieve a similar effect to that 2D and 3D features.
obtained by dynamic gonioscopy, we designed a Variation-
aware Block (VAB) using the paired dark and light feature C. Annealing Loss Function
representations to capture the configurational changes in the 1) Basic Loss Function: As shown in Fig. 5, we can
anterior segment angle and iris shape. Taking into account the obtain multiple outputs, i.e., 2D dark, light and variation
clinical priors, angle width and iris configurations may vary features, as well as 3D dark, light and variation features.
widely depending on illumination, and dynamic changes in iris We cascade all the outputs, and then add a full con-
configuration may be associated with the risk and mechanism nected layer followed by asoft-max layer. Specifically, let
of angle-closure [11], [12]. We hypothesized that, the dark-  N
to-light changes under variation of illumination intensity in X nD2 , X nL 2 , X nD3 , X nL 3 , Yn be a training set containing
n=1
cases of open and appositional angles would be significant, N samples, where X nD2 , X nL 2 denote 2D dark and light
whereas, the synechial angle would change little due to the sequences for n t h sample, X nD3 , X nL 3 denote 3D dark and
permanent contact typical of this diagnosis. To capture dark- light iris representations for n t h sample, and Yn ∈ {1, . . . , C}
to-light variations features - as shown in Fig. 5, the difference denotes the corresponding class label. W2d , W3d , and Wv
between dark and light features is first obtained by an element- are the parameters of the 2D, 3D sub-network and VAB,
wise subtraction across feature maps acquired from the prior respectively. Our basic cross-entropy loss function is defined
step. The fully connected (FC) layers with normalization (BN) as:
then are applied to the output. Thereafter, a sigmoid activation
is employed to obtain variation attention weights, which are L B (W2d,3d,v )
11
N C
used to perform multiplication with dark and light features.
The final output of the block is the concatenation of weighted =− δn,c log(P(Ŷn = c | X nD2 ,L 2 ,D3 ,L 3 ; W2d,3d,v )),
N C
dark and light features. n=1 c=1
To map the inputs, i.e., the 2D sequence and the 3D iris (3)
representation, reconstruction into the feature spaces that best where δn,c is a binary indicator of the ground truth class
satisfy our hypothesis, we apply a contrastive loss func- label, which equals 1 if Yn = c. P(· | ·) denotes the
tion [37] to optimize the sub-networks. For open and appo- probability obtained by our model, in terms of a given sam-
sitional angle samples, the Euclidean distances between dark ple (e.g., X nD2 ,L 2 ,D3 ,L 3 ) being diagnosed as a specific class
and light features are large, while the distances between dark (e.g., Ŷn = c). It is worth noting that the cascaded layers
and light features in synechial angle samples should be small. weight the various outputs and enables our model to learn the
Furthermore, we adopt an annealing strategy to map the inputs contributions of each type of features automatically.

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
260 IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 41, NO. 2, FEBRUARY 2022

2) Annealing Loss Term: Mapping the inputs into the feature training. As βα t  1, the second term encourages the sub-
space that best satisfies our hypothesis, changes under varying networks to map the inputs into a space that is more conducive
illumination will be significant in the case of open and to the extraction of dark-to-light variation features. When t
appositional angle and negligible in the case of synechial increases, this constraint gradually disappears. λ is the trade-
angle, is of importance to obtain accurate variation features. off parameter.
Accordingly, we apply a contrastive loss function [37], which
was first proposed for the face verification task, to optimize IV. E XPERIMENTS
the 2D and 3D sub-networks. Taking the 2D sub-network as
an example, thelabel Z n of A. Data Description
 each paired dark and light input
samples Sn2d = X nD2 , X nL 2 satisfies: A clinical dataset comprising images of 66 eyes with
• Z n = 0, if Sn is diagnostic of synechial angle. open-appositional-synechial angles (annotated by dynamic
• Z n = 1, if Sn is diagnostic of open or appositional angle. gonioscopic examinations) was established by Zhongshan
 Ophthalmic Center, China. All AS-OCT volumes were cap-
D2 L2
For each pair X n , X n in the training set, if Z n = 0, then tured along the perpendicular direction by the CASIA-2
update W2d to decrease Euclidean distance DW2d between the machine (Tomey Inc., Japan). Each AS-OCT volume contains
outputs (i.e., dark and light features) of the 2D parametric 128 slices, each with a resolution of 2144 × 1876 pixels.
sub-network G W2d ; if Z n = 1, then update W2d to increase To acquire dark and light data, AS-OCT imaging of each
DW2d . Here, the Euclidean distance DW2d may be defined as patient was performed first in the dark room (< 1 lux illu-
DW2d = G W2d (X nD2 ) − G W2d (X nL 2 ) 2 . mination by digital light meter). Prior to image acquisition,
The decrease and increase of Euclidean distances in the patients were allowed to undergo a 5-minute dark adap-
output space of 2D sub-network is obtained by minimizing tation. Images under illumination conditions were acquired
the contrastive loss function: at the 15th second after the room lights were turned on
(350-400 lux). It is worth noting that the paired data obtained

N
under light and dark illumination may have slight misalign-
LC (W2d ) = L(W2d , (Z , X D2 , X L 2 )n ), (4)
ment due to the movement of the subject, so we first perform
i=1
a global alignment of dark and light images using an iterative
where the L(·) is defined as diamond search strategy [38]. In practice, in our experiments,
we found that this operation has little effect on the classifi-
L(W2d , (Z , X D2 , X L 2 )n )
1 1 cation performance, suggesting that there may have been few
= (1 − Z ) (DW2d )2 + (Z ) max(0, m − DW2d ) . (5) misalignments in this dataset.
2 2
Particularly, in this in-house dataset, 13 eyes (PACS) are
where (Z , X D2 , X L 2 )n is n t h labeled 2D sample pair, and m diagnosed with appositional ACA, 21 eyes (POAG) are diag-
> 0 is a margin. Open or appositional angle pairs contribute nosed with open ACA, and 32 eyes (PAC/PACG) are with
to the loss function only if the distance DW2d is within this appositional ACA and closed ACA images. PACS is appo-
margin. The contrastive loss function of the 3D sub-network is sitional angle-closure, the eye is diagnosed with PAC/PACG
denoted as LC (W3d ). A direct way to fuse the basic loss and when the adhesive angle appeared. A senior ophthalmolo-
contrastive loss functions would be to weight them, as follows: gist with more than 30 years of experience in glaucoma
analysis performed the gonioscopic examination and annotated
LT ot al = u 1 ∗ L B (W2d,3d,v )+u 2 ∗ LC (W2d )+u 3 ∗ LC (W3d ), every 15◦ ACAs in one eye, with one eye yielding 24 annota-
(6) tions. Notably, appositional and synechial angles may appear
in one eye at the same time.
where u i denotes the corresponding weights (i = 1, 2, 3).
We formed a sequence of AS-OCT slices in each 15◦ region
However, we find that the classification performance of our
as the input of the 2D branch. Considering the computa-
model dropped when this loss was minimized (see Table IV).
tional load, all 2D sequences were cropped and resized to
This may be because the use of the contrastive loss func-
448 × 448 pixels by localizing the ACA regions using the
tion interferes with the learning of 2D and 3D features,
method described in [39]. Meanwhile, we performed the
making the extracted features less powerfully discriminative.
3D iris reconstruction and quantification to obtain the inputs of
To encourage the sub-networks to map the inputs into spaces
the 3D branch. Consequently, we obtained 1584 dark-and-light
more conducive to the extraction of dark-to-light variation
paired inputs, which include a total of 16,896 images. 504 of
representations, while retaining the discriminatory power of
which were annotated with open ACA (5376 images in total),
the learned features, we use an annealing strategy. Specifically,
742 samples were instances of appositional ACA (7915 images
we treat contrastive loss functions as an annealing term,
in total), and 338 samples were instances of synechial ACA
and slowly reduced it during training phase. The final loss
(3605 images in total).
is
LT ot al = λL B (W2d,3d,v ) + βα t (LC (W2d ) + LC (W3d )), (7) B. Evaluation Metrics
here β and 0 < α < 1 are hyper-parameters that controls the We employ several commonly-used metrics to evalu-
annealing schedule, and t is the number of iteration during ate the performance of multi-class classification: weighted

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
HAO et al.: HV-NET FOR ANGLE-CLOSURE ASSESSMENT IN AS-OCT 261

sensitivity (Sen), specificity (Spe), and balanced accuracy TABLE II


(B-Acc): T HE P ERFORMANCE OF O PEN -A PPOSITIONAL -
S YNECHIAL ACA S C LASSIFICATION

Nc
TNi c N
TPi
Spe = ai , Sen = ai ,
TNi + FPi TPi + FNi
i=1 i=1
(Sen+Spe)
B-Acc = ,
2
where TPi is true positives, TNi is true negatives, FPi is false
positives, and FNi is false negatives for the i-th classification
label; and ai represents the percentage of samples whose
ground truth labels are i . We also provide kappa analysis
and F-measure (F1 score) to reflect the trade off between Sen
and Spe.

C. Implementation Details
We evaluated our model in 5-fold cross-validation across
all 1584 samples for more reliable analysis over this relatively
small dataset. We split these samples into 5 subsets, and kept
the number of samples in each subset equal. For each fold,
we used 4 subsets for training and the remaining subset for
testing. It is worth noting that samples in each subset were
divided s based on the subject, so that two samples of one eye
would not be separated between the training and test sets.
The proposed networks were implemented using Python
based on the Pytorch package, and the PC we used contained
two GPUs (i.e, NVIDIA GeForce GTX 1080 Ti 11GB). The
Adam optimizer with recommended parameters was used to
optimize the model, and the batch size was set as 8. The
maximum epoch was 500. For the loss function, following
empirical testing we set β = 100 and α = 0.98, which caused Fig. 6. AUC curves of different deep-learning-based approaches in
the second term being phased out around t = 400. The margin classifying appositional and synechial ACA.
of the contrastive loss was set to 2. The trade-off parameter λ
was set to 1.5.
CNN are more powerful than visual features (e.g., HOG).
Sequence-based deep models achieve higher performances
D. Classification Performances than the slice-based deep models, which underlines the
The proposed HV-Net method was first compared a con- importance of the integration of neighborhood information.
ventional method [40] which was designed for ACAs clas- Compared with the state-of-the-art deep learning method
sification. This method first obtained the HOG feature, and (i.e., 2D and 3D methods), our proposed HV-Net demonstrated
linear SVM [41] was used to classify the ACA subtypes. competitive performance across all metrics. The main reason
We also compared the proposed method to the state-of- may be that the integration of 2D cross-sectional appear-
the-art deep methods by using 2D slices: VGG-16 [33], ance features, 3D iris morphologica features and dark-to-
Inception-V3 [42], ResNet-34 [32], Xception [43], and two light variation features is beneficial for improving diagnostic
specifically designed ACA type classification networks Multi- performance, as these different features can complement each
Level Deep Network (MLDN) [19], Multi-Scale Regions Con- other when distinguishing the types of ACA, especially of
volutional Neural Networks (MSDNN) [39]. In addition, as the appositional and synechial angles. Fig. 7 shows the confusion
2D branch of our model inputs the image sequence, this could matrices of the different methods, compared with MSDN, our
be treated as 3D data. Thus, the some 3D methods were also HV-Net has better performance in distinguishing appositional
applied as baselines: C3D [44], S3D [45], I3D [46] and Multi- and synechial angles.
Sequence Deep Network (MSDN) [22]. Note, MSDN is a It is worth noting that due to the limited condition, there
model specifically proposed for ACA grading. is a data imbalance between the appositional and synechial
1) Classification of the Open-Appositional-Synechial ACAs: angles in our dataset (742 for appositional ACA and 338 for
The classification performances of these methods are reported synechial ACA), in which the appositional ACA is the majority
in Table II. Deep learning methods have better classification class and the synechial ACA is the minority class. Deep
results than conventional methods, and by relatively large learning models trained on the imbalance dataset will exhibit
margins. This shows that, as high-level features are extracted bias towards the majority class and may ignore the minority
in a data-driven manner, representations learned by using class [47], so the algorithm tends to predict the synechial

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
262 IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 41, NO. 2, FEBRUARY 2022

Fig. 7. Confusion matrices of the different methods at sequence level. The numbers in the confusion matrices denote the percentage (above) and
number (below) of the predicted class. Notably, (A)-(C) are results of methods using 2D slices, while (D)-(E) are the results of 3D approaches, so the
total numbers of samples shown are different between them.

angle as the appositional angle, as shown in Fig. 7 (A) - (E). TABLE III
However, our model (see Fig. 7 (F)) performs better in dis- C OMPARISON OF D IFFERENT D EEP -L EARNING -B ASED M ETHODS
IN C LASSIFYING A PPOSITIONAL AND S YNECHIAL ACA
tinguishing synechial angles from appositional angles, which
shows that our algorithm has better robustness for imbalanced
data because of the explicit use of clinical prior.
2) Classification of the Appositional-Synechial ACAs: As
introduced in Section I, the classification of ACAs into open
and angle-closure is easier, and recent studies have shown
excellent performances (AUC ≥ 0.98) [19]. In consequence,
we only report the results in classifying of appositional and
synechial angles. As shown in Fig. 6, the proposed method
receive an AUC = 0.8587, outperforming other deep-learning-
based methods. Table III summaries the performance data of
these models.

E. Ablation Study
Our HV-Net employs three main components to form its in the F1 score. This indicates that the 3D branch provides
classification framework: 2D and 3D sub-networks, Variation- extra information, and the 3D iris morphological feature is
aware Block (VAB) and an annealing loss function, so we profitable for the differentiation of ACA sub-types, consistent
analyze the our model under different scenarios to validate with the conclusion of the epidemiological study.
the performance of each key component of our model. The 2) The Effectiveness of the VAB and Paired Data: As intro-
results of different combinations of these modules are reported duced in Section III-B.3, a key component of our proposed
in Table IV. method is Variation-aware Block designed to capture the dark-
1) The Effectiveness of the 3D Sub-Network: To explore the to-light variation features from the image pairs, with the
3D sub-network contributions, we used the 2D network only as intention of solving the problem of the similar appearance
the backbone. Note that, as shown in Table IV, a 3D network- of appositional and synechial angels in the static AS-OCT
only method is inferior to the backbone, yet a 2D + 3D method images. In this group of experiments, we evaluated the effec-
outperforms the backbone, with an improvement of about 1.4% tiveness of the VAB as well as the paired data. We first

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
HAO et al.: HV-NET FOR ANGLE-CLOSURE ASSESSMENT IN AS-OCT 263

TABLE IV
P ERFORMANCE C OMPARISONS OF A BLATION S TUDIES . W AND A D ENOTE W EIGHTED L OSS AND A NNEALING L OSS , R ESPECTIVELY

TABLE V V. D ISCUSSIONS AND C ONCLUSION


C OMPARISON OF MSDN M ETHOD AND O UR HV-N ET IN C LASSIFYING
A PPOSITIONAL AND S YNECHIAL A NGLE B ASED The conventional ACA classification approaches only focus
ON 15◦ AND 5◦ S EGMENT ACA on binary classification, i.e., open-angle and angle-closure.
In this study, we introduced a deep learning-based method
to identify open-appositional-synechial ACA sub-types from
AS-OCT imagery. However, the identification of ACAs into
three subtypes is not that straightforward, as two challenges
need to be tackled: 2D slices of AS-OCT lack sufficient spatial
information while determining angle status; and second, that
appositional and synechial angle display similar appearances
compared the 2D + 3D + VAB method with a 2D + 3D only in AS-OCT, which makes it difficult to accurately differentiate
method. In the results presented in Table IV, we found that angle-closure subtypes from static AS-OCT images. The main
the 2D + 3D + VAB method slightly improves classification contribution of our work is to incorporate clinical priors into
performance compared with the 2D + 3D only method. This the deep learning algorithm through 3D reconstruction and
is intuitive and reasonable, since the VAB is matched with quantification of the iris, as well as the explicit use of light-
the annealing term in the annealing loss function. Without dark paired data to stimulate the dynamic gonioscopy exam-
any constraints, it is difficult for the VAB to obtain precise ination, aiming to improve the accuracy of appositional and
weak-to-strong illumination variations that satisfy the clinical synechial ACA classification. The ablation study demonstrates
prior, i.e., the weak-to-strong illumination changes of open and that the effectiveness of our specific-designed components.
appositional angles would be significant, while the synechial
angle would change little due to the progression to permanent A. Limitations
contacts. Therefore, we applied a contrastive loss function to In this work, the dataset we used is still small to train
optimize the 2D and 3D sub-networks. such a complex model well, despite the data augmentation
Moreover, we then replaced paired data with only dark/light was applied during the training phase. We have verified the
data as inputs. The corresponding experimental results are effectiveness of each components through ablation studies,
presented in No.5 and No.6 in Table IV. We found that our model does not show significant advantages in classifi-
using dark/light data only achieved similar results to the cation performance when compared with the previous model,
2D + 3D method, which means taht the VAB and the i.e., MSDN, in classifying appositional and synechial ACA.
annealing term are non-contributing, due to the lack of paired This is most likely caused by the dataset being small, the great
data. similarity between the appositional and synechial ACA makes
3) The Effectiveness of the Annealing Loss Function: As it necessary to have more samples to allow the model to
introduced in Section III-C, we used an annealing loss learn more generalized knowledge, and we consider the limited
function to optimize our model. We verified the annealing training data as one of the major limitations of this work.
loss function’s effectiveness by comparing the 2D + 3D + However, we do believe our HV-Net would have significant
VAB + annealing loss method with 2D + 3D + VAB + improvement than the MSND if we train the models on a
weighted loss (i.e., Equation 6) method. We observed that relatively larger dataset. To this end, we added an additional
the annealing strategy improves classification more effec- experiments - with radial areas on every 5◦ ACAs of an
tively performance than the weighted loss function. In addi- eye, yielding 72 ACA segments for each single eye, resulting
tion, when compared with the 2D + 3D + VAB method, in total of 72 × 66 = 4, 752 samples with annotations.
the use of weighted loss makes the performance degradation. We generated label the same for all three consecutive 5◦ ACAs
This implies that the annealing strategy encourage the sub- as the original 15◦ ACAs. Then we train the HV-Net and
networks to map the inputs into the space more conducive MSDN models again on these augmented data. Our HV-Net
to the extraction of dark-to-light variation representations, achieves better performance than MSDN with relatively larger
while not affecting the discriminative power of the learned margin in terms of Acc, Sen and AUC (P < 0.0001, Delong’s
features. test), as illustrated in Table V, which demonstrates that our

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
264 IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 41, NO. 2, FEBRUARY 2022

model has potential in classification on the large-scale data.


It is worth noting that this additional experiment is only for
the validation of data augmentation purpose.

B. Analysis on Failure Cases


We further analyze the failed classification cases. Fig. 8
shows some failed appositional and synechial ACA classifi-
cation samples, by our HV-Net and MSDN. Based on our
observation on cases shown in Fig. 8, we found that proposed
model pays more attention to the dark-to-light change feature:
for some particular samples with appositional angles but have
not very obvious light and dark changes, they tend to be
classified as synechial angles by HV-Net. This leads to a
higher classification accuracy of HV-Net for synechial angles,
which can also be observed from Fig. 7 (E) and (F). It also
shows that our proposed use of the light-dark change as one
of the classification features is feasible, and it is reasonable to
believe that our model will be able to extract light-dark change
representations more accurately as more training samples are
added in the future work.
Fig. 8. Example of failed classifications between appositional and
C. Conclusion synechial angle closure cases of our HV-Net and MSDN. 0 and 1 repre-
sent appositional and synechial ACA respectively. Red (×) means wrong
In this work, we designed a specific approach and strategies classification, green () means correct classification.
to solve the above mentioned issues. We introduced a 2D-
3D Hybrid Variation-aware Network (HV-Net), that utilizes apply it in real clinical scenarios. At present, there is no
both 2D slices and 3D iris representations to obtain 2D cross- other way but gonioscopy to determine synechia, our work
sectional appearance and 3D iris geometrical characteristics, is the first attempt to explicitly use light-dark paired data
respectively. Here, a Variation-aware Block (VAB) is intro- to stimulate the dynamic gonioscopy examination, aiming to
duced to obtain an accurate characterization of configurational distinguish appositional and synechial ACAs, and ablation
changes in angle and iris shape. To obtain the geometrical experiments show the effectiveness of this idea, indicating that
characteristics of the iris and provide global shape information, this perspective is worth further exploration. We will include
we propose an automated framework for 3D iris surface more samples in the future to further improve accuracy and
reconstruction and quantification from AS-OCT imagery. Fur- generalizability with a view to future expansion to clinical and
thermore, we employ an annealing strategy to optimize the commercial applications. In addition, as shown in Table III,
model to encourage the sub-networks to map the inputs into these approaches have a low sensitivity-high specificity, which
more conducive spaces for the extraction of dark-to-light may stem from the similarity between two closed-angle groups
variation representations, while retaining discrimination of the and the data imbalance. For the current model, the synechial
learned features. angle tends to be misclassified into the appositional angle.
The paired-images acquired in different illumination condi- However, the system is used to assist the physician in the
tions were used in this study, and such data acquisition was diagnosis, and the physician can make further observations on
inspired by the clinical examination of dynamic gonioscopy, the appositional angle judged by the model to avoid the missed
so as to achieve a similar effect to that provided by dynamic diagnosis. In addition, physicians can also adjust the sensitivity
gonioscopy: the appositional angle will open a particular and specificity of the model by adjusting the threshold value
angle when under pressure, while the synechial angle will according to the actual needs.
not, due to progression to permanent contacts. We combined Eyes with a large PAS extent are at higher risk for pro-
multiple features (i.e., 2D, 3D, and dark-to-light variation gression and are also more likely to have associated optic
features) to empower the classification system the vital ability neuropathy [48], [49]. In future work, we will collect more
to distinguish between appositional and synechial angles. The clinical data, especially synechial samples, so as to optimize
effectiveness of our proposed method has been extensively the proposed network and further validate the generalization
evaluated using a dataset comprising imagery of 66 individual performance of our model on an external dataset from different
eyes, that is consisting of 1584 paired samples acquired under countries or devices.
different illumination conditions. Compared with several state-
of-the-art methods, our proposed method has demonstrated
better classification performance. R EFERENCES
Our algorithm achieves good results in the automatic clas- [1] P. Foster and G. Johnson, “Glaucoma in China: How big is the problem?”
sification of the open-appositional-synechial ACA sub-types Brit. J. Ophthalmol., vol. 85, no. 11, pp. 1277–1282, 2001.
in the private dataset, demonstrating the potential for clinical [2] P. J. Foster, R. Buhrmann, H. A. Quigley, and G. J. Johnson, “The def-
inition and classification of glaucoma in prevalence surveys,” Brit. J.
application. However, more exploration is still needed to Ophthalmol., vol. 86, no. 2, pp. 238–243, 2002.

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.
HAO et al.: HV-NET FOR ANGLE-CLOSURE ASSESSMENT IN AS-OCT 265

[3] R. Thomas, T. Arun, J. Muliyil, and R. George, “Outcome of laser [26] Y. Ma et al., “ROSE: A retinal OCT-angiography vessel segmentation
peripheral iridotomy in chronic primary angle closure glaucoma,” dataset and new model,” IEEE Trans. Med. Imag., vol. 40, no. 3,
Ophthalmic Surg., Lasers Imag. Retina, vol. 30, no. 7, pp. 547–553, pp. 928–939, Mar. 2021.
Jul. 1999. [27] Y. Zhao et al., “Automated tortuosity analysis of nerve fibers in
[4] H. Fu et al., “AGE challenge: Angle closure glaucoma evaluation in corneal confocal microscopy,” IEEE Trans. Med. Imag., vol. 39, no. 9,
anterior segment optical coherence tomography,” Med. Image Anal., pp. 2725–2737, Sep. 2020.
vol. 66, Dec. 2020, Art. no. 101798. [28] B. Y. Xu, M. Chiang, S. Chaudhary, S. Kulkarni, A. A. Pardeshi,
[5] H.-K. Cho, D. Ahn, and C. Kee, “Evaluation of circumferential angle and R. Varma, “Deep learning classifiers for automated detection of
closure using iridotrabecular contact index after laser iridotomy by gonioscopic angle closure based on anterior segment OCT images,”
swept-source optical coherence tomography,” Acta Ophthalmol., vol. 95, Amer. J. Ophthalmol., vol. 208, pp. 273–280, Dec. 2019.
no. 3, pp. e190–e196, May 2017. [29] D. Williams, Y. Zheng, P. G. Davey, F. Bao, M. Shen, and A. Elsheikh,
[6] Z. D. Soh, S. Thakur, S. Majithia, M. E. Nongpiur, and C.-Y. Cheng, “Reconstruction of 3D surface maps from anterior segment optical
“Iris and its relevance to angle closure disease: A review,” Brit. J. coherence tomography images using graph theory and genetic algo-
Ophthalmol., vol. 105, no. 1, pp. 3–8, Jan. 2021. rithms,” Biomed. Signal Process. Control, vol. 25, pp. 91–98, Mar. 2016.
[7] H. Yang, P. Yu, S. Cringle, X. Sun, and D.-Y. Yu, “Iridal vasculature [30] M. Corsini, P. Cignoni, and R. Scopigno, “Efficient and flexible sampling
and the vital roles of the iris,” J. Nature Sci., vol. 1, no. 8, p. e157, with blue noise properties of triangular meshes,” IEEE Trans. Vis.
2015. Comput. Graphics, vol. 18, no. 6, pp. 914–924, Jun. 2012.
[8] B. Wang et al., “Quantitative iris parameters and association with narrow [31] Y. Zhao et al., “Region-based saliency estimation for 3D shape analysis
angles,” Ophthalmology, vol. 117, no. 1, pp. 11–17, Jan. 2010. and understanding,” Neurocomputing, vol. 197, pp. 1–13, Jul. 2016.
[9] J. Huang, Z. Wang, Z. Wu, Z. Li, K. Lai, and J. Ge, “Comparison [32] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for
of ocular biometry between eyes with chronic primary angle-closure image recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.
glaucoma and their fellow eyes with primary angle-closure or primary (CVPR), Jun. 2016, pp. 770–778.
angle-closure suspect,” J. Glaucoma, vol. 24, no. 4, pp. 323–327, 2015. [33] K. Simonyan and A. Zisserman, “Very deep convolutional networks
for large-scale image recognition,” 2014, arXiv:1409.1556. [Online].
[10] C. Zheng et al., “In vivo analysis of vectors involved in pupil constriction
Available: http://arxiv.org/abs/1409.1556
in Chinese subjects with angle closure,” Investig. Ophthalmol. Vis. Sci.,
[34] S. Xingjian, Z. Chen, H. Wang, D.-Y. Yeung, W.-K. Wong, and
vol. 53, no. 11, pp. 6756–6762, 2012.
W.-C. Woo, “Convolutional lstm network: A machine learning approach
[11] C. K. S. Leung et al., “Dynamic analysis of dark-light changes of the
for precipitation nowcasting,” in Proc. NIPS, 2015, pp. 802–810.
anterior chamber angle with anterior segment OCT,” Investig. Ophthal-
[35] Z. Cui, R. Ke, Z. Pu, and Y. Wang, “Deep bidirectional and uni-
mol. Vis. Sci., vol. 48, no. 9, pp. 4116–4122, 2007.
directional LSTM recurrent neural network for network-wide traf-
[12] C. Y. L. Cheung et al., “Dynamic analysis of iris configuration with fic speed prediction,” 2018, arXiv:1801.02143. [Online]. Available:
anterior segment optical coherence tomography,” Investig. Ophthalmol. http://arxiv.org/abs/1801.02143
Vis. Sci., vol. 51, no. 8, pp. 4040–4046, 2010. [36] R. Q. Charles, H. Su, M. Kaichun, and L. J. Guibas, “PointNet:
[13] J. Tian, P. Marziliano, M. Baskaran, H. T. Wong, and T. Aung, Deep learning on point sets for 3D classification and segmentation,”
“Automatic anterior chamber angle assessment for HD-OCT images,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017,
IEEE Trans. Biomed. Eng., vol. 58, no. 11, pp. 3242–3249, pp. 652–660.
Nov. 2011. [37] R. Hadsell, S. Chopra, and Y. LeCun, “Dimensionality reduction by
[14] Y. Xu et al., “Automated anterior chamber angle localization and learning an invariant mapping,” in Proc. CVPR, 2006, pp. 1735–1742.
glaucoma type classification in OCT images,” in Proc. 35th Annu. Int. [38] J. Cheng et al., “Speckle reduction in 3D optical coherence tomography
Conf. IEEE Eng. Med. Biol. Soc. (EMBC), Jul. 2013, pp. 7380–7383. of retina by A-scan reconstruction,” IEEE Trans. Med. Imag., vol. 35,
[15] S. Ni Ni, J. Tian, P. Marziliano, and H.-T. Wong, “Anterior chamber no. 10, pp. 2270–2279, Oct. 2016.
angle shape analysis and classification of glaucoma in SS-OCT images,” [39] H. Hao et al., “Anterior chamber angles classification in anterior segment
J. Ophthalmol., vol. 2014, pp. 1–12, Aug. 2014. OCT images via multi-scale regions convolutional neural networks,”
[16] H. Fu et al., “Segmentation and quantification for angle-closure glau- in Proc. 41st Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. (EMBC),
coma assessment in anterior segment OCT,” IEEE Trans. Med. Imag., Jul. 2019, pp. 849–852.
vol. 36, no. 9, pp. 1930–1938, Sep. 2017. [40] Y. Xu et al., “Anterior chamber angle classification using multiscale
[17] H. Fu et al., “Multi-context deep network for angle-closure glau- histograms of oriented gradients for glaucoma subtype identification,”
coma screening in anterior segment OCT,” in Proc. MICCAI. Cham, in Proc. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc., Aug. 2012,
Switzerland: Springer, 2018, pp. 356–363. pp. 3167–3170.
[18] P. Amil et al., “Unsupervised feature extraction of anterior chamber [41] J. A. K. Suykens and J. Vandewalle, “Least squares support vector
OCT images for ordering and classification,” Sci. Rep., vol. 9, no. 1, machine classifiers,” Neural Process. Lett., vol. 9, no. 3, pp. 293–300,
p. 1157, Dec. 2019. Jun. 1999.
[19] H. Fu et al., “Angle-closure detection in anterior segment OCT based [42] X. Xia, C. Xu, and B. Nan, “Inception-v3 for flower classification,”
on multilevel deep network,” IEEE Trans. Cybern., vol. 50, no. 7, in Proc. 2nd Int. Conf. Image, Vis. Comput. (ICIVC), Jun. 2017,
pp. 3358–3366, Jul. 2020. pp. 783–787.
[20] P. Li, L. Geng, W. Zhu, F. Shi, and X. Chen, “Automatic angle-closure [43] F. Chollet, “Xception: Deep learning with depthwise separable convo-
glaucoma screening based on the localization of scleral spur in anterior lutions,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR),
segment OCT,” in Proc. IEEE 17th Int. Symp. Biomed. Imag. (ISBI), Jul. 2017, pp. 1251–1258.
Apr. 2020, pp. 1387–1390. [44] D. Tran, L. Bourdev, R. Fergus, L. Torresani, and M. Paluri, “Learning
[21] M. M. Ferreira, G. P. Esteve, G. B. Junior, J. D. S. de Almeida, spatiotemporal features with 3D convolutional networks,” in Proc. IEEE
A. C. de Paiva, and R. Veras, “Multilevel CNN for angle closure glau- Int. Conf. Comput. Vis. (ICCV), Dec. 2015, pp. 4489–4497.
coma detection using AS-OCT images,” in Proc. Int. Conf. Syst., Signals [45] S. Xie, C. Sun, J. Huang, Z. Tu, and K. Murphy, “Rethinking spatiotem-
Image Process. (IWSSIP), Jul. 2020, pp. 105–110. poral feature learning: Speed-accuracy trade-offs in video classification,”
[22] H. Hao et al., “Angle-closure assessment in anterior segment OCT in Proc. ECCV, 2018, pp. 305–321.
images via deep learning,” Med. Image Anal., vol. 69, Apr. 2021, [46] J. Carreira and A. Zisserman, “Quo vadis, action recognition? A new
Art. no. 101956. model and the kinetics dataset,” in Proc. IEEE Conf. Comput. Vis.
[23] J. Hao et al., “Reconstruction and quantification of 3D iris surface for Pattern Recognit. (CVPR), Jul. 2017, pp. 6299–6308.
angle-closure glaucoma detection in anterior segment OCT,” in Proc. [47] J. M. Johnson and T. M. Khoshgoftaar, “Survey on deep learning with
MICCAI. Cham, Switzerland: Springer, 2020, pp. 704–714. class imbalance,” J. Big Data, vol. 6, no. 1, pp. 1–54, Dec. 2019.
[24] J. Console, L. Sakata, T. Aung, D. Friedman, and M. He, “Quantitative [48] R. Thomas, “Five year risk of progression of primary angle closure
analysis of anterior segment optical coherence tomography images: suspects to primary angle closure: A population based study,” Brit. J.
The Zhongshan Angle Assessment Program,” Brit. J. Ophthalmol., Ophthalmol., vol. 87, no. 4, pp. 450–454, Apr. 2003.
vol. 92, no. 12, pp. 1612–1616, 2008. [49] K. K. Ramani, B. Mani, R. J. George, and V. Lingam, “Follow-up of
[25] Y. Zhao et al., “Retinal vascular network topology reconstruction and primary angle closure suspects after laser peripheral iridotomy using
ultrasound biomicroscopy and A-scan biometry for a period of 2 years,”
artery/vein classification via dominant set clustering,” IEEE Trans. Med.
Imag., vol. 39, no. 2, pp. 341–356, Feb. 2020. J. Glaucoma, vol. 18, no. 7, pp. 521–527, 2009.

Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on February 24,2022 at 07:44:18 UTC from IEEE Xplore. Restrictions apply.

You might also like