Professional Documents
Culture Documents
Zhao 2020
Zhao 2020
MULTI-VIEW MAMMOGRAMS
ABSTRACT
1051
Algorithm 1 CvAM
Input: Intermediate feature maps of 4 views in a screening mam-
mography exam, {FCC L , FL
M LO
, FCC
R , FR
M LO
} ∈ Rc×h×w .
for p = CC, M LO do
Calculate bi-lateral attention Ap ∈ R1×h×w for projection p
based on FL R
p and Fp according to Equation 1.
end for
for s = L, R do
Calculate bi-projection attention As ∈ Rc×1×1 for breast s
Fig. 3: General model architecture over multi-view mammogram
based on FsCC and FsM LO according to Equation 2.
classification
end for
for p = CC, M LO do
for s = L, R do
Finally, the feature map of each view are refined by element- Calculate the refined feature map Fs∗ p for side s and projec-
wise multiplication with its corresponding spatial and channel atten- tion p according to Equation 3.
tion maps plus 1: end for
end for
Fps ∗ = (1 + Ap ) ⊗ (1 + As ) ⊗ Fsp (3) Return: Refined feature maps FL CC∗
, FM
L
LO∗
, FCC∗
R , FM
R
LO∗
.
1052
Fig. 4: CAM [19] visualization results of baseline ResNet50 (second row) and CvAM-embeded ResNet50 (third row). The first rows are
input images and the lesions are marked by red circles.
Method single image CC-MLO fusion CC and MLO predictions are fused, the AUC is further boosted to
0.862. This performance gain by score fusion is less significant as
ResNet50 Baseline 0.812 0.836
in single-view approaches, as the inter-view information is already
ResNet50+feature concat[10] N.A. 0.830
used in the feature refinement process.
ResNet50+CBAM[3] 0.825 0.845
Network Visualization with CAM. Apart from high classifica-
ResNet50+CvAM (proposed) 0.855 0.862
tion accuracy, a ideal mammogram classification model should be
also capable of locating abnormalities even if no location annota-
Table 1: Comparison of AUCs of different approaches on DDSM tions were provided in the training phase. To understand the influ-
test set. The AUCs are calculated in both single-image and CC-MLO ence of CvAM, we apply class activation mapping (CAM) [19] to
score fusion mode. the trained ResNet50 model and its CvAM-embedded counterpart to
visualize the areas of the image which is most indicative of cancer.
Figure 4 shows the CAM visualizations for CC and MLO views of
scores are predicted for both CC and MLO views of a breast. Com- 4 breasts in the test set which contain malignant lesions, marked out
bining the prediction of CC and MLO views may increase perfor- in red circles in the first row. In case 1, a tumor is clearly visible
mance because each view contain unique information. We used a in a low-density breast. Both ResNet50 and the CvAM-embeded
simple approach of taking the average score of the two views, and model successfully located the lesion. In more complicated breast
report AUCs based on both single-image and score-fusion. For the images as case 2 and 3, the CAM of ResNet50 is able to located the
multi-view feature fusion approach only the fusion score is provided, lesion in one view but the activation in the other view widely spread
since the CC and MLO views have already been fused to generate a out. The CvAM-embedded model located the lesion in both views,
single score for a breast. showing the ability of exploiting bi-projection information. In case
4, some high-density regions of normal tissues were activated in both
views for ResNet50, while these false positives were suppressed in
3.4. Results the CvAM-embedded activation map. We conjecture that the bene-
fit comes from the bi-lateral attention module which references the
Quantitative Analysis. The AUCs of four compared approaches information in the other breast.
were evaluated based on single image and CC-MLO score fusion,
and is presented in Table 1. The ResNet50 baseline model achieved
a reasonable AUC of 0.812 in single image mode, compared to 4. CONCLUSION
the reported 0.77 AUC in [20] with the same dataset but using
AlexNet [21] as classification model. A simple CC-MLO score We presents a novel cross-view attention module (CvAM) which im-
fusion boosted the performance to 0.836, showing the importance of plicitly learns to focus on the cancer-related local abnormal regions
aggregating information between two views. The multi-view feature and highlighting salient features by exploring cross-view informa-
fusion approach obtained an AUC of 0.83, which is better than the tion among four views of a mammography exam. Experimental re-
single-image baseline but brings less improvement than the score- sults show that the CvAMs-embedded multi-view attention model
level fusion. By integrating attention mechanism, the single-view not only out-performs non-attention and single-view attention mod-
attention approach reaches AUCs of 0.825 and 0.845 in single-image els in classification accuracy, but also provides better lesion locating
and score fusion test respectively, showing moderate improvements ability. In future works, it is promising to extend the current work
over its non-attention baselines. Finally, the proposed multi-view by: 1) incorporating multi-scale modeling to further improve whole
attention model achieves the best single-image AUC of 0.853 of mammogram classification; 2) apply mammogram registration ap-
all compared approaches, which is 3% higher than the single-view proaches [22] to improve the spatial alignment of the left and right
attention model, on account of its ability to exploit both spatial and breast images , such that the proposed bi-lateral attention module
channel attention from multi-view training examples. When the may better learn the importance of asymmetric regions.
1053
5. REFERENCES [14] Lilei Sun, Junqian Wang, Zhijun Hu, Yong Xu, and Zhong-
wei Cui, “Multi-view convolutional neural networks for mam-
[1] Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, mographic image classification,” IEEE Access, vol. 7, pp.
Mark Johnson, Stephen Gould, and Lei Zhang, “Bottom-up 126273–126282, 2019.
and top-down attention for image captioning and visual ques-
[15] Alan Joseph Bekker, Hayit Greenspan, and Jacob Goldberger,
tion answering,” in Proceedings of the IEEE Conference on
“A multi-view deep learning architecture for classification of
Computer Vision and Pattern Recognition, 2018, pp. 6077–
breast microcalcifications,” in 2016 IEEE 13th International
6086.
Symposium on Biomedical Imaging (ISBI). IEEE, 2016, pp.
[2] Jianlong Fu, Heliang Zheng, and Tao Mei, “Look closer to see 726–730.
better: Recurrent attention convolutional neural network for
[16] Sergey Zagoruyko and Nikos Komodakis, “Paying more at-
fine-grained image recognition,” in Proceedings of the IEEE
tention to attention: Improving the performance of convolu-
conference on computer vision and pattern recognition, 2017,
tional neural networks via attention transfer,” arXiv preprint
pp. 4438–4446.
arXiv:1612.03928, 2016.
[3] Sanghyun Woo, Jongchan Park, Joon-Young Lee, and
[17] Inês C Moreira, Igor Amaral, Inês Domingues, António Car-
In So Kweon, “Cbam: Convolutional block attention mod-
doso, Maria Joao Cardoso, and Jaime S Cardoso, “Inbreast:
ule,” in Proceedings of the European Conference on Computer
toward a full-field digital mammographic database,” Academic
Vision (ECCV), 2018, pp. 3–19.
radiology, vol. 19, no. 2, pp. 236–248, 2012.
[4] Jianpeng Zhang, Yutong Xie, Yong Xia, and Chunhua Shen,
[18] P SUCKLING J, “The mammographic image analysis society
“Attention residual learning for skin lesion classification,”
digital mammogram database,” Digital Mammo, pp. 375–386,
IEEE transactions on medical imaging, 2019.
1994.
[5] Qingji Guan, Yaping Huang, Zhun Zhong, Zhedong Zheng,
[19] Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and
Liang Zheng, and Yi Yang, “Diagnose like a radiologist: At-
Antonio Torralba, “Learning deep features for discriminative
tention guided convolutional neural network for thorax disease
localization,” in Proceedings of the IEEE conference on com-
classification,” arXiv preprint arXiv:1801.09927, 2018.
puter vision and pattern recognition, 2016, pp. 2921–2929.
[6] Jo Schlemper, Ozan Oktay, Michiel Schaap, Mattias Heinrich,
[20] Sarah S Aboutalib, Aly A Mohamed, Wendie A Berg, Mar-
Bernhard Kainz, Ben Glocker, and Daniel Rueckert, “Attention
garita L Zuley, Jules H Sumkin, and Shandong Wu, “Deep
gated networks: Learning to leverage salient regions in medical
learning to distinguish recalled but benign mammography im-
images,” Medical image analysis, vol. 53, pp. 197–207, 2019.
ages in breast cancer screening,” Clinical Cancer Research,
[7] Rebecca L Siegel, Kimberly D Miller, and Ahmedin Jemal, vol. 24, no. 23, pp. 5902–5909, 2018.
“Cancer statistics, 2015,” CA: a cancer journal for clinicians,
[21] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton, “Ima-
vol. 65, no. 1, pp. 5–29, 2015.
genet classification with deep convolutional neural networks,”
[8] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, in Advances in neural information processing systems, 2012,
“Deep residual learning for image recognition,” in Proceed- pp. 1097–1105.
ings of the IEEE conference on computer vision and pattern
[22] Saskia van Engeland, Peter Snoeren, JHCL Hendriks, and Nico
recognition, 2016, pp. 770–778.
Karssemeijer, “A comparison of methods for mammogram reg-
[9] Michael Heath, Kevin Bowyer, Daniel Kopans, Richard istration,” IEEE Transactions on Medical Imaging, vol. 22, no.
Moore, and W Philip Kegelmeyer, “The digital database for 11, pp. 1436–1444, 2003.
screening mammography,” in Proceedings of the 5th inter-
national workshop on digital mammography. Medical Physics
Publishing, 2000, pp. 212–218.
[10] Gustavo Carneiro, Jacinto Nascimento, and Andrew P Bradley,
“Automated analysis of unregistered multi-view mammograms
with deep learning,” IEEE transactions on medical imaging,
vol. 36, no. 11, pp. 2355–2365, 2017.
[11] Li Shen, Laurie R Margolies, Joseph H Rothstein, Eugene
Fluder, Russell McBride, and Weiva Sieh, “Deep learning to
improve breast cancer detection on screening mammography,”
Scientific reports, vol. 9, 2019.
[12] Wentao Zhu, Qi Lou, Yeeleng Scott Vang, and Xiaohui Xie,
“Deep multi-instance networks with sparse label assignment
for whole mammogram classification,” in International Con-
ference on Medical Image Computing and Computer-Assisted
Intervention. Springer, 2017, pp. 603–611.
[13] Krzysztof J Geras, Stacey Wolfson, Yiqiu Shen, Nan Wu,
S Kim, Eric Kim, Laura Heacock, Ujas Parikh, Linda Moy,
and Kyunghyun Cho, “High-resolution breast cancer screening
with multi-view deep convolutional neural networks,” arXiv
preprint arXiv:1703.07047, 2017.
1054