Professional Documents
Culture Documents
Visual Quality Driven
Visual Quality Driven
978-1-4799-7061-2/18/$31.00
2018
c IEEE
Fig. 1. Diagram of our two-stage learning. First, we fine-tune the CNN using ground-truth transmission maps, applying the
mean squared error loss in the training process. Second, we take the model and adapt it by including the image restoration
process (b). Finally, we perform new training in the network minimizing a loss function based on image quality metrics.
timated transmission maps, which they state it is the key to Thereby, we only need to estimate t and B to restore an
recover a clean scene. Applying their approach in out of wa- image in the underwater environment. The transmission map
ter images achieve impressive results but do not generalize to t can be estimated using the DehazeNet. Following the prior
underwater scenes. of Drews et al. [8], we can roughly estimate the background
Unlike the methods mentioned above, our approach does light as being the pixel in the degraded image I whose trans-
not rely on a large dataset. Our premise is based and assessed mission map value is the highest, limited by a constant t0 ,
by image quality metrics. This assumption allows us to obtain i.e., B = maxy∈{x|t(x)≤t0 } I(y). The value t0 is chosen as
results that are directly related to the human sense of quality the 0.1% highest pixel.
considering the conditions of underwater images.
Transmission map estimation. To adjust the network to
2. METHODOLOGY our purpose, we proposed to perform supervised training fol-
lowing the guidelines of Cai et al. [11]. It consists in using un-
Our methodology comprises two-phase learning. Initially, derwater images as input to the network and comparing its es-
we perform a supervised training by fine-tuning the De- timated transmission map to the ground-truth maps. The loss
hazeNet [11] to estimate the transmission map. Then, we function used in this stage is the mean squared error (MSE),
restore the input image according to an image formation defined as
model. In the second phase, we minimize a loss function n
composed of quality metrics to perform image restoration 1X 2
Lt (I, J) = (tI (x) − tJ (x)) , (3)
finally. Figure 1 illustrates the whole process. n x=1
Fig. 2. Four examples of images from Ancuti et al.’s dataset restored by five different approaches, including ours.
and Sowmya [1]. Yang and Sowmya showed that the UCIQE
Table 1. Edges conservation (larger is better).
is in accordance with the human visual assessment. Statistics
of chroma, contrast, and saturation are measures correlative Scene He Drews Ancuti Ours
to underwater image quality. The UCIQE value is given by ancuti1 0.0000 0.0000 0.0003 0.0855
ancuti2 0.0000 0.0000 0.0006 0.0192
U CIQE = c1 × σc + c2 × conl + c3 × µs , (10) ancuti3 0.0000 0.0000 0.0008 0.0009
galdran 0.0004 0.0011 0.0062 0.0280
where σc is the standard deviation of chroma, conl is the
contrast of luminance, µs is the average of saturation, and
c1 = 0.4680, c2 = 0.2745, c3 = 0.2576. Table 2. Visual quality using UCIQE metric (larger is better).
Another metric used in our evaluation is border integrity Scene Original He Tarel Drews Ancuti Ours
(Equation 7 in Section 2). It assumes that the restoration pro-
ancuti1 0.8366 1.8894 0.5132 2.7105 1.8259 0.5359
cess should not degenerate the most relevant borders of a de-
ancuti2 0.4490 1.2412 0.4546 0.9285 1.1200 0.5497
graded image. At the same time, it should preserve edges
ancuti3 0.7487 0.9458 0.5766 1.0913 1.0264 4.0781
continuity. Thus, such metric tells us the gain in border con- galdran 0.9045 4.7779 0.5776 1.4077 4.0371 6.9439
servation after a recovering process.
[3] Ke Gu, Guangtao Zhai, Weisi Lin, and Min Liu, “The [13] Jie Li, Katherine A. Skinner, Ryan M. Eustice, and
analysis of image contrast: From quality assessment to Matthew Johnson-Roberson, “Watergan: unsupervised
automatic enhancement,” IEEE Transactions on Cyber- generative network to enable real-time color correction
netics, vol. 46, no. 1, pp. 284–297, 2016. 1 of monocular underwater images,” IEEE Robotics and
Automation Letters, vol. 3, no. 1, pp. 387–394, 2018. 1
[4] Lintao Zheng, Hengliang Shi, and Shibao Sun, “Un-
[14] Risheng Liu, Xin Fan, Minjun Hou, Zhiying Jiang,
derwater image enhancement algorithm based on clahe
Zhongxuan Luo, and Lei Zhang, “Learning aggregated
and usm,” in Information and Automation (ICIA), 2016
transmission propagation networks for haze removal and
IEEE International Conference on, 2016, pp. 585–590.
beyond,” arXiv preprint arXiv:1711.06787, 2017. 1
1
[15] Raanan Fattal, “Single image dehazing,” ACM Trans-
[5] Emanuele Trucco and Adriana T. Olmos-Antillon, actions on Graphics (TOG), vol. 27, no. 3, pp. 72, 2008.
“Self-tuning underwater image restoration,” IEEE Jour- 2
nal of Oceanic Engineering, vol. 31, no. 2, pp. 511–519,
2006. 1 [16] Gershon Buchsbaum, “A spatial processor model for
object colour perception,” Journal of the Franklin insti-
[6] Ran Kaftory, Yoav Y. Schechner, and Yehoshua Y. tute, vol. 310, no. 1, pp. 1–26, 1980. 3
Zeevi, “Variational distance-dependent image restora-
tion,” in Computer Vision and Pattern Recognition [17] Curtis D. Mobley, Light and water: radiative transfer
(CVPR), 2007 IEEE Conference on, 2007, pp. 1–8. 1 in natural waters, Academic press, 1994. 3
[7] Erickson Nascimento, Mario Campos, and Wagner Bar- [18] Cosmin Ancuti, Codruta Orniana Ancuti, Tom Haber,
ros, “Stereo based structure recovery of underwa- and Philippe Bekaert, “Enhancing underwater images
ter scenes from automatically restored images,” in and videos by fusion,” in Computer Vision and Pattern
Computer Graphics and Image Processing (SIBGRAPI), Recognition (CVPR), 2012 IEEE Conference on, 2012,
2009 XXII Brazilian Symposium on, 2009, pp. 330–337. pp. 81–88. 3
1, 3 [19] Jean-Philippe Tarel and Nicolas Hautiere, “Fast visibil-
ity restoration from a single color or gray level image,”
[8] Kaiming He, Jian Sun, and Xiaoou Tang, “Single image
in Computer Vision, 2009 IEEE 12th International Con-
haze removal using dark channel prior,” IEEE Trans-
ference on, 2009, pp. 2201–2208. 3
actions on Pattern Analysis and Machine Intelligence
(TPAMI), vol. 33, no. 12, pp. 2341–2353, 2011. 1, 2