Professional Documents
Culture Documents
Abstract— Improving image resolution, especially spatial only three convolutional layers but provides better
resolution, has been one of the most important concerns on performance than traditional methods. Improvement from
remote sensing research communities. An efficient solution for SRCNN has been actively explored by many researchers to
improving spatial resolution is by using algorithm, known as obtain better performance. For example, Kim et.al proposed
super-resolution (SR). The super-resolution technique that very deep super-resolution (VDSR) [4] that used residual
received special attention recently is super-resolution based on learning with more convolutional layer than SRCNN. Lim et
deep learning. In this paper, we propose deep learning approach al. proposed enhanced deep super resolution network (EDSR)
based on generative adversarial network (GAN) for remote by removing unnecessary modules in the residual block and
sensing images super resolution. We used residual dense
expanding the model size [5]. Tai et al. proposed recursive-
network (RDN) as generator network. Generally, deep learning
with residual dense network (RDN) gives high performance on
supervision and skip-connection to easy the difficulty of
classical (objective) evaluation metrics meanwhile generative training by the method called deep recursive convolutional
adversarial network (GAN) based approach shows a high network (DRCN) [6]. Zhang et al. proposed residual dense
perceptual quality. Experiment results show that combination network (RDN) that combine dense network with residual
of residual dense network generator with generative adversarial training to fully exploit the hierarchical features from all
network training is found to be effective. Our proposed method convolutional layers in the network [7].
outperforms the baseline method in terms of objective
Most of the super resolution methods aim to maximize
evaluation metrics and perceptual quality.
PSNR by minimizing pixel-wise mean squared errors (MSEs)
Keywords— convolutional neural network, generative between the super-resolved image and the target image.
adversarial network, remote sensing, image, residual dense However, measuring pixel-wise differences cannot capture
network, super-resolution perceptual differences between images easily, therefore
higher PSNR does not always produce a perceptually better
I. INTRODUCTION image [8]. To overcome this problem, Ledig et al. proposed
Remote sensing is the process of obtaining information generative adversarial network (GAN) based training for
about targeted objects or areas by measuring its reflected and super resolution, known as SRGAN (super resolution
emitted radiation from a distance. Remote sensing imaging generative adversarial network). Generative adversarial
can cover larger areas than other methods of telemetry data network (GAN) has been introduced by Goodfellow et al. [9]
acquisition but it has low spatial resolution and very low in to produce realistic fake images. It consists of two competing
relation to the dimensions of the sensed object. An effective neural networks, a generator and a discriminator. The
way to increase image spatial resolution at lower cost is by generator tries to generate a realistic image to deceive the
using algorithm based approach, known as super-resolution discriminator, while the discriminator tries to distinguish the
(SR). SR in remote sensing applications is important because generated images from the original. SRGAN employs GAN-
it can assist the visual interpretation of images in remote based training with perceptual loss function for obtaining
sensing application such as surveillance, target detection, visually-pleasing super resolved images rather than
agriculture, land use mapping, meteorology, etc. maximizing PSNR. Another GAN-based architectures and
training strategies have continuously improved to achieve
For many years ago, a number of traditional methods was better performance. Sajjadi et al. proposed EnhanceNet that
used for super-resolution. The simplest and fastest method for using a texture matching loss to encourage super-resolved
SR is interpolation, such as bilinear and bicubic interpolation. results to have the same textures as the ground truth images
Interpolation method projects the initial low resolution (LR) [10]. SRFeat [11] is another GAN-based method that
image onto a high resolution (HR) grid and the missing pixel produces better perceptual quality. It has two discriminators
values are estimated using an interpolation function [1]. and uses the adversarial loss terms in both image and feature
Learning based super-resolution, especially deep learning domains [11]. Another approach is ESRGAN that used
with convolutional neural network (CNN) architecture gets residual in residual dense block and used features before
considerable attention nowadays. The pioneering CNN model activation to improve perceptual quality [12].
for SR is proposed by Dong et.al, known as super-resolution
convolutional neural network (SRCNN) [2] [3]. It consists
In this research, we try to evaluate the implementation of A. Generative Adversarial Training
GAN-based approach for super resolution of remote sensing In this research, we used GAN-based super resolution
images. Different with SRGAN, our proposed method used approach to improve spatial resolution of remote sensing
residual dense network (RDN) [7] that based on residual images. GAN-based super resolution consist of two networks,
learning and dense network as GAN generator and we used generator and discriminator that compete each other. The
ten convolutional layers with different number of filters in diagram of training process can be seen of Fig 3.
discriminator network.
This paper is arranged as follows. In Section II, we
present the proposed method. Section III explains experiment
details. Results and discussion are given on Section IV and
Section V is the conclusion.
II. PROPOSED METHOD
The aim of single image super resolution (SISR) is to
estimate a super resolution image from low resolution (LR)
input image, such as depicted on Fig. 1. Super resolved (SR) Fig 3. GAN-based approach
image has higher resolution (sharper and bigger size) than low The generator network generate super resolution (SR)
resolution input image. image from low resolution (LR) input image. The SR image
obtained from the generator and the ground truth (HR image)
are fed into the discriminator to be assessed continuously until
the SR image and ground truth can’t be distinguished. We
update the parameters in the adversarial networks using
SRGAN loss function (LSRGAN) and repeat the process until
the optimization is finished. Loss function on this approach
uses perceptual loss function such as used on SRGAN [8]. It
Fig. 1. Single image super resolution consists of a content loss (𝐿 ) and an adversarial
loss (𝐿 ).
SISR method that proposed on this research is deep
learning based approach. There are two main phases on deep 𝐿 =𝐿 +𝐿
learning for super-resolution, i.e. training and prediction, such
as depicted on Fig. 2. On the training phase, deep learning Two type of content loss evaluated on this GAN-based
model is trained to analyze statistical relationship between the super resolution, the pixel-wise MSE loss (𝐿 ) and VGG19
low resolution (LR) and its corresponding high resolution
loss (𝐿 ) [8].
(HR) images from a dataset training. On prediction phase, the
trained SR model is then used to predict HR images from a set
of LR images. 𝐿 = ∑ ∑ 𝑌 (𝑎, 𝑏) − 𝑌 (𝑎, 𝑏)
, ,
𝐿 = ∑ ∑ ∅ , 𝑌(𝑎, 𝑏) − ∅ , 𝑌(𝑎, 𝑏)
, ,
V. CONCLUSION
In this paper, we proposed a GAN-based approach for
Fig. 7 Comparison of RDGAN with SRGAN remote sensing images super resolution. Residual dense
network (RDN) was used as generator to produce higher
resolution image from low resolution input image. The dense
B. Qualitative Result network with residual learning and skip connection leads to
Fig. 8 provides example of the super resolved images with better PSNR and SSIM of the super resolved image. Applying
scale factor of 4x on four testing data. We compared our VGG loss as content loss make this GAN based approach has
proposed method (RDGAN with VGG loss function) with better perceptual quality than other evaluated SR methods.
bicubic interpolation, SRCNN, VDSR, RDN, and SRGAN. Our proposed method outperforms SRGAN in terms of full
reference image quality assessments (PSNR and SSIM) and
perceptual quality evaluator (NIQE score) on four testing
datasets consistently.
ACKNOWLEDGMENT
The experiments on this research have been done using
facility of High Performance Computing (HPC) in Research
Center for Informatics, Indonesian Institute of Sciences
(LIPI).
REFERENCES
[1] R. Fernandez-beltran, P. Latorre-carmona, and F. Pla,
“Single-Frame Super-Resolution in Remote Sensing : a
Practical Overview,” International Journal of Remote
Sensing, vol. 38, no. 1, pp. 314–354, 2017.
[2] C. Dong, C. C. Loy, K. He, and X. Tang, “Learning a
Deep Convolutional Network for Image Super-
Resolution,” in Computer Vision – European
Conference on Computer Vision 2014, 2014, pp. 184–
199.
[3] C. Dong, C. C. Loy, and K. He, “Image Super-
Resolution Using Deep Convolutional Networks,” IEEE
Trans. Pattern Anal. Mach. Intell., vol. 38, no. 2, pp.
295–307, 2016.
[4] J. Kim, J. K. Lee, and K. M. Lee, “Accurate Image
Super-Resolution Using Very Deep Convolutional
Networks,” in IEEE Conference on Computer Vision and
Pattern Recognition (CVPR), 2016, pp. 1646–1654.
[5] B. Lim, S. Son, H. Kim, S. Nah, and K. M. Lee,
“Enhanced Deep Residual Networks for Single Image
Super-Resolution,” IEEE Computer Society Conference
Computer Vision Pattern Recognition Work, pp. 1132–
1140, 2017.
[6] Y. Tai, J. Yang, and X. Liu, “Image Super-Resolution
via Deep Recursive Residual Network,” in IEEE
Cionference on Computer Vision and Pattern
Recognition (CVPR), 2017, pp. 3147–3155.
[7] Y. Zhang, Y. Tian, Y. Kong, B. Zhong, and Y. Fu,
“Residual Dense Network for Image Super-Resolution,”
in 2018 IEEE/CVF Conference on Computer Vision and
Pattern Recognition, 2018, pp. 2472–2481.
[8] C. Ledig et al., “Photo-realistic single image super-
resolution using a generative adversarial network,” IEEE
Fig. 8 Visual results with upscaling factor 4x on four testing Conference on Computer Vision and Pattern
datasets (from top: Bdg, Indo, NWPU, and PatternNet dataset) Recognition (CVPR), pp. 105–114, 2017.
[9] I. J. Goodfellow, J. Pouget-abadie, M. Mirza, B. Xu, and Networks,” in The European Conference on Computer
D. Warde-farley, “Generative Adversarial Nets,” in Vision Workshops (ECCVW), 2018.
Neural Information Processing Systems (NIPS), 2014, [13] W. Zhou, S. Newsam, C. Li, and Z. Shao, “PatternNet:
pp. 2672–2680. A benchmark dataset for performance evaluation of
[10] M. S. M. Sajjadi, B. Scholkopf, and M. Hirsch, remote sensing image retrieval,” ISPRS Journal of
“EnhanceNet: Single Image Super-Resolution Through Photogrammetry and Remote Sensing, vol. 145, pp. 197–
Automated Texture Synthesis,” Proceeding of IEEE 209, 2018, doi: 10.1016/j.isprsjprs.2018.01.004.
International Conference on Computer Vision, pp. [14] G. Cheng, J. Han, S. Member, X. Lu, and S. Member,
4501–4510, 2017. “Remote Sensing Image Scene Classification :
[11] S. Park, H. Son, S. Cho, K. Hong, and S. Lee, “SRFeat : Benchmark and State of the Art,” Proceeding of IEEE,
Single Image Super-Resolution with Feature vol. 105, no. 10, pp. 1865–1883, 2017.
Discrimination,” in Proceedings of the European [15] A. Mittal, R. Soundararajan, and A. C. Bovik, “Making
Conference on Computer Vision (ECCV), 2018, pp. 439– a ‘completely blind’ image quality analyzer,” IEEE
455. Signal Processing Letter, vol. 20, no. 3, pp. 209–212,
[12] X. Wang, K. Yu, S. Wu, J. Gu, and Y. Liu, “ESRGAN : 2013.
Enhanced Super-Resolution Generative Adversarial