Professional Documents
Culture Documents
Abstract— A new framelet-based random walks (RWs) method In recent decades, many SAR image fusion approaches
is presented for synthetic aperture radar (SAR) image fusion, have been proposed [1], [11]–[25]. The simplest approach
including SAR-visible images, SAR-infrared images, and Multi- is to calculate the average of all the original images [11],
band SAR images. In this method, we build a novel RWs model
based on the statistical characteristics of framelet coefficients but this approach reduces the contrast and sharpness of
to fuse the high-frequency and low-frequency coefficients. This the fused image. To resolve this problem, many methods
model converts the fusion problem to estimate the probability based on multi-scale transform (MST) have been presented,
of each framelet coefficient being assigned each input image. such as ratio of low-pass pyramid (RLP) [12], gradient
Experimental results show that the proposed approach improves pyramid (GP) [11], Laplace pyramid (LAP) [11], discrete
the contrast while preserves the edges simultaneously, and out-
performs many traditional and state-of-the-art fusion techniques wavelet transform (DWT) [1], [13], “à trous” wavelet trans-
in both qualitative and quantitative analysis. form (AWT) [14], [15], curvelet transform (CVT) [16], [17],
dual-tree complex wavelet transform (DTCWT) [18], nonsub-
Index Terms— Image fusion, framelet transform, random
walks, SAR image, infrared image, visible image. sampled contourlet transform (NSCT) [19], [20] and nonsub-
sampled shearlet transform (NSST) [21]. In addition, many
I. I NTRODUCTION other image fusion techniques have been developed. For
all input images were given labels. They focused on solving Compared to previous approaches, our method can obtain
a labeling problem to estimate the probability of each input better fused image.
image pixel being assigned each labeled input image. And the The remainder of this paper is organized as follows.
merged weights for image fusion can be calculated by these In Section II, we give a brief introduction of framelet transform
probabilities. and random walks. Section III presents a novel random walks
Paper [32] and [33] are spatial domain methods. They might approach for SAR image fusion in framelet domain. The
smooth the merged weights excessively, which is bad for experimental results are given in Section IV. In the end,
image fusion. In order to solve this problem, we transform the Section V summarizes the main conclusions of this paper.
input images to a multi-scale framelet domain by performing
nonsubsampled framelet transform (NFT). II. P RELIMINARIES
As a generalization of the orthogonal basis, the framelet
In this section, we shall briefly review some concepts and
relaxes the requirements of the orthogonality and linear inde-
principles with regard to framelet transform and random walks.
pendence to bring in redundancy [35]. Therefore, it is more
precise than wavelet in extracting the spatial information.
Moreover, translation invariance is a desirable property in A. Framelet Transform
image processing [36]. It can be achieved by nonsubsampling For simplicity, we only show the framelets in the univariate
in framelet decomposition. In addition, the framelet transform setting, and the bivariate framelets can be obtained by tensor
can afford the effective sparse representation of the images product of the univariate one.
which include rich textures. More importantly, this transform A tight frame of L 2 (R) can be defined by a countable
also has properties of fast decomposition and perfect recon- function subset of χ ⊂ L 2 (R), if
struction [37]. In recent years, it has been widely used in image
processing, such as image inpainting [38], image deblur- f = f, gg ∀ f ∈ L 2 (R) (1)
ring [35], image denoising [36], [39], pansharping [40], [41] g∈χ
and image sparse unmixing [42]. where f, g represents the inner product for any two func-
In our proposed method, we first perform NFT on the tions f, g ∈ L 2 (R).
input images to obtain their framelet coefficients. The inherent Given a collection of dilations and shifts of a finite set
speckle noise of SAR image affects the further processing. = {ψ 1 , . . . ψ p } ⊂ L 2 (R), it is defined as a wavelet system
Therefore, we give a speckle reduction method based on hard W () as follows:
thresholding the high-frequency coefficients. Then, a novel
random walks fusion model is built in pre-denoised framelet W () = {ψd,q
t
= 2d/2 ψ t (2d η − q) | t = 1, . . . , p, d ∈ Z,
domain. In this model, we can obtain the probabilities of q ∈ Z} (2)
framelet coefficients being assigned each input image by
finding a global optimal solution. Then, the fusion weights of When W () forms a tight frame, it’s called a wavelet tight
high-frequency and low-frequency coefficients are estimated frame, and each ψ t ∈ is named as a framelet.
by these probabilities, respectively. In the end, the fused image To construct the compactly supported tight frame system,
is reconstructed by employing the corresponding inverse NFT one first obtains a compactly supported scaling function φ ∈
on the merged framelet coefficients. L 2 (R) with a refinement mask
h 0 satisfying
In random walks model, we first employ to build a φ (2w) =
h 0 (w)
φ (w) (3)
weighted undirected graph to model the statistical character-
istics of framelet coefficients. For high-pass subband fusion, where φ can be obtained by the Fourier transform of φ and
the absolute high-frequency coefficient values are selected to
h0 is a 2π-periodic function with h 0 (0) = 1. h 0 (w) =
h (k)e −ikw , h is named as a low-pass filter.
measure the compatibility of the high-frequency coefficient k∈Z 0 0
node with the input image node, which is good for fusing the Then, for the given scaling function φ, a tight frame can
edge contour information of input images. For low-pass sub- be constructed by looking for a suitable set of framelets =
band fusion, we take neighborhood information into account. {ψ 1 , . . . ψ p }, which is denoted in the Fourier domain by
The local variance of low-frequency coefficient is proposed to t (2w) =
ψ h t (w)
φ (w), t = 1, 2, . . . , p (4)
construct compatibility function, which is helpful to improve
contrast and avoid artifacts. where theframelet masks h t are 2π-periodic functions. Here
The main contribution of this paper is to propose a new
h t (w) = k∈Z h t (k)e−ikw , h t is called a high-pass filter.
SAR image fusion model based on random walks in framelet The unitary extension principle (UEP) in [43] gives the
domain. This model provides a method for solving high- condition for W () to form as a tight frame system, i.e., the
pass or low-pass subband fusion problem that can be for- masks h0,
h1, . . . ,
h p satisfy
mulated as an optimization problem. It not only considers
the correlation of the framelet coefficients, but also takes
p
h 0 (w)
h 0 (w + γ π)+
h t (w)
h t (w + γ π) = δ(γ ), γ = 0, 1
the local characteristics of coefficients into account, which is
t =1
good for merging framelet coefficients. In addition, we provide (5)
both subjective and objective comparisons with traditional
and state-of-the-art methods on three different image sets. for almost all w ∈ R. Here δ(γ ) is a delta function.
YANG et al.: RWs FOR SAR IMAGE FUSION IN FRAMELET DOMAIN 853
Fig. 2. Graph used in RW. The blue nodes are unlabeled nodes and red
nodes are labeled nodes.
Fig. 4. Correlation of framelet coefficients. (a) Neighborhood correlation. (b) Scale correlation. (c) Subband correlation.
of framelet coefficients in the next subsection. Then, a new The correlation between the high-pass subbands is also
random walks model is built based on these characteristics. useful for random walks in framelet domain. In addition,
Furthermore, we give framelet coefficients fusion rule. In the we can also use (8) to capture the correlation between con-
end, the effect of noise in SAR images is discussed. nected high frequency coefficients (blue nodes) at the same
location in adjacent high-pass subbands, as shown in Fig. 4(c).
A. Statistical Characteristics of Framelet Coefficients
Let the source image be {Ir,c , r, c = 1, 2, . . . , M}, where r B. Random Walks Model in Framelet Domain
and c represent row and column, respectively. M is the size 1) Weighted Undirected Graphs in Framelet Domain:
of row and column. As mentioned in section II-A, given the Weighted undirected graph will play an important role in
J -scale NFT of an image with one low-pass filter and two random walks model.
high-pass filters, Hs,k,r,c denotes the (r, c)th high-frequency In this model, let I = {I 1 , I 2 , . . . , I K } denotes the set of
coefficient in the kth high-pass subband of the sth scale and input images and VY = {y1, y2 , . . . , y K } denotes the set of
L s,r,c denotes the (r, c)th low-frequency coefficient in the sth labeled nodes, where a labeled node yl ∈ VY is associated
scale. Where s = 1, 2, . . . , J, k = 1, 2, . . . , 8 and J is the with the lth input image I l ∈ I , l = 1, 2, . . . , K .
largest scale in the decomposition. For high-pass subbands, let V XHs,k = {x s,k,r,c
H } denotes the set
As mentioned in [39] and [44], we have the following of unlabeled nodes in the kth subband of sth scale. It should be
statistical properties of the framelet coefficients. Neighbor- noted that the superscript H represents the high-pass subband
ing framelet coefficients are often highly dependent – if in this paper.
a specific framelet coefficient is large or small, then its Then, we define the unlabeled set:
neighboring framelet coefficients are very likely to also be
J
8
large or small [44]. In other words, each framelet coeffi- V XH = V XHs,k = {x iH = x s,k,r,c
H
,
cient (black node) is related to its neighborhood (blue nodes), s=1 k=1
as shown in Fig. 4(a). i = 8M 2 (s − 1) + M 2 (k − 1) + (c − 1)M + r } (9)
According to [39], large or small framelet coefficients
incline to propagate across scales. That is to say each par- where i represents the spatial position of high-frequency
ent coefficient (red node) connects to its child coefficient coefficient in all high-pass subbands.
(blue node), as shown in Fig. 4(b). In this paper, we give the As shown in Fig. 5(a), we define a weighted undirected
definition of the correlation coefficient between transformed graph G H = (V H , E H ) based on the correlation of high-
subbands. It can express the correlation between connected frequency coefficients, where V H = VY ∪ V XH is the set of
coefficients in the same subband of adjacent scales. nodes. Let V H be arranged in a way that:
Let A = {ar,c } and B = {br,c } be the matrices
V H = {v 1H = y1 , v 2H = y2 , . . . , v KH = y K ,
(e.g., the transformed subbands). The correlation coefficient
between A and B is defined as (8). v KH+1= x 1H , . . . , v KH+i = x iH , . . . , v KH+8 J M 2 = x 8HJ M 2 } (10)
M M
c=1 [ar,c −A][br,c −B]
Furthermore, E H = E 1H ∪ E 2H ∪ E 3H ∪ E 4H is the set of
R(A, B) =
r=1
M M edges, where edges in E 1H are drawn between nodes in V XH
M M
c=1 [ar,c − A] c=1 [br,c − B]
2 2
r=1 r=1 and VY , edges in E 2H are drawn between nodes within V XHs,k
(8) and edges in E 3H are drawn between nodes in V XHs ,k and V XHs ,k ,
1 M M 1 M M
1 2
where A = M 2 r=1 c=1 ar,c , B = M 2 r=1 c=1 br,c s1 and s2 represent the adjacent scales. In addition, edges
YANG et al.: RWs FOR SAR IMAGE FUSION IN FRAMELET DOMAIN 855
Fig. 5. Graph used in RW based on framelet domain. The red and blue nodes are unlabled nodes and the green nodes are label nodes. (a) High-pass subbands.
(b) Low-pass subbands.
in E 4H are drawn between nodes in V XHs,k and V XHs,k , k1 and k2 Likewise, we can define the compatibility function for low-
1 2
represent the adjacent high-pass subbands. pass subbands as follows:
H
Each edge in E is given a weight to measure the compat- ⎧
⎪
⎨w1,m,n (v m , v n ) ∈ E 1
L L L L
ibility between the adjacent nodes in V H . L L L L
wm,n = w (v m , v n ) = w2,m,n (v m , v n ) ∈ E 2L (14)
L L L
Therefore, we can define the compatibility function for high- ⎪
⎩ L
pass subbands as follows: λw3,m,n (v mL , v nL ) ∈ E 3L
⎧ H
⎪ w (v mH , v nH ) ∈ E 1H where w1,m,n
L , w2,m,n
L and w3,m,n
L are weight functions defined
⎪
⎪ 1,m,n
⎪
⎪ on E 1 , E 2 and E 3 , respectively. λ is employed to balance
L L L
⎨w H (v mH , v nH ) ∈ E 2H
wm,n = w (v m , v n ) =
H H H H 2,m,n
(11) the weight w3,m,n
L . We will explain more details about these
⎪
⎪ αw3,m,n
H (v mH , v nH ) ∈ E 3H weights in the following subsection.
⎪
⎪
⎪
⎩ H 2) Compatibility Functions: In this paper, the function
βw4,m,n (v mH , v nH ) ∈ E 4H w1,m,n
H measures the compatibility of mth high-frequency
where w1,m,n
H , w2,m,n
H , w3,m,n
H and w4,m,n
H are weight functions coefficient node with nth input image node. The large absolute
high-frequency coefficient values in high-pass subbands cor-
defined on E 1 , E 2 , E 3 and E 4 , respectively. α and β are
H H H H
rected graph G L = (V L , E L ) for low-pass subbands, where are used to evaluate the compatibility between connected
V L = VY ∪ V XL is the set of nodes. Let V L be arranged in a high frequency coefficients. As previous random walks
way that: approaches [28]–[33], the Gaussian weighting function defined
by Euclidean distance could measure the compatibility
V L = {v 1L = y1 , v 2L = y2 , . . . , v KL = y K , between two adjacent pixels. But this function can’t effectively
v KL +1 = x 1L , . . . , v KL +i = x iL , . . . , v KL +J M 2 = x JL M 2 } (13) capture the geometric property of input images due to its
isotropy. Moreover, it has drawbacks in the treatment of
Moreover, E L = E 1L ∪ E 2L ∪ E 3L is the set of edges, where edges and contours. In our work, we apply the Gaussian
edges in E 1L are drawn between nodes in V XL and VY , edges weighting function in framelet domain. It not only measures
in E 2L are drawn between nodes within V XLs and edges in E 3L the compatibility between two connected framelet coefficients
are drawn between nodes in V XLs and V XLs . but also overcomes the shortcomings in the spatial domain.
1 2
856 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 27, NO. 2, FEBRUARY 2018
Fig. 6. Original images used in our experiments. (a) SAR-1. (b) VIS-1.
(c) SAR-2. (d) VIS-2. (e) SAR-3. (f) IR-3. (g) SAR-4. (h) IR-4.
(i) S-BAND-5-1. (j) S-BAND-5-2. (k) S-BAND-6-1. (l) S-BAND-6-2.
D
where Hs,k,r,c indicates denoised high-frequency coefficients B. Image Fusion Evaluation Metrics
of SAR image. Ts,k represents threshold in different high-pass In order to objectively assess the performances of different
subbands. image fusion approaches, many fusion evaluation metrics
In this paper, the computing method of the threshold Ts,k have been presented in recent years [47]–[53]. It is generally
is adaptive to different high-pass subband characteristics. known that only one evaluation metric can’t reflect the image
√ √ fusion quality in quantitative analysis. Therefore, to make
σs,k 2 ln N 2 2σs,k
Ts,k = √ = √ (35) a comprehensive evaluation for fused image, we adopt five
ln( N ) ln N popular fusion evaluation metrics, which are briefly introduced
where N denotes the number of high-frequency coefficients. as follows. For simplicity, we denote A and B as two original
The variance of noise σs,k is not known in framelet domain, images and indicate F as the fused image.
it is estimated by the median function. Fusion evaluation metrics Q W and Q E are proposed by
Piella and Heijmans [49] based on the universal image quality
medi an(|Hs,k |)
σs,k = (36) index(UIQI) [47], [48].
0.6745 The evaluation metric Q W is defined as
However, the blurred edges may be generated during the
pre-denoising process, which will affect the result of image QW = c(w)(λ(w)Q 0 (A, F|w)+(1−λ(w))Q 0 (B, F|w))
fusion. In future work, we will further study how to incorpo- w∈W
rate the noise item into the image fusion model to make the (37)
approach have a better robustness to noise.
where the weight c(w) is the normalized saliency of w among
Based on the above discussion, we summarize the presented
all the local sliding windows. Q 0 (A, F|w) and Q 0 (B, F|w)
approach for image fusion in Algorithm 1.
are computed by the proposed method in [47] in a local sliding
window w. The local weight λ(w) is computed by
IV. E XPERIMENTAL R ESULTS AND C OMPARISON
s(A|w)
A. Original Images λ(w) = (38)
s(A|w) + s(B|w)
In this paper, we assume that there are only two input
images I 1 and I 2 . As shown in Fig. 6, six groups of original where the saliency measures s(A|w) and s(B|w) are computed
images divided into three categories are applied to verify the with the variance of A and B in a local sliding window w,
effectiveness of the presented image fusion approach. Among respectively.
them, there are two groups of SAR-Visible (S-V) images The evaluation metric Q E is defined as follows [49]:
(see Fig. 6(a-d)), two groups of SAR-Infrared (S-I) images
Q E = Q W (A, B, F)Q W (A , B , F )τ (39)
(see Fig. 6(e-h)) and two groups of Multi-band SAR (M-S)
images (see Fig. 6(i-l)). Furthermore, we assume that the two where A , B , and F are the corresponding edge images of
original images in each group have been registered. A, B, and F, respectively. τ is a parameter that represents
YANG et al.: RWs FOR SAR IMAGE FUSION IN FRAMELET DOMAIN 859
Fig. 7. SAR-VIS-1 fused images. (a) LAP. (b) RLP. (c) DWT. (d) DTCWT. (e) CVT. (f) NSCT-P. (g) NSCT-S. (h) NSST-M. (i) GF. (j) GRW. (k) NFT.
(l) Our.
Fig. 8. SAR-VIS-2 fused images. (a) LAP. (b) RLP. (c) DWT. (d) DTCWT. (e) CVT. (f) NSCT-P. (g) NSCT-S. (h) NSST-M. (i) GF. (j) GRW. (k) NFT.
(l) Our.
the contribution of the edge images compared to the source The evaluation metric Q Y based on structural similarity
images. index measure (SSIM) is defined as [52]:
The evaluation metric Q G based on gradient assesses the ⎧
performances of edge information transferred from the original ⎪
⎪λ(w)SS I M(A, F|w)+(1−λ(w))SS I M(B, F|w),
⎪
⎨i f SS I M(A, B|w) ≥ 0.75
images to the fused image [50]. It is computed by
QY = (42)
⎪
⎪max{SS I M(A, F|w), SS I M(B, F|w)},
i, j (Q
AF (i, j )w A (i, j ) + Q B F (i, j )w B (i, j )) ⎪
⎩
QG = (40) i f SS I M(A, B|w) < 0.75
i, j (w (i, j ) + w (i, j ))
A B
Fig. 9. SAR-IR-1 fused images. (a) LAP. (b) RLP. (c) DWT. (d) DTCWT. (e) CVT. (f) NSCT-P. (g) NSCT-S. (h) NSST-M. (i) GF. (j) GRW. (k) NFT.
(l) Our.
Fig. 10. SAR-IR-2 fused images. (a) LAP. (b) RLP. (c) DWT. (d) DTCWT. (e) CVT. (f) NSCT-P. (g) NSCT-S. (h) NSST-M. (i) GF. (j) GRW. (k) NFT.
(l) Our.
Fig. 14. Multi-band SAR-1 fused images. (a) LAP. (b) RLP. (c) DWT. (d) DTCWT. (e) CVT. (f) NSCT-P. (g) NSCT-S. (h) NSST-M. (i) GF. (j) GRW.
(k) NFT. (l) Our.
Fig. 15. Multi-band SAR-2 fused images. (a) LAP. (b) RLP. (c) DWT. (d) DTCWT. (e) CVT. (f) NSCT-P. (g) NSCT-S. (h) NSST-M. (i) GF. (j) GRW.
(k) NFT. (l) Our.
In order to balance the evaluation metrics, we suggest using The analysis of parameters α, β and λ are shown in Fig. 13.
J = 3. When α increases, the Q W and Q E become smaller, while
As shown in Fig. 12, when δ increases, the Q W becomes Q G , Q P and Q Y become bigger. In order to balance the
smaller, while Q G , Q P and Q Y become bigger. Furthermore, evaluation metrics, we suggest using α = 0.6. Next, β is
Q E increases first and then decreases with increase of δ. analyzed in the same way, which is set as 0.8 in this paper.
The values of the metrics changes slowly when δ >= 10.0. At last, when λ increases, all evaluation metrics have little
Therefore, δ is set as 10.0 in this paper. change. Therefore, we suggest using λ = 1.0.
862 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 27, NO. 2, FEBRUARY 2018
TABLE I TABLE II
SAR-VIS SAR-IR
[48] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image Jingkai Wang received the B.S. degree from the
quality assessment: From error visibility to structural similarity,” IEEE School of Mathematical Sciences, Inner Mongolia
Trans. Image Process., vol. 13, no. 4, pp. 600–612, Apr. 2004. University, Hohhot, China, in 2013. He is currently
[49] G. Piella and H. Heijmans, “A new quality metric for image fusion,” in pursuing the Ph.D. degree with the School of Math-
Proc. 10th Int. Conf. Image Process. (ICIP), 2003, pp. 173–176. ematics and Systems Science, Beihang University,
[50] C. S. Xydeas and V. Petrović, “Objective image fusion performance Beijing, China. His current research interests include
measure,” Electron. Lett., vol. 36, no. 4, pp. 308–309, Feb. 2000. image processing and remote sensing image fusion.
[51] J. Zhao, R. Laganiere, and Z. Liu, “Performance assessment of combina-
tive pixel-level image fusion based on an absolute feature measurement,”
Int. J. Innov. Comput., Inf. Control, vol. 3, no. 6(A), pp. 1433–1447,
Dec. 2007.
[52] C. Yang, J.-Q. Zhang, X.-R. Wang, and X. Liu, “A novel similarity based
quality metric for image fusion,” Inf. Fusion, vol. 9, no. 2, pp. 156–160,
Apr. 2008.
[53] Z. Liu, E. Blasch, Z. Xue, J. Zhao, R. Laganiere, and W. Wu, “Objective
assessment of multiresolution image fusion algorithms for context
enhancement in night vision: A comparative study,” IEEE Trans. Pattern
Anal. Mach. Intell., vol. 34, no. 1, pp. 94–109, Jan. 2012.
[54] P. Sonneveld, “CGS, a fast Lanczos-type solver for nonsymmetric linear
systems,” SIAM J. Sci. Statist. Comput., vol. 10, no. 1, pp. 36–52,
Jan. 1989.
Xiaoyuan Yang received the B.S. degree from the Ridong Zhu received the B.S. degree in information
Department of Mathematics, Liaoning University, and computation science from Tianjin Polytechnic
Shenyang, China, in 1985, the M.S. degree from the University, China, in 2014. He is currently pursuing
Shenyang Computation Institute, Chinese Academy the Ph.D. degree with the School of Mathematics
of Sciences, Shenyang, in 1988, and the Ph.D. and Systems Science, Beihang University, China.
degree from the Department of Computer Science, His current research interests include pattern recog-
Beihang University, Beijing, China, in 2007. She nition and visual tracking.
is currently a Professor with the Department of
Mathematics and Systems Science, Beihang Uni-
versity. Her research interests include computational
harmonic analysis and image processing.