Professional Documents
Culture Documents
Sparse-Adaptive Hypergraph Discriminant Analysis For Hyperspectral Image Classification
Sparse-Adaptive Hypergraph Discriminant Analysis For Hyperspectral Image Classification
net/publication/335745481
CITATIONS READS
129 250
6 authors, including:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Luo Fulin on 08 June 2023.
Authorized licensed use limited to: Wuhan University. Downloaded on June 13,2020 at 03:40:32 UTC from IEEE Xplore. Restrictions apply.
LUO et al.: SAHDA FOR HSI CLASSIFICATION 1083
analysis (SAHDA) method to implement the dimensionality According to the incidence matrix, we can obtain the
reduction of the HSI. SAHDA first uses the sparse repre- degrees of xi and e j , which are represented as
sentation to reveal the intrinsic structure relationships of the
n
n
HSI adaptively. Then, we use the intraclass sparse coefficients div = w j h i j d ej = |e j | = hi j . (3)
to construct an adaptive hyperedge. Finally, we construct j =1 i=1
a dimensionality-reduction model to compute the projection In a low-dimensional space, we preserve the structures of
matrix and the hyperedge weights adaptively, which can be a hypergraph and compact the homogeneous data as close as
solved by the alternating direction method of multipliers possible. That is to say, the vertices on the same hyperedge
(ADMM). SAHDA is more robust to data and can better should be close in low-dimensional space, while the similarity
reveal the intrinsic properties of the HSI. Experiments on a of each hyperedge can be effectively calculated by an adaptive
hyperspectral data set achieve better performance than BH and weight. Thus, the objective function can be denoted
2 as
SGDA. T
1 wi V x j VT xk n
The rest of this letter is organized as follows. Section II J (V, W) = min − +α wi2
die v v
(x j ,xk )∈ei k
details our method. Experimental results are presented in 2 dj d
ei ∈E i=1
Section III to demonstrate the effectiveness of the proposed
method. Finally, Section IV draws some conclusions.
n
s.t. tr(V XX V) = 1,
T T
wi = 1 (4)
II. SAHDA i=1
To represent the intrinsic properties of the HSI adaptively, where W = [wi ]ni . α > 0 is a balanced parameter.
a novel dimensionality-reduction method, termed SAHDA, For the optimal problem of (4), we construct an augmented
was proposed in this letter, as shown in Fig. 1. First, spare Lagrangian function with a Lagrangian multiplier δ and λ as
representation is adopted to represent the intrinsic relation-
ships of the HSI adaptively. Then, according to the intra- L(V, W, δ, λ)
2
T
class sparse coefficients, we construct an adaptive hypergraph 1 wi V x j
− V T x
k
model. Finally, an adaptive dimensionality reduction model =
is designed to learn the weights of the hyperedges and the 2 die
ei ∈E (x j ,xk )∈ei d vj dkv
projection matrix. n
n
Suppose a hyperspectral data set X = [x1 , x2 , . . . , +α wi +δ
2
wi −1 +λ(1−tr(VT XXT V)). (5)
xn ] ∈ D×n contains n pixels with D spectral bands. i=1 i=1
(xi ) ∈ {1, 2, . . . , c} denotes the class label of xi , where c
is the class number of the land cover. The low-dimensional Then, we use the ADMM to update the solution of (5).
embedding of X denotes Y = [y1 , y2 , . . . , yn ] ∈ d×n , where We first fix W to update V, and the objective function can be
d is the embedding dimension. We can get Y = VT X with a represented as
projection matrix V ∈ D×d , where d << D. 2
Sparse representation aims to represent a data point with T
1 wi V xj V xk
T
L(V, λ) = v − d v
a dictionary and obtain the representation coefficients as
de
e ∈E i (x ,x )∈e k
sparse as possible. For the representation coefficients, most 2 d
i j k i j
of them are zero, and only a few elements are nonzeros
that corresponding data points possess strong relevance. These + λ(1 − tr(V XX V)) T T
2
nonzero coefficients corresponding to data points can reveal
1 h i j h ik wi VT x j VT xk
the intrinsic properties of the data. For a data point xi , its = − v
sparse coefficients can be obtained by the following problem: 2 die d v d k
ei ∈E (x j ,xk )∈X j
min si 1 s.t. xi − Xsi < ε (1) + λ(1 − tr(V XX V)) T T
si
= tr(VT XL H XT V) + λ(1 − tr(VT XXT V)) (6)
where ε > 0 is the error tolerance. || • ||1 is the l1 -norm
−1/2 T −1/2
that controls the sparsity of coefficients. si = [si,1 , si,2 , where L H = I − Dv HWD−1
e H Dv is the
. . . , si,i−1 , 0, si,i+1 , . . . , si,n ] are the sparse coefficients. hyper-Laplacian matrix. I is an identity matrix.
Authorized licensed use limited to: Wuhan University. Downloaded on June 13,2020 at 03:40:32 UTC from IEEE Xplore. Restrictions apply.
1084 IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 17, NO. 6, JUNE 2020
Authorized licensed use limited to: Wuhan University. Downloaded on June 13,2020 at 03:40:32 UTC from IEEE Xplore. Restrictions apply.
LUO et al.: SAHDA FOR HSI CLASSIFICATION 1085
Fig. 3. Classification map with different methods, where the first and second rows show the results of SVM and NN, and the first to fourth columns show
the results of RAW, BH, SGDA, and SAHDA.
TABLE I
C LASSIFICATION R ESULTS OF E ACH C LASS (%)
Fig. 4. Classification results with SVM (the first row) and NN (the second
row) under different embedding dimensions.
methods can reduce the redundancy and preserve the valuable E. Results With Different Numbers of Training Samples
information to enhance the discriminant performance of the In this section, we analyzed the results with different
HSI. For all the compared methods, SAHDA possesses the numbers of training samples. We selected randomly 5, 10,
best AA, OA, and KC. The visualized results are shown 15, 20, 25, and 30 samples from each class for training and
in Fig. 3. The proposed method achieves a smoother region repeated ten times under each condition. Table II shows the
than RAW, BH, and SGDA, especially in the areas of Road and average OAs with STD and KCs.
water because SAHDA can adaptively represent the intrinsic In Table II, the results indicate that the accuracies improve
multiple relationships and the similarity of the homogeneous with the increasing of training samples because more informa-
samples. tion can be used to construct the training model. Furthermore,
the proposed method achieves better OAs and KCs than the
D. Dimensionality Analysis other methods under each case.
To analyze the effect of embedding dimension, 30 samples
were randomly selected from each class for training and the F. Parameter Analysis
other samples were used for testing. We repeated ten times For SAHDA, it has a parameter α to adjust the weights.
under each case and Fig. 4 shows the results of each method To evaluate the influence of α, we randomly selected 30 sam-
under different embedding dimensions. ples from each class as the training set and the remaining
According to Fig. 4, the classification accuracies improve samples were considered as the test set. We set α to 0.01, 0.1,
and then reach a stable value with the increasing of embedding 1, 5, 10, 15, 20, 25, and 30, and Fig. 5 shows the average OAs
dimension, because the increasing dimension can obtain more with a ten-time-repeated experiment under each condition.
Authorized licensed use limited to: Wuhan University. Downloaded on June 13,2020 at 03:40:32 UTC from IEEE Xplore. Restrictions apply.
1086 IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 17, NO. 6, JUNE 2020
TABLE II
C LASSIFICATION R ESULTS W ITH D IFFERENT N UMBERS OF T RAINING S AMPLES (OA±STD (KC)%)
View publication stats Authorized licensed use limited to: Wuhan University. Downloaded on June 13,2020 at 03:40:32 UTC from IEEE Xplore. Restrictions apply.