Professional Documents
Culture Documents
Real Time Recursive Hyperspectral Sample and Band Processing PDF
Real Time Recursive Hyperspectral Sample and Band Processing PDF
Chang
Real-Time Recursive
Hyperspectral
Sample and Band
Processing
Algorithm Architecture and
Implementation
Real-Time Recursive Hyperspectral Sample
and Band Processing
Chein-I Chang
Real-Time Recursive
Hyperspectral Sample
and Band Processing
Algorithm Architecture and Implementation
Chein-I Chang
Center for Hyperspectral Imaging in Remote Sensing (CHIRS)
Information and Technology College
Dalian Maritime University
Dalian, China
Remote Sensing Signal and Image Processing Laboratory (RSSIPL)
University of Maryland, Baltimore County (UMBC)
Baltimore, MD, USA
vii
viii Preface
Two types of PHSI are designed and developed in this book to process
hyperspectral data in real time. One is sample-wise PHSI which processes
hyperspectral data sample by sample in a progressive manner, with full bands of
each data sample vector being processed. This is generally used to address the issue
of BIS process. An approach to implementing sample-wise PHSI developed in this
book is called progressive hyperspectral sample processing (PHSP), which is
similar to the BIS concept. In other words, all data sample vectors are fully
processed sample by sample progressively in real time. Most hyperspectral imaging
algorithms currently being used can be interpreted by PHSP one way or another.
The other type is band-wise PHSI which operates hyperspectral data by
processing bands in a progressive manner, where data samples in each of individual
bands are processed separately and sequentially one band after another. This type is
generally used to address the issue associated with the BSQ process. An approach to
band-wise PHSI is called progressive hyperspectral band process (PHBP), which is
derived from the BSQ idea with slight modifications. It can be further extended to
recursive hyperspectral band process (RHBP) in which data samples can be
processed in real time as new bands come in. In this book, both PHBP and RHBP
will be referred to as RHBP where RHBP can be considered as a recursive version
of PHBP. More specifically, data processing by RHBP can take place as new
incoming bands feed in, while at the same time, the processed results are output
at nearly the same time provided the computing time is negligible. Unfortunately,
such RHBP seems to have received little interest to date. For example, if we assume
that the use of each spectral band is specified by a bit in terms of whether or not it is
selected for data processing, the number of bits used to process data samples can be
interpreted as the number of bands used to process data. The proposed RHBP
materializes this concept to make it applicable to hyperspectral data processing.
Unlike PHSP, which seems natural and easy to understand, RHBP is rather new.
On some occasions RHBP has been confused with band selection (BS) methods.
This deserves more clarification. RHBP is a new theory developed for hyperspectral
data transmission and communications where hundreds of spectral channels can be
fine-tuned progressively for data processing. It is not a BS process. More specifi-
cally, it deals with issues of transmitting and communicating hyperspectral data
band by band but not issues of selecting bands. Thus, RHBP does not need to solve
band selection optimization problems to find optimal bands. In fact, BS and
progressive band process are completely separate subjects. They may be correlated
but concepts are rather different. BS must select all necessary bands prior to data
processing. In this case, it requires that prior knowledge of how many bands need to
be determined prior to the selection of optimal bands. In this case, a recently
developed Virtual Dimensionality (VD) has been used to estimate the band number
and a band selection criterion is then used to find all required optimal bands. As a
result, BS generally cannot be implemented in real time. In contrast, RHBP
performs data processing as new incoming bands are fed in without having to
determine the band number and selecting bands. Such an advantage is significant
in data transmission and communication, a task which cannot be accomplished by
BS. Moreover, RHBP offers many advantages that cannot be provided by most
Preface ix
University, China. Also, I would like to thank the members of the Center for
Hyperspectral Imaging in Remote Sensing (CHIRS) at Dalian Maritime University,
Professors Meiping Song (宋梅萍), Sen Li (李森), Chunyan Yu (于纯妍), Lin
Wang ( ) and Yulei Wang ( ).
In addition, my appreciation is also expanded to my colleagues, Professor Pau-
Choo Chung ( ) with the Department of Electrical Engineering, National
Cheng Kung University, Professor Yen-Chieh Ouyang ( ) with the
Department of Electrical Engineering, National Chung Hsing University, Professor
Chinsu Lin ( ) with the Department of Forestry and Natural Resources at
National Chiayi University, Professor Chia-Hsien Wen ( ) with Providence
University, Dr. Ching-Wen Yang ( ) who is the Director of Computer
Center, Taichung Veterans General Hospital and Ching-Tsong Tsai ( )
with Tunghai University. Specifically, I would like to particularly thank my former
Ph.D. student Dr. Hsiao-Chi Li for carrying out nearly all experiments presented in
Chaps. 2, 6, 10–12, 16 and current Ph.D. student Ms. Yao Li for doing all
experiments in Chaps. 14, 15, 18–20. This book cannot be completed without
their involvements.
I would also like to acknowledge several universities in Taiwan and China for
their financial and professional support during my visit to China and Taiwan,
particularly, Chang Jiang Scholar Chair Professorship (教育部長江學者講座教
授) from Dalian Maritime University (大連海事大學) in China along with strong
and enthusiastic supports from its president Dr. Yuqing Sun ( ), Dis-
tinguished Chair Professorship of Remote Sensing Technology (遙測科技傑出講
座教授) from the National Chung Hsing University, and Adjunct Chair Professor-
ship from Providence University and Chair Professorship with the Taichung Vet-
erans General Hospital.
Finally, I would also like to thank many of my special friends in Taichung
Veterans General Hospital (TCVGH), Dr. San-Kan Lee ( ) (former
Superintendent of TCVGH), Dr. Ping-Wing Lui ( ) (Deputy Super-
intendent of TCVGH), Dr. Yong Kie Wong ( ) (Deputy Superinten-
dent at Show Chwan Health Care System), Dr. Clayton Chi-Chang Chen
( ) (Chairman of Radiology at TCVGH), Dr. Yen-Chuan Ou
( ) (Head of Department of Medical Research at TCVGH), Dr.
Bor-Jen Lee ( )( at TCVGH), Dr. Jyh-Wen Chai
( ) (Chief of Radiology at TCVGH), Dr. Man-Yee Chan ( )
(Chief of Oral and Maxillofacial Surgery at TCVGH), Dr. Francis S.K. Poon
( ) (Director, Clinical Informatics Research and Development Center
at TCVGH) and Dr. Siwa Chan ( ) at TCVGH who have unselfishly provided
their expertise and resources for the research described in this book during my stay
in Taichung, Taiwan. Last but not least, I would also like to thank my very close and
special friend, General Manager, Vincent Tseng ( ) at Bingotimes,
Inc. ( ). All their supports are greatly appreciated.
Preface xi
Finally and most importantly, I would like to thank my family members, older
sister Feng-Chu Chang ( ), younger sister Mann-Li Chang ( ), youn-
ger brother Chein-Chi Chang ( ), sister-in-law Chuen-Chin Kuo ( ) as
well as my niece Chia-Fang Lue ( ), also my three nephews, Yu-Wei Wayne
Chang ( ), Yu-Cheng Channing Chang ( ) and Yu-Rui Raymond
Chang ( ). Their supports cannot be underestimated as this book was being
prepared during the most difficult time in my life after my mother passed away.
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.2 Recursive Hyperspectral Sample Processing . . . . . . . . . . . . . 5
1.2.1 Sample Spectral Statistics-Based Recursive
Hyperspectral Sample Processing . . . . . . . . . . . . . . . 7
1.2.2 Signature Spectral Statistics-Based Recursive
Hyperspectral Sample Processing . . . . . . . . . . . . . . . 8
1.3 Recursive Hyperspectral Band Processing . . . . . . . . . . . . . . . 8
1.3.1 Band Selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.3.2 Progressive Hyperspectral Band Processing . . . . . . . 9
1.3.3 Recursive Hyperspectral Band Processing . . . . . . . . 10
1.4 Scope of Book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
1.4.1 Part I: Fundamentals . . . . . . . . . . . . . . . . . . . . . . . . 12
1.4.2 Part II: Sample Spectral Statistics-Based
Recursive Hyperspectral Sample Processing . . . . . . . 12
1.4.3 Part III: Signature Spectral Statistics-Based
Recursive Hyperspectral Sample Processing . . . . . . . 13
1.4.4 Part IV: Sample Statistics-Based Recursive
Hyperspectral Band Processing . . . . . . . . . . . . . . . . 13
1.4.5 Part V: Signature Statistics-Based Recursive
Hyperspectral Band Processing . . . . . . . . . . . . . . . . 14
1.5 Real Hyperspectral Images to Be Used in This Book . . . . . . . 14
1.5.1 AVIRIS Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.5.2 HYDICE Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
1.5.3 Hyperion Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
1.6 Synthetic Images to Be Used in this Book . . . . . . . . . . . . . . . 24
1.7 How to Use this Book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
1.8 Notations and Terminology Used in the Book . . . . . . . . . . . . . 28
xiii
xiv Contents
Part I Fundamentals
2 Simplex Volume Calculation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
2.2 Determinant-Based Simplex Volume Calculation . . . . . . . . . . 33
2.3 Geometric Simplex Volume Calculation . . . . . . . . . . . . . . . . . 34
2.4 General Theorem for Geometric Simplex
Volume Calculation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
2.5 A Mathematical Toy Example . . . . . . . . . . . . . . . . . . . . . . . . 43
2.6 Real Image Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
2.7 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
3 Discrete-Time Kalman Filtering for Hyperspectral Processing . . . . 49
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
3.2 Discrete-Time Kalman Filtering . . . . . . . . . . . . . . . . . . . . . . 50
3.2.1 A Priori and A Posteriori State Estimates . . . . . . . . . 51
3.2.2 Finding an Optimal Kalman Gain K(k) . . . . . . . . . . . 53
3.2.3 Orthogonality Principle . . . . . . . . . . . . . . . . . . . . . . 54
3.2.4 Discrete-Time Kalman Predictor and Filter . . . . . . . . 56
3.3 Kalman Filter-Based Linear Spectral Mixture Analysis . . . . . . 58
3.4 Kalman Filter-Based Hyperspectral Signal Processing . . . . . . . 61
3.4.1 Kalman Filter-Based Hyperspectral Signal
Processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
3.4.2 Kalman Filter-Based Spectral Signature Estimator . . . 63
3.4.3 Kalman Filter-Based Spectral Signature Identifier . . . 64
3.4.4 Kalman Filter-Based Spectral Signature
Quantifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
3.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
4 Target-Specified Virtual Dimensionality
for Hyperspectral Imagery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
4.2 Review of VD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
4.3 Eigen-Analysis-Based VD . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
4.3.1 Binary Composite Hypothesis Testing Formulation . . 81
4.3.2 Discussions of HFC Method and MOCA . . . . . . . . . . 85
4.4 Finding Targets of Interest . . . . . . . . . . . . . . . . . . . . . . . . . . 86
4.4.1 What Are Targets of Interest? . . . . . . . . . . . . . . . . . . 87
4.4.2 Second-Order-Statistics (2OS)-Specified
Target VD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
4.4.3 HOS-Specified Target VD . . . . . . . . . . . . . . . . . . . . 97
4.5 Target-Specified VD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
4.5.1 Target-Specified Binary Hypothesis Testing . . . . . . . 100
4.5.2 ATGP-Specified VD Using ηl ¼ tlATGP 2 . . . . . . . . 103
pffiffiffiffi
4.5.3 ATGP-Specified VD Using ηl ¼ tlATGP . . . . . . . 103
4.5.4 Discussions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
Contents xv
Glossary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 655
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 659
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 681
About the Author
xxiii
Chapter 1
Introduction
Acronyms
α Abundance vector
^
α Estimate of abundance vector α
αj jth abundance fraction
^j
α Estimate of jth abundance fraction (αj)
A Weighting or mixing matrix
Az Area under an ROC curve
Bl lth band image
bl lth band image represented as a vector
C Total number of classes
Cj jth class
d Desired signature vector
D Desired signature matrix
Dλ Eigenvalue diagonal matrix
δ Detector or classifier
Δ Database
ε Error threshold
ej jth endmember
I Identity matrix
I(x) Self-information of x
k(.,.) Kernel function
K Total number of skewers used in Pixel Purity Index
K Sample covariance matrix
λ Eigenvalue of sample covariance matrix (K)
^λ Eigenvalue of sample correlation matrix (R)
l Index of band number
L Number of spectral channels or bands
Λ Eigenvector matrix
μ Global mean
μj Mean of jth class
mj jth signature vector
M Signature or endmember matrix
n Noise vector
N Total number of image pixel vectors in a band image, i.e., N ¼ nr nc
nB Number of bands being processed
nBS Number of bands to be selected
nl Set of first l bands
ns Number of first s sample vectors being visited
nVD Number of endmembers or signatures estimated by virtual dimensionality
p Number of endmembers
PD Detection power or probability
PF False alarm probability
P⊥U Projector to reject undesired target signatures in U
1.1 Introduction 3
1.1 Introduction
The development of progressive hyperspectral imaging (PHSI) arose from how data
are acquired and collected as well as how data are processed. In general,
hyperspectral imaging sensors use two data acquisition formats (Schowengerdt
1997). One is referred to as the band-interleaved-by-sample (BIS) or the band-
interleaved-by-pixel (BIP) process (Fig. 1.1a), which acquires data sample vectors
with full band information sequentially sample by sample, or the band-interleaved-
by-line (BIL) process shown in Fig. 1.1b, which acquires data sample vectors with
full band information sequentially line by line, where (x,y) indicates the spatial
coordinate of a data sample vector or pixel vector and λ is a wavelength parameter
used to specify spectral bands. Note that the pixels highlighted by red in Fig. 1.1a
and the line highlighted by red in Fig. 1.1b are the data sample vectors currently
being processed.
Another format is referred to as Band SeQuential (BSQ) process shown in
Fig. 1.2, which acquires band images band by band sequentially where (x,y)
indicates the spatial coordinate of a data sample vector or pixel vector and λ is a
wavelength parameter used to specify spectral bands, and the band image
highlighted by RED in Fig. 1.2 is the band data currently being processed.
According to Figs. 1.1 and 1.2, to process a designed hyperspectral imaging
algorithm in real time, it must follow one of these two data acquisition formats.
With this interpretation the BIP/BIS/BIL format generally results in real-time
processing since the data sample vectors are processed in a raster fashion progres-
sively sample by sample or line by line. The PHSI theory presented in Chang (2016)
is developed to address this issue. To further facilitate the real-time capability of
PHSI, the first part of this book (i.e., Part II) extends PHSI to recursive
hyperspectral sample processing (RHSP) so that RHSP can be carried out by
4 1 Introduction
Fig. 1.1 Hyperspectral imagery acquired by BIP/BIS/BIL format. (a) BIP/BIS; (b) BIL
updating data sample vectors recursively via recursive equations. On the other
hand, the BSQ format allows users to fully process each band image band by
band progressively. However, owing to the very high correlation among bands
resulting from hundreds of contiguous spectral channels used by hyperspectral
imaging sensors, it is anticipated that much interband redundancy can be removed
without loss of much information. Thus, effectively carrying out this task becomes
a challenging issue. The third part of this book (i.e., Part III) takes up this issue. In
particular, it extends PHSI to recursive hyperspectral band processing (RHBP),
which can be implemented more efficiently and effectively in real time by updating
data sample vectors to allow progressive hyperspectral band processing (PHBP) to
be processed not only progressively band by band but also recursively by recursive
equations.
1.2 Recursive Hyperspectral Sample Processing 5
such as anomaly detection, which finds targets whose signatures are spectrally
distinct from those of the surrounding sample vectors. The other is to use
signature characterization, such as a linear mixing model (LMM) used in linear
spectral mixture analysis (LSMA) for data unmixing. Thus, depending on which
one is used, the recursive equations derived for these two categories are also quite
different.
To date, most hyperspectral imaging algorithms have been designed and developed
on the basis of a single-pixel vector. This means that the algorithms process data
sample vectors pixel by pixel without taking into account interpixel spatial corre-
lation. Examples include target detection, such as OSP (Harsanyi and Chang 1994)
and automatic target generation process (ATGP) (Ren and Chang 2003);
endmember extraction/finding, such as the Pixel Purity Index (PPI) (Boardman
1994), N-FINDR (Winter 1999), and a simplex growing algorithm (SGA) (Chang
et al. 2006a, b); and LSMA, such as fully constrained least squares (FCLS) method
(Heinz and Chang 2001). Thus, technically speaking, all these algorithms can be
processed in real time once their required prior knowledge is obtained. Unfortu-
nately, in real-world applications, obtaining such prior knowledge is either very
expensive or nearly impossible. Under such circumstances, algorithms must be
designed in an unsupervised manner, in which case the required knowledge can be
obtained only directly from the data to be processed a posteriori. Like target
detection (Sect. 1.2.1), PMP/PSP will require two passes or stages to perform
data processing in real time where the first pass/stage process is to acquire neces-
sary a posteriori knowledge required by the algorithm to be implemented and the
second pass/stage is to run the algorithm in real time. However, there is a disad-
vantage of using this method. If the obtained a posteriori knowledge needs to be
updated sample by sample, PMP/PSP must be repeatedly implemented over and
over again. This will result in significant increases in computing time, not to
mention that such time delays will make it impossible to implement the algorithm
in real time. Part III of this book is dedicated to resolving this issue.
RHBP can be considered a parallel theory to RHSP (Sect. 1.2), both of which
correspond to two completely different data acquisition formats, BIP/BIS/BIL and
BSQ, respectively. While these two processes appear similar, the techniques devel-
oped for both are quite different because recursive equations derived for BIP/BIS/
BIL are based on a covariance/correlation matrix or signature matrix formed by
data sample vectors, whereas the recursive equations derived for BSQ are from a
band matrix, which is formed by band images, not data sample vectors.
In recent years, hyperspectral band processing has generally been performed by
either data compression via linear/nonlinear transforms or data reduction via band
selection (BS), each of which has its own merits and applications. However, from a
real-time processing point of view, BS seems to be a better option since data
compression generally requires the entire data set before transforms are carried
out and cannot be implemented in a causal manner; thus, it cannot be implemented
1.3 Recursive Hyperspectral Band Processing 9
in real time. This will be particularly important for data communications and
transmission in future satellite hyperspectral data processing. Thus, Parts IV
and V of this book will focus on how BS can be extended and expanded from
various perspectives.
and structures not specified by particular applications that actually determine how
bands are selected. To mitigate this problem, an alternative approach to BS was
developed by Chang et al. (2011a, b, c, d) and Chang (2013) and is called
progressive band dimensionality processing (PBDP), in which bands are processed
without determining the value of nBS. This was further extended to progressive
band selection by Chang and Liu (2014) and Chang (2013). To properly select
bands, both BS and PSDP approaches require band prioritization (BP) criteria to
first rank bands so that bands can be selected according to their priorities. However,
different BP criteria generally result in different sets of selected bands. In partic-
ular, when BP is designed based on data statistics and structures, the bands selected
by BP may have nothing to do with applications. Another issue is that if a band is
selected by its high priority, it is very likely that its adjacent bands may also be
selected owing to their high correlation with the selected band. This furthermore
indicates that these bands may be redundant and should not be selected. To resolve
such a dilemma, band decorrelation (BD) is required to remove bands having high
correlation with selected bands. However, this also leads to another challenging
issue: determining an appropriate threshold for BD so that BD between bands and
bands below this threshold will be not selected. Apparently, such a threshold cannot
be fixed and must adapt to the data set and applications. To avoid the use of BP and
BD, we need to reinvent the wheel and seek a completely different approach. This
book develops a new approach, called PHBP, that does not rely on BP and
BD. Instead, PHBP is performed using algorithms that can be specified by a
particular application. For example, in applications of hyperspectral target detec-
tion, Chang et al. (2015a, b) and Chang and Li (2016a, b) developed PHBP versions
of AD and CEM, respectively.
One issue associated with implementing PHBP developed in Sect. 1.3.2 is that it
must repeatedly recalculate causal band matrices as new band images come in. As
the number of bands grows large, such calculation becomes excessive and may be
impossible to manage. To address this computational issue, PHBP is further
extended to RHBP, which not only can implement PHBP progressively band by
band but can also update causal band matrices recursively band by band via
recursive equations. As a consequence, there are several salient differences between
PHBP and RHBP. First, RHBP can be considered a Kalman-filtering-like process
carried out by recursive equations. Second, owing to its recursive nature, RHBP can
also be implemented in real time. Third, RHBP can be easily implemented in
hardware, which provides further feasibility in chip design. Finally and most
importantly, RHBP has great promise in future hyperspectral data communication
and transmission.
1.4 Scope of Book 11
This book is comprised of five parts. Part I: Fundamentals provides needed knowl-
edge to help readers follow in subsequent chapters. Parts II–V are devoted to RHSP
and RHSP. In particular, Parts II and III design and develop techniques for sample
spectral statistics-based RHSP and signature spectral statistics-based RHSP,
respectively, whereas Parts IV and V follow treatments similar to those in Parts II
and III to design and develop techniques for sample spectral statistics-based RHBP
and signature spectral statistics-based RHBP, respectively.
12 1 Introduction
Part II discusses recursive real-time RHSP for target detection, which makes the
covariance/correlation matrix K/R recursive to update targets of interest as well as
progressive to process data sample vectors. The other is real-time RHSP for signa-
ture generation, to be treated in Part III, which finds signatures recursively on a
single-pixel processing basis while progressively processing data sample vectors in
real time. As described in Chang (2016), causality is a prerequisite for real-time
processing. In this case, the concept of causality must be included in design and
development of target detection algorithms where causal sample covariance matrix
(CSCVM) and causal sample correlation matrix (CSCRM) are introduced as causal
versions of the commonly used global sample covariance matrix and global sample
correlation matrix, respectively. By virtue of these newly defined causal covariance/
correlation matrices, hyperspectral target detection can be carried out in real time
based on a BIP/BIS/BIL data acquisition format. Most importantly, the profiles of
various degrees of sample-wise progressive background suppression by CSCVM
and CSCRM can also be generated for visual assessment, with target detection
taking place recursively which cannot be accomplished by conventional target
detection.
Two chapters (Chaps. 5 and 6) on recursive real-time hyperspectral target
detection are included in Part II:
• Chapter 5: Recursive Real Time Hyperspectral Sample Processing for Active
Target Detection: Constrained Energy Minimization;
• Chapter 6: Recursive Real Time Hyperspectral Sample Processing for Passive
Target Detection: Anomaly Detection.
1.4 Scope of Book 13
Following the same logic used to derive Part II, a parallel theory is also developed
in Part IV for real-time RHBP based on the BSQ data acquisition format. Thus, Part
IV presents recursive real-time causal band covariance/correlation-based RHBP for
target detection, which makes use of the causal band covariance/correlation matrix
to update targets of interest band by band, while Part V addresses recursive real-
time RHBP for finding signatures band by band.
Since the process is carried out band by band in real time, causal versions of
band covariance/correlation matrices similar to CSCVM and CSCRM are also
introduced as causal band sample covariance matrix (CBSCVM) and causal band
sample correlation matrix (CBSCRM) for target detection. Two chapters
(Chaps. 13 and 14) are included as counterparts to Chaps. 5 and 6 in Part II:
14 1 Introduction
Although Part V can also be considered as the counterpart to Part III, it should be
noted that in Part III signatures are updated recursively using full band information
through RHSP. By contrast, in Part V signatures are updated band by band while the
number of signatures is fixed by RHBP. As a result, there is no need to determine
how many signatures must be generated. Instead, RHBP allows users to determine
how many bands must be used when the performance of algorithms shows little
improvement as the number of bands increases.
Six chapters (Chaps. 15–20) are included in Part V:
• Chapter 15: RHBP of Automatic Target Generation Process
• Chapter 16: RHBP of Orthogonal Subspace Projection
• Chapter 17: RHBP of Linear Spectral Mixture Analysis
• Chapter 18: RHBP of Growing Simplex Volume Analysis
• Chapter 19: RHBP of Iterative Pixel Purity Index
• Chapter 20: RHBP of Fast Iterative Pixel Purity Index
Three real hyperspectral image data sets are frequently used in this book in
experiments. Two are AVIRIS real image data sets, Cuprite Mining District in
Nevada and Purdue’s Indian Pine test site in Indiana. A third image data set is the
HYperspectral Digital Imagery Collection Experiment (HYDICE) image scene
(Basedow et al. 1992). Each of these three data sets is briefly described in what
follows.
Two AVIRIS data sets presented in this section are Cuprite data and Lunar Crater
Volcanic Field (LCVF) data, which can be used for different purposes in
applications.
1.5 Real Hyperspectral Images to Be Used in This Book 15
One of the most widely studied hyperspectral image scenes available in the public
domain is the Cuprite Mining District in Nevada (Fig. 1.3a) with a region of interest
shown as a subscene in Fig. 1.3b, which is a 224-band image 350 350 pixels in
size collected over the Cuprite Mining District site in Nevada in 1997.
It is well understood mineralogically. As a result, a total of 189 bands were used
for experiments where bands 1–3, 105–115, and 150–170 were removed prior to the
analysis owing to water absorption and low signal-to-noise ratio (SNR) in those
bands.
Although there are more than five minerals in the data set in Fig. 1.4a, which is
an enlarged figure of Fig. 1.3b, the ground truth available for this region only
provides the locations of the pure pixels: alunite (A), buddingtonite (B), calcite (C),
kaolinite (K), and muscovite (M). The locations of these five pure minerals are
labeled A, B, C, K, and M, respectively, and shown in Fig. 1.4b.
The five pure pixels shown in Fig. 1.4b are carefully verified using laboratory
spectra in Fig. 1.5 provided by the USGS (available at http://speclab.cr.usgs.gov).
Fig. 1.3 Cuprite Mining District image scene (a) original Cuprite Mining District image scene;
(b) the image cropped from the center region of the original scene in (a) (350 350)
16 1 Introduction
Fig. 1.4 (a) Spectral band number 170 of Cuprite Mining District AVIRIS image scene; (b)
spatial positions of five pure pixels corresponding to minerals alunite (A), buddingtonite (B),
calcite (C), kaolinite (K), and muscovite (M)
1 1
Calcite
Calcite
0.9
Alunite
Kaolinite Alunite
0.8 0,8
0.7 Kaolinite
Reflectance
Reflectance
Muscovite
0.6 0,6
Buddingtonite
Buddingtonite
0.5
0.4 0,4
Muscovite
0.3
0,2
0.2
20 40 60 80 100 120 140 160 180 200 220 400 700 1000 1300 1600 1900 2200 2500
Band number Wavelength (nm)
It is recommended that bands 1–3, 105–115, and 150–170, owing to the low
water absorption and low SNR in those bands, be removed prior to data processing.
As a result, a total of 189 bands are used for experiments, and the reflectance and
radiance signatures of the five pure pixels shown in Fig. 1.4b are shown in Fig. 1.6a,
b. The steps to produce the spectra in Fig. 1.6a, b follow:
1. Download the laboratory reflectance data at http://speclab.cr.usgs.gov/.
2. Use Spectral Angle Mapper (SAM) (Chang 2003a, b) as a spectral similarity
measure to identify the five pixels in Fig. 1.4a that correspond to the five
reflectances obtained in step 1 using the following procedure.
1.5 Real Hyperspectral Images to Be Used in This Book 17
a
7000
Muscovite Alunite
Buddingtonite
6000 Calcite
Alunite Kaolinite Kaolinite
Muscovite
5000
Reflectance
4000 Buddingtonite
3000
Calcite
2000
1000
0
400 600 800 1000 1200 1400 1600 1800 2000 2200 2400
Wavelength(nm)
b
12000
Alunite
Kaolinite
Muscovite Buddingtonite
Calcite
10000 Muscovite
Kaolinite
8000 Alunite
Radiance
6000
4000
Calcite
Buddingtonite
2000
0
400 600 800 1000 1200 1400 1600 1800 2000 2200 2400
Wavelength(nm)
Fig. 1.6 (a) Reflectance signatures of five minerals marked in Fig. 1.4b in wavelengths;
(b) radiance signatures of five minerals marked in Fig. 1.4b in bands
18 1 Introduction
U
D
U
D
D
U
Silica
Alunite
Kaolinite
Kaolinite +
Muscovite U
D
Propylitic
Muscovite 1 km
U D
A second AVIRIS image scene to be used in our experiments is shown in Fig. 1.8a.
It is the Lunar Crater Volcanic Field (LCVF) located in Northern Nye County,
Nevada, which is one of the earliest hyperspectral image scenes studied in the
literature. Atmospheric water bands and low-SNR bands were removed from the
1.5 Real Hyperspectral Images to Be Used in This Book 19
a
vegetation
cinders
shade
anomaly
rhyolite
dry lake
b
1800
anomaly
1600 cinders
dry lake
1400 rhyolite
shade
1200 vegetation
Radiance
1000
800
600
400
200
0
0.40 µm 2.5 µm
Wavelength
Fig. 1.8 (a) AVARIS LCVF scene; (b) spectra of anomaly, cinders, dry lake, rhyolite, shade,
and vegetation
data, reducing the image cube from 224 to 158 bands. The LCVF image has a 10 nm
spectral resolution and 20 m spatial resolution.
The image scene in Fig. 1.8a is relatively simple compared to the Cuprite Mining
District scene in Fig. 1.4a, where there are five targets of interest: red oxidized
basaltic cinders, rhyolite, playa (dry lake), vegetation, and shade, whose radiance
spectra are plotted in Fig. 1.8b. In addition to these five target signatures, there is an
interesting target that is an anomaly two pixels in size located in the upper left dry
lake with its spectral signature also plotted in Fig. 1.8b.
20 1 Introduction
The HYDICE image scene shown in Fig. 1.9a is 200 74 pixel vectors in size, and
its ground truth is provided in Fig. 1.9b, where the center and boundary pixels of the
objects are highlighted by red and yellow, respectively. The upper part contains
fabric panels 3, 2, and 1 m2 in size from the first column to the third column. Since
the spatial resolution of the data is 1.56 m2, the panels in the third column are
treated as subpixel target. The lower part contains different vehicles with sizes of
4 8 m (the first four vehicles in the first column) and 6 3 m (the bottom vehicle
in the first column) and three objects in the second column (the first two have a size
of two pixels and the bottom one has three pixels), respectively. In this particular
scene there are three types of targets with different sizes, small targets (panels of
three different sizes, 3, 2, and 1 m2) and large targets (vehicles of two different
sizes, 4 8 m and 6 3 m and three objects two and three pixels in size) to be used
to validate and test target detection performance.
Figure 1.9c shows an enlarged HYDICE scene from the same flight for visual
assessment. It has a size of 33 90 pixel vectors with 10 nm spectral resolution and
1.56 m spatial resolution, where five vehicles lined up vertically to park along the
tree line in a field, where the red (R) pixel vectors in Fig. 1.9d show the center pixel
of the vehicles, while the yellow (Y) pixels are vehicle pixels mixed with back-
ground pixels.
Fig. 1.9 HYDICE vehicle scene; (a) image scene; (b) ground truth map; (c) five vehicles; (d)
ground truth of (c)
1.5 Real Hyperspectral Images to Be Used in This Book 21
a b
p11, p12, p13
Fig. 1.10 (a) A HYDICE panel scene containing 15 panels; (b) ground truth map of spatial
locations of the 15 panels
A third enlarged HYDICE image scene (Fig. 1.10a) is also cropped from the
upper part of the image scene in Fig. 1.9a, b marked by a square.
It has a size of 64 64 pixel vectors, with 15 panels in the scene. This particular
image scene was studied in depth in Chang (2003a, b, 2013, 2016). The scene
contains a large grass field background, a forest at the left edge, and a barely visible
road running along the right edge. Bands 1–3 and 202–210 are low signal/high
noise bands, and bands 101–112 and 137–153, which are water vapor absorption
bands, were removed. The spatial resolution is 1.56 m, and the spectral resolution is
10 nm. There are 15 panels located in the center of the grass field. They are arranged
in a 5 3 matrix, as shown in Fig. 1.10b, which provides the ground truth map of
Fig. 1.10a with the panel signatures in watts per square meter. Each element in this
matrix is a square panel and denoted by pij with rows indexed by i ¼ 1, . . . , 5 and
columns indexed by j ¼ 1, 2, 3.
For each row i ¼ 1, . . . , 5, the three panels were painted using the same material,
but they have three different sizes. According to the ground truth, the panels in rows
2 and 3 are made of the same materials with slightly different colors of paint, a light
olive parachute and a dark olive parachute. Consequently, if panels could be
detected in row 2, then panels in row 3 would also be detectable, and vice versa.
The same principle is also applicable to panels in rows 4 and 5, which are also made
of the same green fabrics with/without spectral tags. Nevertheless, they were still
considered different materials. For each column j ¼ 1, 2, 3, the five panels have the
same size. The sizes of the panels in the first, second, and third columns are 3 3 m,
2 2 m, and 1 1 m respectively.
Thus, the 15 panels have 5 different materials and 3 different sizes. Figure 1.10b
shows the precise spatial locations of these 15 panels, where red (R) pixels are a
panel’s center pixels and yellow pixels (Y) are panel pixels mixed with the
background. The 1.56 m spatial resolution of the image scene suggests that the
panels in the second and third columns, denoted by p12, p13, p22, p23, p32, p33, p42,
p43, p52, p53 in Fig. 1.10b, are one pixel in size. Additionally, except for the panel in
the first row and first column, denoted by p11, which also has a size of one pixel, all
other panels located in the first column are two-pixel panels; these are the panel in
22 1 Introduction
the second row with two pixels lined up vertically, denoted by p211 and p221, the
panel in the third row with two pixels lined up horizontally, denoted by p311 and
p312, the panel in the fourth row with two pixels also lined up horizontally, denoted
by p411 and p412, and the panel in the fifth row with two pixels lined up vertically,
denoted by p511 and p521. Since the size of the panels in the third column is 1 1 m,
they cannot be seen visually from Fig. 1.10a because its size is less than the
1.56 m pixel resolution.
It is worth noting that panel pixel p212, marked yellow in Fig. 1.10b, is of
particular interest. Based on the ground truth, this panel pixel is not a center
panel pixel representing a pure panel pixel. It is marked by a Y pixel to indicate
that it is a boundary panel pixel. However, in our extensive and comprehensive
experiments, conducted in later chapters, this Y panel pixel is always extracted as
the one with the most spectrally distinct signature compared to the R panel pixels in
row 2. This indicates that a signature of spectral purity is not equivalent to a
signature of spectral distinction. As a matter of fact, in many cases, panel pixel
p212 instead of panel pixel p221 is the first panel pixel extracted by endmember
finding algorithms to represent the panel signature in row 2, as will be demonstrated
experimentally in many chapters throughout the book. Also, because of such
ambiguity, the panel signature in the second row can be considered to be
represented by either p221 or p212. This implies that the ground truth of R panel
pixels in the second row in Fig. 1.10b may not be as pure as was originally thought.
Figure 1.11 plots the five panel spectral signatures obtained from Fig. 1.10b,
where the ith panel signature, denoted by pi, was generated by averaging the red
panel center pixels in row i. These panel signatures will be used to represent the
target knowledge of the panels in each row.
7000
P1
6000 P2
P3
P4
5000 P5
Radiance
4000
3000
2000
1000
0
0 20 40 60 80 100 120 140 160 180
Band
tree
road
A visual inspection and the ground truth in Fig. 1.10a, b also reveal four
background signatures (Fig. 1.12), which can be identified and marked interferer,
grass, tree, and road. The interferer is a large rock located in the woods/forest. The
trees are selected from the woods/forest located at the left edge. The road runs
vertically along the right edge of the woods/forest. The grass is a dirt field where
panels are located. These four signatures, along with the five panel signatures in
Fig. 1.11, can be used to form a nine-signature matrix for a linear mixing model to
perform supervised LSMA.
The data to be studied were collected by the Hyperion sensor mounted on the Earth
Observer 1 (EO-1) satellite. Hyperion uses a high-resolution hyperspectral imager to
record Earth surface images of approximately 7.5 100 km with a 30 m spatial
resolution and a 10 nm spectral resolution. Image scene EO1H0150322011201110K3
was downloaded from the USGS Earth Explorer website (EarthExplorer. 01 Sept.
2013 http://earthexplorer.usgs.gov/). The image used is a Hyperion L1 data product,
which includes 198 channels of calibrated spectral information. This image was then
cropped to select a region of interest (ROI) covering the western side of the Chesa-
peake Bay Bridge. The resulting image cube is a square with spatial dimensions of size
64 64 pixels and 198 spectral bands ranging from 426 to 2396 nm. The image was
then compared with higher-resolution areal imagery to identify five distinct areas of
interest (AOIs). The areas include a beach, Westinghouse Bay, Mezick Ponds,
farmland, and a large corporate building. Four pixels were selected at random from
each of these areas and used to generate an average spectral signature for the material
contained in the area, with the exception of the farmland, which used seven pixels
owing to the fact that it is composed of four disjoint regions in the image where four
pixels were selected from the largest farmland region and a single pixel was chosen
from the three remaining regions. A color image of the region and the spectral
signatures of each of the five AOIs are shown in Fig. 1.13.
24 1 Introduction
a b
c
10000
Beach
9000 Westinghouse Bay
Mezick Ponds
8000 Farm Land
Corporate Building
7000
6000
5000
4000
3000
2000
1000
0
0 50 100 150
Fig. 1.13 (a) Cropped Hyperion image scene EO1H0150322011201110K3. (b) Cropped Hype-
rion image scene EO1H0150322011201110K3 with five areas of interest (AOIs): beach (dark
blue), Westinghouse Bay (green), Mezick Ponds (orange), farmland (light blue), and a corporate
building (magenta). (c) Average spectral signatures for each of the five AOIs in (b)
Since real images generally do not have complete ground truth about the
endmembers, we must rely on synthetic images that are simulated by complete
knowledge to conduct quantitative analysis for performance evaluation.
Several synthetic images recently developed in Chang (2013, 2016) and Wu
et al. (2009) can be used for this purpose. These synthetic images were
1.6 Synthetic Images to Be Used in this Book 25
a b
BKG
Fig. 1.14 (a) Cuprite Mining District AVIRIS image scene; (b) spatial positions of five pure
pixels corresponding to minerals A, B, C, K, and M; (c) five mineral reflectance spectra and
background signature (b); (d) five mineral radiance spectra and background signature in (b)
custom-designed and simulated based on the Cuprite Mining District image data
available on the USGS website http://aviris.jpl.nasa.gov/. This Cuprite Mining
District image scene (Fig. 1.4b) can be used to simulate synthetic images for our
study. Although there may be more than five minerals in the data set in Fig. 1.4a, the
ground truth available for this region only provides the locations of the five pure
pixels: alunite (A), buddingtonite (B), calcite (C), kaolinite (K), and muscovite
(M) (Fig. 1.4b). For clarity, Figs. 1.4a, b and 1.6a, b are reproduced in Fig. 1.14a–d,
which will be used to simulate synthetic images with various scenarios. An area
marked “BKG” in the upper right corner of Fig. 1.14a was selected to find its
sample mean, that is, the average of all pixel vectors within the area BKG, denoted
by b and plotted in Fig. 1.14b, which will be used to simulate the background for the
image scene in Fig. 1.15.
The synthetic image designed here simulates 25 panels shown in Fig. 1.15, with
5 panels in each row simulated by the same mineral signature and 5 panels in each
column having the same size.
Among the 25 panels are five 4 4 pure-pixel panels for each row in the first
column and five 2 2 pure-pixel panels for each row in the second column, the five
2 2 mixed pixel panels for each row in the third column, and the five 1 1
subpixel panels for each row in the fourth column and the fifth column, where the
mixed and subpanel pixels were simulated according to the legends in Fig. 1.15.
Thus, a total of 100 pure pixels (80 in the first column and 20 in the second column),
referred to as endmember pixels, were simulated in the data by the five
endmembers, A, B, C, K, and M. The BKG in Fig. 1.14a is empirically selected
since it seems more homogeneous than other regions. Nevertheless, other areas can
also be selected for the same purpose. This b-simulated image background is
further corrupted by an additive noise to achieve a certain SNR, which is defined
26 1 Introduction
c
7000
Kaolinite
5000
Reflectance
4000
3000
Buddingtonite
Background
2000
Calcite
1000
0
500 1000 1500 2000 2500
Wavelength (nm)
d
12000
Mean of five signatures
Kaolinite Muscovite
10000
Alunite
8000
Radiance
Buddingtonite
6000
4000 Calcite
2000
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
The first type of target insertion is a target implantation (TI), which can be
simulated by inserting clean target panels into a clean image background by
replacing their corresponding background pixels, referred to as TI1, and into a
clean image background corrupted by additive Gaussian noise by replacing their
corresponding background pixels, referred to as TI2. In both scenarios, there are
still 100 pure panel pixels, all located in the first and second columns. A third
scenario, referred to as TI3, is simulated by adding a Gaussian noise to TI1, in
which case the clean panels are corrupted by an additive Gaussian noise. As a result,
there are no pure signatures in TI3. The design of TI scenarios is to see how the
Gaussian noise term, n, plays in model (1) to respond to three criteria, OP, simplex
volume, and fully constrained abundance fractions, where panels are clean in TI1
and TI2 while clean panels in TI3 are contaminated by additive Gaussian noise.
A second type of target insertion is target embeddedness (TE), which can also be
simulated by embedding clean target panels into a clean image background plus
additive Gaussian noise by superimposing target pixels over the background pixels.
Like TI, we can also simulate TE1, TE2, and TE3 as counterparts of TI1, TI2, and
TI3, respectively, with the only key difference among them being how target panels
are inserted. With TE, there are no endmembers present in TE1, TE2, and TE3
scenarios. The need to design TE scenarios is in seeing that adding a background
signature b to a standard signal detection model,
r ¼ s þ n, ð1:2Þ
yields
r ¼ s þ b þ n, ð1:3Þ
which is the same signal/background + noise (SBN) model used by Thai and Healey
(2002). By virtue of (1.3) it is a signal detection model with abundance sum-to-one
constraint (ASC) violated. In this case, it is interesting to find how a pure signature
as an endmember can be found through various algorithms.
28 1 Introduction
Since TI2 and TE2 are of major interest for the experiments conducted in this
book, they will be referred to as TI and TE for simplicity throughout the book,
without the suffix 2.
Although this book can be studied chapter by chapter, it can also be used as a
reference book without having to go back and forth between chapters. To make
each chapter standalone and self-contained, the image scenes used for the experi-
ments are briefly described in each chapter, with more details in Sects. 1.6 and 1.7
in this chapter. In addition, many descriptions and statements may be repeated and
reiterated in the introduction of each chapter, so that readers do not have to waste
time flipping through other chapters looking for the information they need. Thus,
readers who already possess knowledge on a particular topic can skip these descrip-
tions and go straight to the desired sections. Finally, the book is arranged in five
different parts such that each part not only can be read independently but is also
closely tied to the other parts. Most importantly, each part is arranged as an
integrated entity on a specific theme.
Since this book deals primarily with real hyperspectral data, image pixels are
generally mixed and not necessarily pure, and the term endmember is not used;
instead, a general term like signature or signature vector is used. In addition,
because we are only interested in target analysis, the term targets instead of
materials is used throughout the book. To draw a distinction between a target
pixel and its spectral signature vector, we use the notation t to represent the target
pixel, r to denote an image pixel vector, and s or m to indicate its spectral signature
vector. We also use bold uppercase for matrices and bold lowercase for vectors.
When a vector is used in the text, it is generally described by a row vector specified
by an open parenthesis “()” with transpose “T” as a superscript. When a vector is
used in mathematical derivations, it is generally described by a column vector with
closed brackets “[]” instead of “()” for better illustration. In other words, open
parenthesis “(.)” and bracket “[.]” are interchangeable in vector representation
whichever is more appropriate. An italic uppercase L will be used for the total
number of spectral bands, K for the sample spectral covariance matrix, and
R for the sample spectral correlation matrix. Also, δ*(r) is used to represent a
detector or classifier that operates on an image pixel vector r, where the superscript
“*” in δ*(r) specifies what type of detector or classifier is to be used. It should be
noted that δ*(r) is a real-valued function that takes the form of an inner product of a
T
filter vector w with r, that is, δ* ðrÞ ¼ w* r, with the filter vector w* specified by a
particular detector or classifier. We also use α and α ^ to represent the abundance
vector and its estimate, where the hat over the α denotes “estimate.”
Part I
Fundamentals
Abstract Using maximal simplex volume (SV) as an optimal criterion for finding
endmembers is a common approach and has been widely adopted in the literature.
However, very little work has been reported on how SV is calculated because it can
be simply done by finding the determinant of a matrix formed by vertices of a
simplex. Interestingly, it turns out that calculating SV is much more complicated
and involved than one might think. This chapter investigates this issue from two
different aspects, through eigenanalysis and a geometric approach. The
eigenanalysis takes advantage of the Cayley–Menger determinant to calculate the
SV, referred to as determinant-based SV (DSV) calculation. The major issue with
this approach is that when the matrix is of ill-rank, calculating the determinant runs
into a singularity problem. To deal with this issue, two methods are generally
considered. One is to perform data dimensionality reduction to make the matrix
of full rank. The drawback of this method is that the original volume has been
shrunk and the found volume of a dimensionality-reduced simplex is not the true
original SV. The other is to use singular value decomposition to find singular values
for calculating the SV. An issue arising from this method is its instability in
numerical calculations. An alternative to eigenanalysis is a geometric approach
derived from the simplex structure whose volume can be calculated by the product
of the base and height of a simplex. The resulting approach is called geometric SV
(GSV) calculation. This chapter explores DSV and GSV calculations, with further
discussions in Chaps. 11 and 12.
2.1 Introduction
Endmembers are defined as pure spectral signatures and provide very important
information for hyperspectral data exploitation. Accordingly, finding endmembers
has attracted considerable interest in recent years (Chang 2003a, b, 2013, 2016, and
references therein). Using maximal simplex volume (SV) as an optimal criterion is
a common practice in finding endmembers owing to the fact that the purity of
endmembers can be specified by the geometric convexity of a simplex. Therefore,
many well-known SV-based endmember finding algorithms (EFAs) have been
Since the matrix in (2.1) is generally not a square matrix, DSV computation cannot
be done directly by simply multiplying eigenvalues because the matrix is of ill rank.
Three comments on finding DSV via (2.1) are in order:
1. For a given set of p endmembers in a hyperspectral image with band dimen-
sionality, L, the dimensionality of a simplex Sp formed by these p endmembers
has at most p 1. Because p is usually much smaller than full spectral dimen-
sionality, L, calculating DSV via (2.1) generally requires DC/DR to reduce L to
p 1 to find the determinant of a square matrix. Once a simplex Sp is reduced to
( p 1)-dimensional spectral space, the matrix used to calculate DSV in (2.1)
becomes a ðp 1Þ ðp 1Þ square matrix, in which case DSV is simply the
product of the eigenvalues of the reduced square matrix.
2. According to the LU factorization of n n matrix A (Moon and Stirling 2000),
the determinant of A, |A|, can be calculated as
Y
A ¼ PLU ¼ PLU ¼ n uii ; ð2:2Þ
i¼1
v2 h
v1
b c
v3
h
v3
v2
v2
v1
v1
Fig. 2.1 (a) Two-dimensional three-vertex simplex S3 formed by two edge vectors v1, v2; (b)
three-dimensional four-vertex simplex S4 formed by three edge vectors v1, v2, v3; (c) three-
dimensional parallelotope (or parallelepiped) P3 formed by the same edge vectors
h3
x
h1
y h2
36 2 Simplex Volume Calculation
Proof Let Sj+1 be a ( j + 1)-vertex simplex, and Sj is its base which is a j-vertex
simplex. Assume that hj is the height of Sj+1, which is the perpendicular distance
(altitude or height) of the apex of a j-vertex simplex Sj contained in a subspace
linearly spanned by hSji. For example, h2 is the height of a three-vertex S3 with
two-vertex simplex S2 formed by connecting two data sample points as its base.
Similarly, h3 is the height of S4 with a three-vertex simplex S3 formed by three data
sample points as a triangle as its base. The volume DSV(Sj+1) of a ( j + 1)-vertex
simplex Sj+1 given by (2.1) can be reexpressed by GSV(Sj+1) of a ( j + 1)-vertex
simplex Sj+1 with a j-vertex simplex Sj as its base and hj as its height, as follows:
ð hj
j1
ð hj j1
h h¼0 h
GSV Sjþ1 V Sjþ1 ¼ V Sj dh ¼ GSV Sj j1
h¼0 hj hj for j2
hj
hj h¼0 hj j hj
¼ GSV Sj j1 ¼ GSV Sj j1 ¼ GSV Sj :
j hj hj j
ð2:5Þ
Note that (2.5) is not defined for j ¼ 1 since GSV(S1) ¼ 0. However, if we further
define h1 ¼ GSV(S2) as the initial condition, which is the height of a zero-
dimensional one-vertex simplex, S1, then according to (2.5) GSV(Sj+1) can be
calculated by the previous V(Sj) for j 2 recursively as
hj
GSV Sjþ1 ¼ GSV Sj
j ð2:6Þ
¼ ðGSVðS2 Þ=j!Þhj hj1 h2 ¼ ð1=j!Þhj hj1 h2 h1 ;
□
Using Theorem 2.1 we can further derive the following theorem.
2.3 Geometric Simplex Volume Calculation 37
By virtue of (2.8) and (2.9) the GSV of a ( j + 1)-vertex simplex Sj+1, V(Sj+1), can
be expressed as
GSV Sjþ1 ¼ ð1=j!ÞV Pj ¼ ð1=j!ÞDetðME Þ: ð2:10Þ
It is also worth noting that the volume derived by (2.12) is only applicable when
ME is a square matrix. More details on (2.12) can be found in Li et al. (2015a, b).
When no DR is performed, we must deal with the issue of an ill-ranked matrix
determinant for a nonsquare matrix in (2.12). In this case, SVD can be used.
Generally, SVD can be applied to any m n matrix such as ME with the following
two valid relations:
1. ME * ME ¼ VΣ * U* UΣV* ¼ V Σ * Σ V* ;
2. ME ME * ¼ UΣ * V* VΣU* ¼ U Σ * Σ U* :
38 2 Simplex Volume Calculation
< m 2 – m1 , m 3 – m1 >
m 2 – m1 m2 P[^m – m ]
2 1
m1 A
m3
t3
and
n o
ATGP
tjþ1 ¼ arg maxr P⊥
Uj r ; ð2:15Þ
with Uj ¼ m1 m2 mj . Then
ATGP, OP SGA, OP
t ¼ t ¼ h* ð2:16Þ
jþ1 jþ1 j
and
Y j
V Sjþ1 ¼ GSV m1 ; . . . ; mj ; mjþ1 ¼ m1 m2 i¼2 1=i 1 ! h*i
Yjþ1
¼ m1 m2 i¼3 1=i 1 ! tSGA i
, OP
Yjþ1
¼ m1 m2 i¼3 1=i 1 ! tATGPi
, OP
;
ð2:17Þ
where h
j is the maximal height perpendicular to the simplex
Sj ¼ S m1 ; . . . ; mj1 ; mj . Note that h
j is defined as the height of the simplex
specified by Sjþ1 ¼ S m1 ; . . . ; mj1 ; mj ; mjþ1 . It is perpendicular to its base,
40 2 Simplex Volume Calculation
Sj ¼ S m1 ; . . . ; mj1 ; mj , and can be obtained using tSGA ATGP
jþ1 in (2.14) or tjþ1 in
(2.15).
Proof Using Fig. 2.3 as an illustrative example, we can proceed with the proof by
mathematical induction as follows.
1. Initial condition: j ¼ 2 (degenerated simplex)
Since the lowest-dimensional simplex is a one-dimensional two-vertex
degenerated simplex, let m1 and m2 be two initial endmembers to form a (m1,
m2)-degenerated simplex obtained by the line segment connected by m1 and m2,
denoted m1m2. There are two ways to obtain m1 and m2. One is to find a data
sample vector with maximal vector length as m1, and then m2 is a data sample
vector with maximal distance from m1. The other is find two data sample vectors
with maximal distance as m1 and m2. In either case, SV(m1,m2) ¼ ||m2 m1||.
To find the third endmember m3, m3 should be the one that yields the maximal
height or maximal OP perpendicular to the line segment m1m2. More specifi-
cally, let U2 ¼ ½m1 m2 , and hU2i is a subspace linearly spanned by two
endmembers, m1 and m2. Then any data sample vector r can be decomposed
into the form r ¼ rOP þ rU2 , where rOP 2 hU2 i⊥, which is orthogonal to the space
hU2i and rU2 2 hU2 i. Therefore, tATGP 3 and tSGA
3 can be decomposed into the form
ATGP
t3 ATGP, OP
¼ t3 þ t3 ATGP, U2 ATGP, OP
, with t3 2 hU2 i⊥ and tATGP
3
, U2
2 hU2 i, and tSGA
3
¼ t3SGA, OP þ tSGA
3
, U2
, with tSGA
3
, OP
2 hU2 i⊥ and tSGA
3
, U2
2 hU2 i, respectively, as
shown in Fig. 2.4, where there are three ways to represent the third endmember
m3 via the vector Om3.
(a) Om3 ¼ Om1 þ m1 m3 with m1 m3 ¼ m1 A þ Am3 , Am3 ¼ m3OP , m1 A ¼
OP
¼ h* ;
3 , m1 m3 ¼ m1 A þ Am3 , and m3
mU 2
2
(b) Om3 ¼ Om2 þ m2 m3 with m2 m3 ¼ m2 A þ Am3 , Am3 ¼ m3OP , m3 A ¼
OP
mU ¼ h* ;
3 and m3
2
2
(c) Om
OP3 ¼
OA*þ Am3 where OA ¼ Om1 þ m1 A, OA ¼ Om2 þ m2 A and
m ¼ h .
3 2
Note that the spatial locations of m1, m2, and m3 are included in the information
⊥
of m1 A ¼ mU 31 and m3 A ¼ m32 in the PU2 -specified hyperplane. As a result, none
2 U2
maxr V ðm
1 ; m2 ; rÞ ¼
GSVðm1 ; m2; m3 Þ ð2:18Þ
¼ ð1=2Þm1 m2 h*2 ¼ ð1=2Þm1 m2 t3ATGP, OP :
2.4 General Theorem for Geometric Simplex Volume Calculation 41
However,
maxr GSVðm1 ; m2 ; rÞ ¼ GSV m1 ; m2 ; tSGA
SGA, OP 3
ð2:19Þ
¼ ð1=2Þm1 m2 t3 :
Equating (2.18) to (2.19) yields h*2 ¼ t3ATGP, OP ¼ tSGA
3
, OP
. In this case, m3
can be either tATGP or t SGA
with the same maximal OP given by
ATGP, OP 3 SGA, OP 3
h2 ¼ t3
*
¼ t3 .
2. Now assume that (2.16) and (2.17) are true for any positive integer
j endmembers. Then we would like to prove that (2.16) and (2.17) are also
true for j + 1 endmembers.
Since hUji is the space linearly spanned by the j endmembers m1, m2, . . ., mj,
it is obvious that it also contains the simplex formed by m1, m2, . . ., mj, denoted
by S(m1, m2, . . ., mj) because a simplex is a linear convex set. In this case, the
height of the simplex S(m1, m2, . . ., mj) must be perpendicular to the simplex S
(m1, m2, . . ., mj), thus hUji. More specifically,
also using Fig.2.4OPasan illustra-
tive example, by replacing m3 with mj+1, m3OP ¼ h*2 with mjþ1 ¼ h* , and
j
m1, m2 with m1 , m2 , . . . , mj , any data sample vector r can be decomposed into
⊥
the form r ¼ rOP þ rUj , where rOP 2 Uj and rUj 2 Uj . Thus, the data sample
vector that yields the maximal height should the one lying in the hyperplane ⊥
specified by P⊥ P r
Uj r over all the data sample vectors r with maximal Uj
according to (2.15).
In other words,
GSV m1 ; m2 ; . . . ; mj ; tjþ1
ATGP
ATGP, OP
¼ GSV m1 ; m2 ; . . . ; mj1 ð1=jÞ tjþ1 ð2:6Þ
¼ GSV m1 ; m2 ; . . . ; mj ð1=jÞmaxhj hj ð2:15Þ
¼ GSV m1 ; m2 ; . . . ; mj1 ð1=jÞ h*j
m1 A
m2
U2
m 32
O
42 2 Simplex Volume Calculation
Y j ATGP, OP
¼ km 2 m 1 k ð 1= ð i 1 Þ Þ t ð1=jÞ
t ATGP, OP
ðby inductionÞ
i jþ1
Y j
i¼2
¼ ð1=j!Þkm2 m1 k i¼2 i
h* ð2:6Þ
¼ maxtjþ1 GSV m1 ; m2 ; . . . ; mj ; tjþ1 ¼ GSV m1 ; m2 ; . . . ; mjþ1 : ð2:20Þ
Y j
¼ ð1=j!Þkm2 m1 k h *
i¼2 i
ð2:6Þ
ð2:21Þ
ATGP, OP
¼ GSV m1 ; m2 ; . . . ; mj ; tjþ1 ð2:15Þ:
ATGP
By virtue of (2.20) and (2.21), both tjþ1 and tSGA
jþ1 yield the same maximal SV,
Yj1 *
that is, ð1=j!Þm2 m1 h , by (2.6). More specifically, the OP of
i¼2 i
ATGP
tjþ1 and tSGA projected on the P⊥
Uj -specified hyperplane should be the same,
ATGP, OP jþ1 SGA, OP
t ¼ t . □
jþ1 jþ1
1 1 1
GSV Sjþ1 ¼ ð1=j!Þ DetðME Þ ¼ ð1=j!ÞDet
m1 m2 mjþ1
¼ ð1=j!ÞDet½ m
e2 m e3 m e jþ1 ð2:22Þ
Y j
¼ ð1=j!Þm2 m1 i¼3
h *
i ¼ GSV S jþ1 :
a b c
7 1 8
6 7
0.5
5 6
4 0
z
z
3 5
-0.5 4
2
1 -1 3
10 1 8
8 6.8 7 0.5 1 6 8
6 6.6 0 0.5 4 7
0 6
y 4
2 6 6.2 6.4 x y -0.5 -1 -1 -0.5 x y 2 0 4
5
x
Fig. 2.5 (a) Triangle; (b) regular triangle; (c) 3D simplex with one vertex in origin, in 3D space
44 2 Simplex Volume Calculation
Table 2.1 Comparative analysis of volume among various methods of simplexes in three-
dimensions
Method
SVD- PCA- PCA-
Simplex GSV DSV DSV GSV DSV
Arbitrary 2D simplex (Fig. 2.5a) 21.8518 N/A 155.5426 21.8518 21.8518
Regular 2D simplex (Fig. 2.5b) 1.1547 N/A 1.2172 1.1547 1.1547
Arbitrary 3D simplex (Fig. 2.5c) 15.8333 15.8333 15.8333 N/A N/A
data dimensionality of three, this four-vertex simplex has the same dimensionality
as its data dimensionality, three.
Table 2.1 tabulates the values of a SV calculated by various versions of SV
calculation methods using GSV and DSV, where “N/A” indicates “not applicable.”
As we see in Table 2.1, all three methods, GSV, DSV, and SVD-DSV, yield the
same SV value when the dimensionality of the simplex equals the data dimension-
ality. In this case, no dimensionality reduction is needed, and SVD could be used as
an alternative method to obtain the SV. But this result does not justify the use of
SVD for calculating the volumes of simplexes in a different dimensional space.
However, in all cases, GSV still yields the same SV as DSV when the dimension-
ality of a simplex is smaller than that of the data space, where the SV values
calculated using SVD-DSV are not consistent.
A real image scene in Fig. 2.6 (also shown in Fig. 1.10), acquired by the airborne
HYperspectral Digital Imagery Collection Experiment (HYDICE) sensor in August
1995 from a flight altitude of 10,000 feet with a ground sampling distance of
approximately 1.56 m, was used for experiments. It has 210 spectral channels
ranging from 0.4 to 2.5 μm with a spectral resolution of 10 nm. The low signal/
high noise bands, bands 1–3 and 202–210, and the water vapor absorption bands,
bands 101–112 and 137–153, were removed. Thus, a total of 169 bands were used
for the experiments. The scene has a size of 64 64 pixel vectors.
To compare the SV values calculated by GSV with and without PCA as a DR,
the SGA developed by Chang et al. (2006a, b) was used to find a set of data sample
vectors known as endmembers from the image. The largest simplex is formed by a
set of 170 vertices specified by 170 data sample vectors found using SGA in a
169-dimensional data space. Since SGA finds endmembers progressively one after
another starting from a single endmember as a zero-vertex m0 simplex up to
169
170 endmembers mj j¼0 as a 169-vertex simplex, we used these progressively
found endmembers to form a set of growing simplexes one by one, progressively
denoted by Sj, j ¼ 0, . . . , 169, where j is the number of endmembers found in order
by SGA. S0 is formed by a single vertex m0, S2 is a line segment formed by two
2.6 Real Image Experiments 45
a b
70 12 64
GSV x 10 GSV
60 PCA-DSV PCA-DSV
10
50
40 8
Log10(SV)
SV
30
6
20
10 4
0
2
–10
–20 0
0 20 40 60 80 100 120 140 160 180 0 20 40 60 80 100 120 140 160 180
number of vertices number of vertices
Fig. 2.7 Comparison of SV values calculated by (a) GSV versus PCA-DSV and (b) PCA-GSV
versus PCA-DSV
1.4
GSV
1.2 PCA-DSV
Accumulated time (sec)
0.8
0.6
0.4
0.2
0
0 20 40 60 80 100 120 140 160 180
number of vertices
Fig. 2.8 Comparisons of cumulative computing times required by GSV and PCA-DSV
hðSjþ1 Þ
than the dimensionality of the simplex Sj+1, j + 1, that is, jþ1 < 1 . This peak
occurred at the 73rd endmember and was the 78th endmember when DR was
performed.
To further compare the performance of each method, a comparative analysis of
the computational complexity is provided in Table 2.2, where p is the number of
endmembers used to specify the vertices of simplexes and L is the data
dimensionality.
Since GSV only involves inner products, it has a complexity of OðpLÞ compared
to DSV, which has a complexity of Oðp3 Þ. In real-world problems, high data
dimensionality is always an issue. When a DR technique or SVD is performed
prior to SV calculation, thecomputational
complexity of implementing PCA-DSV
and SVD-DSV becomes O pL þ Oðp Þ and Oðp2 LÞ, respectively.
2 3
A major advantage of GSV over DSV is that GSV calculates SV using (2.6),
which only uses mathematical products. This cannot be done using DSV since the
volume of Sj can only be calculated one at a time. Figure 2.8 shows the accumulated
time required by GSV to calculate SV by multiplying height h(Sj+1) ¼ hj by the SV
of the previous j-vertex simplex Sj compared to the accumulated time required by
PCA-DSV, where the computing time was obtained by an average of 100 trials. The
experimental results show that GSV can significantly reduce the computing time of
SV calculation.
2.7 Conclusions 47
2.7 Conclusions
SV has been widely used as a criterion for algorithm design to find endmembers.
Since the number of endmembers is generally much smaller than the data dimen-
sionality, the matrix formed by endmembers, called the endmember matrix used to
calculate SV via its matrix determinant, is usually not of full rank. As a result,
calculating the determinant of an endmember matrix runs into a singularity prob-
lem. Two approaches are used to deal with this issue. One approach involves using
DC or DR, which reduces the original data space to a lower-dimensional data space
to make an endmember matrix of full rank. Unfortunately, as shown in Li
et al. (2015a, 2015b) and Li et al. (2016a, b), the SV found by a dimensionality-
reduced data space is not the true SV obtained directly from the original data space.
The other approach is to use SVD or the Moore–Penrose pseudo-inverse to find
SV. While the SV found by this approach is correct, it suffers from two issues,
numerical instability and excessive computing time when data dimensionality and
the number of endmembers are high. Interestingly, the issue of finding the true SV
has been overlooked by all SV-based endmember finding algorithms, including a
well-known N-FINDR, SGA. This chapter investigates this important issue and
further develops a GSV-based method to resolve the dilemma. The proposed
GSV-based method has several benefits that a DSV-based method cannot provide.
One is that no DR is required. Another is that GSV can calculate the true SV
without experiencing the numerical issues. Third, the computational complexity is
significantly reduced. Last but not least, when GSV is applied to endmember
finding algorithms, it can avoid finding incorrect endmembers caused by numerical
errors resulting from use of the matrix determinant to calculate the SV.
Chapter 3
Discrete-Time Kalman Filtering
for Hyperspectral Processing
3.1 Introduction
processed; and (3) innovations information, which retains only information that is
contained in the new data sample (i.e., the current data sample) but that cannot be
obtained from the processed information. One way to satisfy these requirements is
through the use of a recursive process. In statistical signal processing, the Kalman
filter (KF) is probably the most widely used linear operator that processes data
samples recursively in a very effective manner. It implements two equations, a state
equation, which updates internal states via the state transition matrix, and a mea-
surement or input/output equation, which updates outputs via the Kalman gain
obtained by error covariance matrix and state estimates. Since the update equations
derived for state estimates and outputs can be implemented causally and recur-
sively, the most appealing and attractive feature of the KF in real-world applica-
tions is its ability and capability in real-time processing.
Kalman filtering has been widely used in statistical signal processing for the
purpose of parameter estimation (Poor 1994; Gelb 1974). It makes use of a
measurement equation (input/output equation) and a state equation (process equa-
tion) to recursively estimate and update states and outputs. Here, we follow the
treatment in Gelb (1974) and briefly describe its idea with detailed derivations.
Suppose that there is a discrete-time linear dynamic system specified by the
white noise vectors {u(k)}, output vectors {y(k)} and system gain matrix {H(k)}.
Now let {x(k)} be the system state vectors that describe the states of the system over
time. Two equations are of interest. One is the state equation given by
where Φ(k + 1,k) is the state transition matrix, E½uðkÞuT ðlÞ ¼ QðkÞδkl, and δkl is the
delta function. Figure 3.1 depicts a diagram of it.
A second equation is the measurement equation, which is the output of the
system resulting from the state {x(k)} given by
As a matter of fact, when the system is observed, it is also corrupted by white noise
{v(k)} to yield the observation equation given by
x(k) x(k+1)
F (k+1,k)
3.2 Discrete-Time Kalman Filtering 51
where E½vðkÞvT ðlÞ ¼ RðkÞδkl and E½uðkÞvT ðlÞ ¼ CðkÞδkl , with E[.] being the
expectation operator. Figure 3.2 illustrates the final system output {y(k)} from an
observation process given by Fig. 3.2.
As an example, assume that a random process, {x(k)} can be observed only
through an observation process, {z(k)}. Let an estimate of x(k) at the time instant tk
Xk
k
based on k observations fzðiÞgi¼1 be given by ^x ðkÞ ¼ 1k i¼1
zðiÞ. Thus, with k + 1
observations fzðiÞgkþ1
i¼1 ,
X
1 Xkþ1 k 1 k 1
^
x ð k þ 1Þ ¼ z ð i Þ ¼ zð i Þ þ z ð k þ 1Þ
k þ 1 i¼1 k þ 1 k i¼1 kþ1
k 1 ð3:4Þ
¼ ^x ðkÞ þ z ð k þ 1Þ
kþ1 kþ1
1
¼x^ ðk Þ þ ðzðk þ 1Þ x^ ðkÞÞ:
kþ1
P ( k – 1 | k – 2) P ( k – 1 | k – 1) P ( k | k – 1) P ( k | k )
xˆ ( k – 1 | k – 2) xˆ ( k – 1 | k – 1) xˆ ( k | k – 1) xˆ ( k | k )
x kk can be considered an updated estimate of ^x kk 1 using the
case, ^
incoming measurement at time instant tk, z(k) (Figs. 3.2 and
3.3).
In other words, given an a priori estimate, ^x kk 1 , we are interested in
updating the state estimate, ^x kk , in linear and recursive forms to avoid using a
growing memory filter as follows:
^x kk ¼ KðkÞ^x kk 1 þ KðkÞzðkÞ; ð3:5Þ
where both KðkÞ and K(k) are time-varying matrices yet to be determined. Define
the estimation error vector
x kk ¼ xðkÞ ^x kk ) ^x kk ¼ xðkÞ þ e
e x k k ; ð3:6Þ
x kk 1 ¼ xðkÞ ^x kk 1 ) ^x kk 1 ¼ xðkÞ þ e
e x kk 1 ; ð3:7Þ
x kk ¼ KðkÞ þ KðkÞHðkÞ I xðkÞ þ KðkÞx kk 1 þ KðkÞvðkÞ: ð3:8Þ
e
By definition, E½vðkÞ ¼ 0. Also, if E xe kk 1 ¼ 0, the estimate ^x kk in (3.6)
x kk ¼ 0 which
is an unbiased estimate for any given state vector x(k), that is, E e
means that only if the term in brackets in (3.8) equal to zero as follows
or, alternatively,
x kk ¼ ^x kk 1 þ KðkÞ zðkÞ HðkÞ^x kk 1 ;
^ ð3:11Þ
x kk ¼ ðI KðkÞHðkÞÞ^x kk 1 þ KðkÞvðkÞ:
e ð3:12Þ
3.2 Discrete-Time Kalman Filtering 53
T
Since P kk 1 ¼ E e
x k k 1 e
x kk 1 and
x kk 1 vT ðkÞ ¼ E vðkÞe
E e x T k k 1 ¼ 0 ð3:14Þ
due to the fact that the measurement error is uncorrelated, (3.13) is reduced to
P kk ¼ ðI KðkÞHðkÞÞP kk 1 ðI KðkÞHðkÞÞT
þ K k RðkÞKT ðkÞ: ð3:15Þ
which is equivalent to minimizing the length of the estimation error vector. To find
the optimal value of K(k), we differentiate (3.17) using (3.15) with respect to K(k)
via the relation
∂
trace ABAT ¼ 2AB ð3:18Þ
∂A
and obtain
2ðI KðkÞHðkÞÞP kk 1 HT ðkÞ þ 2KðkÞRðkÞ ¼ 0: ð3:19Þ
The extrapolation of the state estimate vector and error covariance matrix
between two measurements can be performed by (3.1) as follows:
^x kk 1 ¼ Φðk, k 1Þ^x k 1k 1 ; ð3:22Þ
P kk 1 ¼ Φðk, k 1ÞP k 1k 1 þ Qðk 1Þ; ð3:23Þ
K kk 1 ¼ Φðk, k 1ÞKðk 1Þ: ð3:24Þ
h
2 i
then ^y kk 1 ¼ arg min 0 E y0 kk 1 yðkÞ , if and only if
y k k1
k 1: ð3:26Þ
h 2 i
E y0 k k 1 yð k Þ
h 2 i
¼ E y0 kk 1 ^y kk 1 þ ^y kk 1 yðkÞ
h 2 i
¼ E y0 kk 1 ^y kk 1 ð3:27Þ
þ 2Eh y0 kk 1 ^y kki 1 ^y kk 1 yðkÞ
2
þ E ^y kk 1 yðkÞ :
Since y0 kk 1 ^y kk 1 is also a linear combination of the observations
zðlÞ0 l k 1 , the middle term in (3.27) is E y0 kk 1 ^y kk 1 Þ
^y kk 1 yðkÞ ¼ 0, according to (3.26). This implies that
h 2 i
E y0 k k 1 yð k Þ
h 2 i h 2 i
¼ E y0 kk 1 ^y kk 1 þ E ^y kk 1 yðkÞ ð3:28Þ
h 2 i
E ^ y k k 1 yð k Þ :
Because y0 kk 1 is arbitrarily chosen, it proves the sufficiency of (3.26).
Necessity of (3.26)
Suppose that y0 kk 1 is a linear estimate based on the observations
Now we construct a new linear estimate y kk 1 by
h T i
E yðkÞ y0 kk 1 zðjÞ
y k k 1 ¼ y0 k k 1 þ zðjÞ: ð3:30Þ
E½z2 ðjÞ
h 2 i
¼ E yð k Þ y0 k k 1
h T i
E yðkÞ y0 kk 1 zðjÞ h T i
0
2 E y ð k Þ y k k 1 zð j Þ
E½z2 ðjÞ
0 h T i 12
E yðkÞy0 kk1 zðjÞ
þ@ E½z2 ðjÞ
A ðE½zðjÞÞ2
h T i2
h 2 i E yðkÞ y0 kk 1 zðjÞ
¼ E yð k Þ y0 k k 1 ð3:31Þ
h i E½z2 ðjÞ
< E yðkÞ y0 kk 1
2
;
which implies that y kk 1 is a better linear estimate than y0 kk 1 . Thus,
y0 kk 1 cannot satisfy (3.26).
□
By virtue
of the
orthogonality principle we can further
define
an innovation
process e kk 1 resulting from the estimate of ^y kk 1 by
e kk 1 ¼ zðkÞ ^y kk 1 ¼ zðkÞ HðkÞ^x kk 1 ; ð3:32Þ
where ^ y kk 1 is the least-squares estimate of y(k) given the observation
zðlÞ0 l k 1 . As a result, E e kk 1 ¼ 0 and
T
E e kk 1 e kk 1 ¼ Py kk 1 þ RðkÞ δkl , where Py(k|k 1) is the
covariance matrix of ey kk 1 in the estimate of ^y kk 1 and given by
h T i
P y k k 1 ¼ E e
y k k 1 e
y k k 1 : ð3:33Þ
Innovations Theorem e kk 1 ¼ zðkÞ ^y kk 1 is a white noise process
with
the same covariance matrix as v(k), that is,
E e kk 1 eT ll 1 ¼ E½vðkÞvT ðlÞ:
v(k)
F (k,k-1) D
z(k)
xˆ ( k | k – 1)
e(k) - yˆ ( k | k – 1)
H(k)
K(k) F (k+1,k)
xˆ ( k | k – 1)
predictor xˆ ( k + 1 | k ) D
xˆ ( k | k – 1)
filter xˆ ( k | k )
Xk 1
^x k þ 1k ¼ E xð k þ 1 Þe T
l l 1 P y ð l Þ þ R ðl Þ e l l 1
Xk1 l¼0
1
¼ E xðk þ 1ÞeT ll 1 Py ðlÞ þ RðlÞ e ll 1 þ E xðk þ 1ÞeT kk 1
l¼0 1
Py ðkÞ þ RðkÞ e kk 1
¼ Φðk þ 1:kÞ^ x kk 1 þ KðkÞe kk 1 ;
ð3:34Þ
discrete-time KF:
x kk ¼ ^x kk 1 þ KðkÞe kk 1
^
¼ ^x kk 1 þ KðkÞzðkÞ ^y kk
1 ð3:35Þ
¼ ^x kk 1 þ KðkÞ zðkÞ HðkÞ^x kk 1 ;
or, alternatively,
x kk ¼ Φðk, k 1Þ^x k 1k 1 þ KðkÞe kk 1
^
¼ Φðk, k 1Þ^x k 1k 1 þ KðkÞzðkÞ ^y kk
1 ð3:36Þ
¼ Φðk, k 1Þ^x k 1k 1 þ KðkÞ zðkÞ HðkÞ^x kk 1 ;
1
KðkÞ ¼ E xðk þ 1ÞeT kk 1 Py kk 1 þ RðkÞ ; ð3:37Þ
T
with Py kk 1 ¼ E e y k k 1 e
y kk 1 ¼ HðkÞP kk 1 HT ðkÞ:
As a result of (3.35)–(3.37), K(k) in (3.37) can be reexpressed as the following
recursive equations:
1
KðkÞ ¼ P kk 1 HT ðkÞ HðkÞP kk 1 HT ðkÞ þ RðkÞ ; ð3:38Þ
where
P k þ 1k ¼ Φðk þ 1, kÞA kk 1 ΦT ðk þ 1, kÞ þ QðkÞ; ð3:39Þ
T
A kk 1 ¼ P kk 1 KðkÞ Py kk 1 þ RðkÞ K ðkÞ; ð3:40Þ
x k þ 1k ¼ Φðk þ 1, kÞ^x kk 1 þ KðkÞe kk 1 ;
^ ð3:41Þ
KðkÞeðkÞ Φðkþ1, kÞ Kðkþ1Þeðkþ1Þ
x kk 1 ! ^x kk
^ ! ^x k þ 1k ! ^x k þ 1k þ 1 :
ð3:42Þ
present in the image with appropriate abundance fractions, α1, α2, . . . , αp with αj
corresponding to the abundance fraction of the jth endmember mj, as follows:
where δnk is Kronecker’s notation defined by δkk ¼ 1 for k ¼ l and δkl ¼ 0 for l 6¼ k.
Similarly, we can additionally assume that the noise v(k) in (3.45) is also an
additive white noise given by
E vðkÞ vT ðlÞ ¼ RðkÞ ¼ σ 2v δkl Ipp : ð3:46Þ
k
rðiÞi¼1 ^ k þ 1k denote the minimum MSE prediction of the abundance
. Let α
^ kk denote the minimum MSE estimate of
vector α(k) at the position k + 1 and α
k
the abundance vector α(k) at position k based on all image pixel vectors rðiÞi¼1 that
have already been visited up to position k. Similarly, P k þ 1 k and P kk
represent the one-step MSE prediction error covariance matrix and MSE estimation
error covariance matrix at position k. Then a KF that performs LSU can be
described in the following procedure.
• Initial conditions:
and Q(k),
R(k)
^ 0 1 ¼ E½αð0Þ is the mean of α(0),
α
P 0 1 ¼ Cov½αð0Þ is the covariance matrix of α(0);
• Kalman gain at k:
1
KðkÞ ¼ Φðk þ 1, kÞP kk 1 MT MP kk 1 MT þ R
• Abundance update at k:
^ k k ¼ α
α ^ kk 1 þ KðkÞ rðkÞ M^
α k k 1 ;
• Abundance prediction at k + 1:
^ k þ 1k ¼ Φðk þ 1, kÞ^
α α k k ;
In Sect. 3.3 KF-LSMA, the measurement equation is described by an LMM, and the
state equation is included to perform abundance vector estimation for spectral
unmixing. Therefore, KF-LSMA actually operates on a hyperspectral image cube
and performs mixed pixel classification by estimating abundance vectors via its
state equation, while each image pixel vector in the image cube is used as an input
to the measurement equation. In contrast, KF-SCSP techniques are rather different
from KF-LSMA in the sense that a KF-SCSP technique is a one-dimensional
(1D) signal processing technique to exploit the spectral variability of a single
signature vector across its spectral coverage to perform spectral analysis without
referencing other data sample vectors. As a result, the sample correlation used in the
state equation by KF-LSMA is not available in KF-SCSP techniques.
This and the following sections present several new applications of Kalman
filtering to spectral estimation, identification, and quantification for hyperspectral
62 3 Discrete-Time Kalman Filtering for Hyperspectral Processing
vector that enables capturing the matching signature vector in identifying the
unknown signature vector. According to their functionality, both KF-SSE and
KF-SSI were developed as signature vector estimators, but not abundance vector
estimators, as KF-LSMA was originally designed. This is because no LMM is used
to unmix abundance fractions of image endmembers. Unlike KF-SSE or KF-SSI,
KF-SSQ can be considered a follow-up signature filter. It models its state equation
as a zero-holder interpolator by taking a KF-SSE-estimated signature vector, that is,
a spectral estimate in its measurement equation as a system gain vector, to achieve
spectral quantification of the estimated signature vector. As a result, the quantifi-
cation of the spectral signature value at the current lth band is used as a prediction of
the spectral value at the next adjacent band, the (l + 1)st band. Its ability in spectral
quantification makes KF-SSQ particularly useful in applications of chemical/bio-
logical (CB) defense, where the lethal level of a CB agent is determined by its
concentration (Wang et al. 2004b) and the collected samples do not have to be
spectrally correlated the way image pixel vectors do in an image cube. Therefore,
the quantification of a CB agent is crucial in damage control assessment.
where the index l denotes the lth band, c ¼ ðc1 ; c2 ; . . . ; cL ÞT is a system gain vector,
ϕðl þ 1, lÞ is a transition parameter from the lth band to the (l + 1)st band, and
u ¼ ðu1 ; u2 ; . . . ; uL ÞT and v ¼ ðv1 ; v2 ; . . . ; vL ÞT are white noise vectors, all of which
must be determined a priori. According to Chang and Brumbley (1999a, b), to
implement (3.47) and (3.48) recursively, the initial condition to start the recursive
algorithm between the measurement and state equations is set to ^t 1 ¼ 0 by
assuming that the estimate of the true spectral signature at the first band is
0. The KF is then implemented recursively until it reaches the last band [see
recursive formulas (6–10) in Chang and Brumbley (1999a)]. Comparing (3.47)–
(3.48) to (3.43)–(3.44), there are several salient differences between KF-LSMA
and KF-SSE. First, KF-LSMA uses (3.43) and (3.44) to model the same image
vector. That is, (3.43) represents the status of the image pixel r, which is linearly
mixed by a set of p known image endmembers, m1, m2, . . . , mp and (3.44)
64 3 Discrete-Time Kalman Filtering for Hyperspectral Processing
T
specifies its corresponding abundance vector α ¼ α1 ; α2 ; . . . ; αp , which is
unknown but needs to be estimated via (3.43). Therefore, KF-LSMA requires
the prior knowledge of image endmembers to be used in the linear mixture model
(3.43). As a rather different approach, KF-SSE is used to estimate a target
signature vector t rather than an abundance vector α. It treats a true signature
vector t ¼ ðt1 ; t2 ; . . . ; tL ÞT as a state vector in (3.48) and estimates the state vector
band by band via an observable vector r ¼ ðr 1 ; r 2 ; . . . ; r L ÞT specified by the
measurement equation (3.47). As a result, the variables in both (3.47) and (3.48)
are scalars and represent spectral signature values of the lth band of the target
signature vector t within the observed signature vector r. In this case, the mea-
surement equation (3.47) is simply a true signature vector corrupted by noise u
¼ ðu1 ; u2 ; . . . ; uL ÞT , not a linear mixture model specified by (3.43). Another major
difference is that both (3.47) and (3.48) are linear functions of the same lth band of
the true spectral signature t, tl. This is different from (3.43) to (3.44), which are
linear functions of two distinct vectors, image vector r(k) and abundance vector
α(k) with different dimensions L and p, respectively. Therefore, compared to
KF-LSMA, whose input and output are an image pixel vector r(k) and an estimate
of the abundance vector α(k), α ^ KFLU ðkÞ , respectively, the input of KF-SSE is
simply the lth band spectral value of the observable spectral signature vector r, rl,
and its output is an estimate of tl that is the lth band spectral value of the true
signature vector t, denoted by ^t l
KFSSE
. Furthermore, the state in (3.48) is designed
to predict the spectral value of the (l + 1)st band of the true signature vector t by
updating the currently estimated spectral value of its lth band according to the
spectral variation between two adjacent bands. This is quite different from the
state in (3.44), which is included to keep track of changes in abundance vectors
between two adjacent image pixel vectors. In addition, (3.48) models the rela-
tionship of the spectral signature values between one wavelength and the next
adjacent wavelength as a Gaussian–Markov process specified by the state transi-
tion parameter ϕðl þ 1, lÞ and the additive Gaussian noise, v ¼ ðv1 ; v2 ; ; vL ÞT . If
the variance σ 2v or standard deviation of v is set too low, then the state equation
may not be effective enough to capture real variations in spectral values of one
material. Accordingly, the standard deviation of the state noise is always assumed
to be high.
In the previous section, KF-SSE was used to estimate a true signature vector
t through an observed signature vector r. By remodeling its measurement and
state equations, KF-SSE can be reinterpretated as a spectral signature identifier,
which is referred to as a KF-based spectral signature identifier (KF-SSI. To perform
spectral signature idenitification, it always assumes that there is a database or
3.4 Kalman Filter-Based Hyperspectral Signal Processing 65
K
spectral signature library available for this purpose, denoted by Δ ¼ fsk gk¼1 , which
is made up of K spectral signature vectors, s1, s2 , . . ., sk.
Suppose that the observable spectral signature vector is r ¼ ðr 1 ; r 2 ; . . . ; r L ÞT , to
K
be identified via Δ ¼ fsk gk¼1 . The idea is to use the measurement equation to
describe the observable spectral signature vector r as a noise-corrupted matching
spectral signature vector sk selected from the database or spectral signature library
Δ. An auxiliary spectral signature vector t is then introduced in the state equation to
model the state that describes the ability to identify a given matching signature
vector s in the observable signature vector r. In other words, the identification of the
observable spectral signature vector r is performed by finding a particular spectral
signature vector sk from the Δ that best matches r. In this case, the measurement
equation (3.47) and the state equation (3.48) can be remodeled respectively as
follows:
and
where the matching spectral signature vector sk ¼ ðsk1 ; sk2 ; ; skL ÞT is included in
the measurement equation (3.51) and therefore is the signature vector to be used to
match the observable signature vector r ¼ ðr 1 ; r 2 ; ; r L ÞT . The auxiliary spectral
signature vector t in (3.51) then serves to bridge the observable signature vector
r and the matching signature vector sk to dictate the ability of a given signature
vector sk to match the observable signature vector r. As a result of introducing the
target signature vector t in (3.51), the system gain parameter cl in (3.49) is
reexpressed in (3.51) and becomes a spectral-varying parameter specified by skl/tl.
Interestingly, the use of such a spectral-varying system gain parameter by cl ¼ skl =tl
takes care of the effects resulting from a matching signature vector sk in (3.51). This is
a key difference between KF-SSE and KF-SSI. Therefore, a KF using the pair (3.51)
and (3.50) to perform spectral signature identification is called a KF-SSI. In this case,
the input is provided by the observable spectral signature vector r ¼ ðr 1 ; r 2 ; . . . ; r L ÞT
K
to be identified by a particular signature vector in a database, Δ ¼ fsk gk¼1 , via a
KF-SSI. And the output of the KF-SSI is the estimate of the auxiliary signature
T
vector t, ^t
KFSSI
ðsÞ ¼ ^t 1 ðsÞ,^t 2 ðsÞ, . . . ,^t l
KFSSI KFSSI KFSSI
ðsÞ . To implement (3.50)
66 3 Discrete-Time Kalman Filtering for Hyperspectral Processing
and (3.51) recursively, the initial condition used to start the recursive algorithm
between the measurement and state equations is set to ^t 1
KFSSI
ðsÞ ¼ 0 by assuming
that the estimate of the auxiliary signature vector t at the first band is 0. Then the
KF is implemented recursively until it reaches the last band. Since a KF is a
MSE-based estimation technique, the least-squares error (LSE) between ^t l
KFSSI
T
ðsÞ and the observable spectral signature r ¼ ðr 1 ; r 2 ; . . . ; r L Þ is used instead of
MSE as a quantitative measure defined by
rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
2ffi
XL
LSEðr; sk Þ ¼ r^t ðsk Þ¼
KFSSI KFSSI
l¼1
r l ^t l ðsk Þ : ð3:52Þ
So far, KF-SSE and KF-SSI were developed in the two previous subsections in
attempts to perform signature vector estimation and identification from an observ-
able signature vector r. This subsection presents another new application of
KF-SSE in spectral quantification to quantify a signature vector r, which is referred
to as a Kalman filter-based spectral signature quantifier (KF-SSQ).
One of the great challenges in hyperspectral data exploitation is spectral quan-
tification. This is particularly important for CB defense, where the concentrations of
targets are of major interest rather than their shapes and sizes as generally encoun-
tered in image processing. The concentration of an agent is a key element in the
assessment of a threat (Kwan et al. 2006). In recent years, many algorithms have
been developed for quantification (Goetz and Boardman 1989; Shimabukuro and
Smith 1991; Settle and Drake 1993; Smith et al. 1994b; Tompkins et al. 1997;
Heinz and Chang 2001; Kwan et al. 2006). However, most of them are image
analysis-based linear spectral unmixing methods that estimate the abundance frac-
tions of image endmembers assumed to be present in a linear mixture model such as
the fully constrained least-squares (FCLS) method (Heinz and Chang 2001). The
technique developed in this subsection is signature vector-based and allows one to
quantify a particular material substance present in a single signature vector of
interest band by band without resorting to an LMM. Once again, a direct application
of KF-SSE in spectral quantification is not applicable since the state equation
described by (3.48) is the spectral value tl but not its abundance fraction αl,
which is of interest in quantification. To estimate the abundance fraction αl in the
state equation, the state tl in (3.48) is replaced by its abundance fraction αl and the
system gain cl in (3.48) is replaced by the lth band of the target signature vector t, tl.
68 3 Discrete-Time Kalman Filtering for Hyperspectral Processing
The state transition parameter ϕðl þ 1, lÞ is set to 1 such that the state equation is
actually a zero-holder interpolator. The resulting measurement equation and state
equation become
By virtue of the pair (3.53) and (3.54), a KF-SSQ can be designed and implemented
to quantify the target signature t ¼ ðt1 ; t2 ; . . . ; tL ÞT present in the data sample vector
r in tems of its abundance fraction, αl. In this case, the inputs of the KF-SSQ are an
observable signature vector r ¼ ðr 1 ; r 2 ; . . . ; r L ÞT and target signature vector
t ¼ ðt1 ; t2 ; . . . ; tL ÞT . Its output is the estimate of αl associated with target signature
vector t, denoted by α ^ lKFSSQ ðtÞ.
Two scenarios can be implemented for the KF-SSQ. One is that the target
signature vector t in (3.53) is assumed to be known a priori. However, in many
applications, we may not have such prior knowledge. Under this circumstance,
another scenario is that t can be estimated from the observable signature vector r
¼ ðr 1 ; r 2 ; . . . ; r L ÞT via KF-SSE, where the tl in (3.53) is then replaced by its KF-SSE
estimation, ^t l
KFSSE
. The resulting measurement equation becomes
r l ¼ αl^t l
KFSSE
þ ul : ð3:55Þ
The pair (3.54) and (3.55) specifies the implementation of a KF-SSQ for unknown
target signature vector t, where the input is KF-SSE-estimated ^t l
KFSSE
specified by
(3.55) and the output is the estimate of αl in (3.54) corresponding to ^t l
KFSSE
, denoted
KFSSQ ^ KFSSE
^l
by α t .
The KF-SSQ described by (3.54) and (3.55) is quite different from KF-LSMA,
which assumes all target signature vectors, that is, image endmembers, are known a
priori and then quantifies all the known signatures present in the data vector r via
the FCLS method developed by Heinz and Chang (2001) using an LMM. However,
the second scenario of the KF-SSQ specified by (3.54) and (3.55) can be
implemented without prior knowledge of the target signature vector t, a task that
cannot be accomplished by KF-LSMA.
In what follows, the algorithmic steps of each of the three KF-SCSP techniques,
KF-SSE, KF-SSI, and KF-SSQ, are described. Assume that r ¼ ðr 1 ; r 2 ; . . . ; r L ÞT
and t ¼ ðt1 ; t2 ; . . . ; tL ÞT are the observable signature vector and the target signature
vector to be estimated, respectively.
KF-SSE
1. Initial conditions:
(a) Preset the values of σ u and σ v.
3.4 Kalman Filter-Based Hyperspectral Signal Processing 69
2. Start rl with l ¼ 1:
(a) Calculate the Kalman gain:
1
K l ¼ ϕ P cl cl P cl þ σ u :
lþ1 l l l1 l l1
P ¼ ½1 K l cl P :
l l l l1
^t ¼ϕ ^t :
lþ1l lþ1l ll
3. Increase l by 1 and process rl with l ¼ 2, repeat the five recursive steps outlined
in step 2 until l ¼ L.
KFSSE T
4. Output KF-SSE-estimated ^t
KFSSE
¼ ^t 1 ; ^t 2 ; . . . ; ^t L , with ^t l
KFSSE KFSSE KFSSE
KF-SSI
1. Initial conditions:
K
(a) Assume that Δ ¼ fsk gk¼1 is a given spectral library or database and sk is a
matching signature selected from Δ.
(b) Preset the values of σ u and σ v.
(c) Set ^t 1j0 ðsk Þ ¼ 0 and P ¼ Var½t1 ¼ 1.
1 0
(d) Set cl ¼ skl =^t l for all l ¼ 1, 2, . . . , L.
(e) Set ϕ ¼ 1 for all l ¼ 1, 2, . . . , L.
lþ1 l
70 3 Discrete-Time Kalman Filtering for Hyperspectral Processing
2. Start rl with l ¼ 1.
(a) Calculate the Kalman gain:
1
K l ¼ ϕ P cl cl P cl þ σ u :
lþ1 l l l1 l l1
P ¼ ½1 K l cl P :
l l l l1
^t ¼ϕ ^t :
lþ1l lþ1l ll
3. Increase l by 1 and process rl with l ¼ 2. Update cl ¼ skl =^t l and repeat the five
recursive steps outlined in step 2 until l ¼ L.
T
4. Output KF-SSI-estimated ^t
KFSSI
ðsk Þ ¼ ^t 1 ðsk Þ,^t 2 ðsk Þ, . . . ,^t L
KFSSI KFSSI KFSSI
ðsk Þ ,
with ^t l ðsk Þ ¼ ^t l , for all 1 l L obtained from steps 2 and 3.
KFSSE
KF-SSQ
1. Initial conditions:
(a) Preset the values of σ u.
(b) Set σ v ¼ 0.
(c) Assume that α(t) is the abundance vector corresponding to the target signa-
ture t. Set cl ¼ tl for l ¼ 1, 2, . . . , L.
^ 1j0 ðtÞ ¼ 0 and P ¼ Var½α1 ¼ 1.
(d) Let α
1 0
(e) Set ϕ ¼ 1 for all l ¼ 1, 2, . . . L.
lþ1 l
3.5 Conclusions 71
2. Start rl with l ¼ 1.
(a) Calculate the Kalman gain:
1
K l ¼ ϕ P cl cl P cl þ σ u :
lþ1 l l l1 l l1
P ¼ ½1 K l cl P :
l l l l1
^
α ¼ϕ α
^ :
lþ1l lþ1l l l
3.5 Conclusions
This chapter reviews Kalman filtering, which is one of the most widely used real-
time processing techniques in statistical signal processing. Its application to
hyperspectral imaging was first explored by Chang and Brumbley (1999a, b) to
develop KF-based LSMA (KF-LSMA). The concept of Kalman filtering was
further explored by Chang and Wang (2006) and Chang (2013), which developed
KF-based techniques for spectral signature characterization rather than KF-LSMA
for LSU. Accordingly, several new applications of Kalman filtering in
hyperspectral signature analysis were developed for spectral estimation, identifica-
tion, and quantification by rederiving the measurement and state equations, which
resulted in various techniques referred to as KF-based spectral characterization
signal processing (KF-SCSP) techniques.
Chapter 4
Target-Specified Virtual Dimensionality
for Hyperspectral Imagery
Abstract Virtual dimensionality (VD) was first envisioned and coined by Chang
(Hyperspectral Imaging: Techniques for Spectral Detection and Classification,
Kluwer Academic/Plenum Publishers, New York, 2003) and later written about
with details in Chang and Du (IEEE Transactions on Geoscience and Remote
Sensing 42:608–619, 2004). It was originally developed for the purpose of finding
an appropriate number of signatures required by linear spectral mixture analysis to
form an effective linear mixing model for data unmixing. To go beyond its original
intent VD was further defined as the number of spectrally distinct signatures present
in hyperspectral data where the intention is to leave room for users to specify what
spectrally distinct signatures are. Unfortunately, since its inception in 2003, many
techniques developed to estimate VD have overlooked this issue and, thus, have
produced various values for VD. As a result, different conclusions have been drawn
about VD. This issue was realized by Chang (Hyperspectral data processing:
algorithm design and analysis, Wiley, Hoboken, 2013), where VD was defined by
two types of criteria, data characterization-driven criteria and data representation-
driven criteria. This chapter revisits VD and follows a different route. It extends VD
to target-specified VD from various target signal source perspectives of interest
instead of focusing on data as treated in Chang (Hyperspectral data processing:
algorithm design and analysis, Wiley, Hoboken, 2013). In particular, it generalizes
VD in two different approaches. One is to develop a unified theory for VD of
hyperspectral imagery, where the value of VD can be completely determined by
targets of interest. With this newly developed theory VD techniques can be cate-
gorized according to targets that can be characterized by various orders of statistics
into first-order statistics, second-order statistics, and high-order statistics, where a
binary composite hypothesis testing problem is formulated with a target of interest
considered as a desired signal under the alternative hypothesis, while the null
hypothesis represents an absence of the target. With this interpretation many VD
estimation techniques can be unified under the same framework, including those
defined in Chang (Hyperspectral data processing: algorithm design and analysis,
Wiley, Hoboken, 2013), for example, the Harsanyi–Farrand–Chang method or the
maximum orthogonal complement algorithm (MOCA) (Kuybeda et al., IEEE
Transactions on Signal Processing 55:5579–5592, 2007). A second approach is to
define VD as being completely determined by real target signal sources instead of
data spectral statistics. Once again, a binary composite hypothesis testing problem
is also used to determine whether a specific target signal source is a desired signal
source to account for VD. Interestingly, this approach is more practical in real-
world problems because VD is actually determined by real target signal sources
directly extracted from the data for a particular application rather than eigenvectors
specified by data spectral statistics, which are independent of applications.
4.1 Introduction
Linear spectral mixture analysis (LSMA) is one of most important and fundamental
applications in hyperspectral data exploitation. When implemented, it requires prior
knowledge of the signatures used to form a linear mixing model (LMM) to find the
abundance fraction of each of the signatures present in a data sample vector.
Unfortunately, in reality, such prior knowledge is generally not available. Under
such a circumstance two challenging issues must be addressed. One is determining
the number of signatures, p. The other is finding these p signatures that are used to
form an LMM. In the past, the value of p was determined empirically, and no
specific criteria were proposed until virtual dimensionality (VD) was introduced in
2003 (Chang 2003a, b). Once the value of p is determined, a follow-up issue is to
find the desired p signatures. To resolve this issue, various techniques have been
developed, for example, endmember extraction algorithms (EEAs) discussed in
great detail in Part II of Chang (2013) as well as endmember finding algorithms
(EFAs) studied extensively in Parts II and III of Chang (2016) or unsupervised
target detection algorithms such as the automatic target generation process (ATGP),
unsupervised nonnegativity constrained least-squares (UNCLS) method, and
unsupervised fully constrained least-squares (UFCLS) method in Chang (2016).
Interestingly, it was shown in Gao et al. (2015a, b) and Chang (2016) that signatures
extracted by EEAs or found by EFAs are not necessarily desired signatures that are
suitable for the LMM used by LSMA. In other words, EEAs and EFAs are not the
best algorithms for finding signatures to form an LMM for LSMA, as many
researchers once expected. This leads to a crucial clarification of how to properly
use VD for interpretation. Specifically, the value of VD, nVD, determined for
endmembers should not be the same value of nVD determined for signatures used
by LSMA. The same principle holds for the number of anomalies. This implies that
the value of nVD must be determined by applications, not solely by data statistics.
This chapter investigates this important issue, which so far has not been recognized
or explored.
Many techniques have been developed to estimate the number of signal sources in
hyperspectral imagery. The earliest one was developed by Harsanyi et al. (1994a, b),
referred to as the Harsanyi–Farrend–Chang (HFC) method, which has been shown
to be by far the most effective and promising. However, the concept of VD was not
realized until Chang (2003a), which modeled the VD estimation problem as a
binary composite hypothesis testing problem with covariance eigenvalues and
4.2 Review of VD 75
correlation eigenvalues being treated as signal sources under the null and alterna-
tive hypotheses, respectively. It then developed a Neyman–Pearson detector (NPD)
to determine the value of nVD as specified by a prescribed false alarm probability.
The idea of using a binary hypothesis testing problem was also later adopted by the
maximal orthogonal complement algorithm (MOCA) (Kuybeda et al. 2007), which
used a maximum likelihood detector (Bayesian detector) instead of adopted by
NPD to determine the rank of rare signals in the data to estimate nVD. An idea of
combining the HFC method with MOCA was investigated in Chang et al. (2011b),
where an approach called maximal orthogonal subspace projection (MOSP) was
developed to estimate nVD. Most recently, MOSP was further extended to high-
order statistics (HOS) HFC methods (Chang et al. 2014a, b).
Another approach is to use a linear model for data representation to estimate
nVD. One is signal subspace estimation (SSE) and hyperspectral signal identifica-
tion by minimum error (HySime) proposed by Bioucas-Dias and Nascimento
(2005, 2008). However, there are two major issues with the SSE/HySime method.
The first has to do with the fact that the value of nVD estimated by SSE/HySime is
heavily reliant upon the method used to estimate the noise covariance matrix. The
second and most serious issue is that the HySime-estimated value is a single fixed
value and completely determined by the data statistics and linear representation
regardless of application. In other words, the SSE/HySime-estimated nVD is the
same for all applications, such as anomaly detection, endmember finding/extrac-
tion, and LSMA. Apparently, this is generally the case. Thus, from a practical point
of view, it is not an appropriate measure for estimating nVD. A second one is
LSMA-based methods proposed by Chang et al. (2010a, b), where the value of
nVD is determined by the minimal unmixed error produced by the number of
signatures used to form an LMM for LSMA.
The aforementioned techniques have one thing in common: their estimated
values for nVD have nothing to do with applications. Although the value estimated
by the HFC method with its variants can be made to vary by adjusting different false
alarm probabilities, which may be somewhat specified by applications, the signal
sources used by HFC-based methods via a binary composite hypothesis testing
formulation are eigenvalues or eigenvectors, not real target signal sources. Accord-
ingly, they are not as effective as the signal sources found directly from applica-
tions. To resolve this dilemma, this chapter develops a target-specified VD (TSVD)
theory for hyperspectral imagery where the value of nVD is completely determined
by the targets of interest characterized by specific applications with algorithms
custom-designed for finding them.
4.2 Review of VD
Determining the value of nVD can be very challenging. One of the earliest attempts
at doing so involved using eigenanalysis, such as eigenvalues or singular values, as
a criterion to find a significant gap that would indicate how many eigenvalues/
76 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
since various real target signal sources may be projected onto the same eigenvec-
tors. Accordingly, the spectral properties of real target signal sources are very likely
compromised. Thus, a crucial and critical issue in designing and developing
effective techniques for VD estimation is the signal sources that must be determined
by real targets of interest, not eigenvalues/eigenvectors. To the author’s best
knowledge, the ATGP-NPD developed in Chang et al. (2011a, b, c, d) and a theory
of HOS-based VD in Chang et al. (2014a, b) are the only approaches that make use
of real target signal sources to estimate VD. This chapter is believed to be the first to
extend the idea of ATGP-NPD and HOS-VD to derive a unified theory of TSVD.
Table 4.1 summarizes various VD estimation techniques available in the liter-
ature, along with their criteria, assumptions, and constraints.
4.3 Eigen-Analysis-Based VD
HySime Second MSE: LMM Yes Yes None Yes Minimization Fixed
order
OSP Second LSE: Yes No ε No Minimization Vary/ε
order OSP-LMM
MOSP Second OSP No No PF No NPD: ATGP-generated targets Vary/PF
order
SDD Non- Detector power No No PF No NPD: nonparametric PDs Vary/PF
parametric
SPICE Second LSE + V + SPT No No ε No LSE Vary/ε
order
GENE Second LSE Yes Yes Nmax, PF Yes NPD: LMM-fitting error Vary/PF
order
Second-order Second LSE Yes No PF No NPD: second-order-statistics-gener- Vary/PF
statistics order ated targets
LSE model Second LSE Yes Yes PF No NPD: LS-generated targets Vary/PF
order
NCM Prior Likelihood NCM Yes σ2 No Bayesian algorithm for ID Vary/σ 2
function
HOS-VD HOS OSP No No PF No NPD: HOS-generated targets Vary/PF
k-NN Manifold NN No No k (NN) Yes Average ID/pixel Vary/k
79
80 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
NPD, developed in the HFC method, which assumes no prior knowledge and signal
detection performance is evaluated by the false alarm probability, PF, at a signif-
icant level. As a result, the number of the qth-order spectral targets varies and is
determined by a prescribed value of PF. The TSVD for hyperspectral imagery is
derived by integrating the concept of specifying spectral targets of any given order
into a binary composite hypothesis testing detection formulation to derive a detec-
tor that can determine the value of nVD for a specific type of spectral target, such as
anomaly, endmember, and others. Under this umbrella many existing VD estima-
tion techniques can be treated as the first two orders of statistics-based techniques,
for example, HFC/noise-whitened HFC (NWHFC) as a first-order-statistics-based
technique and MOCA, SSE/HySime as 2OS-based techniques.
As noted in the introduction, VD was originally designed to resolve the issue of
how many signatures are required to form an LMM used by LSMA for data
unmixing and was used subsequently to address similar issues in various applica-
tions, such as the number of endmembers, nE, to be extracted, the number of
dimensions, nDR, required to be retained after data dimensionality reduction, the
number of bands, nBS, to be selected by BS, and the number of anomalies, nA,
present in the data. Unfortunately, when VD was defined in Chang (2003a), it did
not specify what a spectrally distinct signature should be. Accordingly, a single
value of VD, nVD, cannot be adequately used for all applications. Apparently, the
previously developed eigenanalysis-based approaches are not able to cope with this
issue. This chapter reinvents a wheel to revolutionize the concept of VD and
develops a rather different theory for VD from those discussed in Chang (2003a,
Chap. 17) and Chang (2013, Chap. 5). First, finding nDR, nBS, nE, and nA requires
different values. That is, the value of nVD estimated for nDR, nBS, nE, and nA should
be different in each case and cannot be the same value determined by a single
constant value. For example, nDR and nBS are generally determined by data struc-
tures and statistics compared to nE and nA, which are indeed determined by targets
of interest. By far most techniques available in the literature belong to the first
category and have no reference to any specific targets. Therefore, they are better for
estimating nDR and nBS, not nE and nA, since there is no prior knowledge but only
data structures and statistics available for use in DR and BS. Examples of such
techniques include the HFC method and HySime. On the other hand, for nE and nA
to be estimated effectively, the techniques used to estimate the value of nVD must be
those that can capture the spectral characteristics of specific targets of interest, not
data statistics or structure. This means that techniques used to estimate nDR and nBS
may not be suitable for estimating nE and nA. Therefore, new approaches must be
sought. This chapter presents a theory of TSVD for hyperspectral imagery where
the value of nVD is completely determined by targets of interest characterized by
their spectral statistics.
4.3 Eigen-Analysis-Based VD 81
Binary hypothesis testing has been widely used for signal detection in noise where
two hypotheses represent signal presence and absence. This approach was first
applied to determine the value of nVD by the HFC method (Harsanyi et al. 1994a, b).
The idea of the HFC method is very simple. It assumes that the hyperspectral
signatures are unknown nonrandom and deterministic signal sources and noise is
white with zero mean. Under this circumstance, the signatures of interest will be
N
contributed to the first-order statistics, which is the sample mean. Let fri gi¼1 be data
XN
sample vectors and μ the sample mean vector given by μ ¼ r . The HFC
i¼1 i
XN
method first calculates the sample autocorrelation matrix, RLL ¼ r r T , and
i¼1 i i
XN
sample covariance matrix, KLL ¼ i¼1
ðri μÞðri μÞT , then finds the differ-
ence between their corresponding eigenvalues. More specifically, let
^λ 1 ^λ 2 ^λ L and fλ1 λ2 λL g be two sets of eigenvalues
generated by RLL and KLL , called correlation eigenvalues and covariance
eigenvalues, respectively, where L is the number of spectral channels. If a
hyperspectral signal source is present in the data, there should be some spectral
dimension l with 1 l L such that ^λ l > λl owing to the fact that the signal source
will contribute to the sample mean in the sample correlation matrix RLL but not to
the sample covariance matrix KLL . It should be noted that the sample correlation
N N
matrix RLL is invariant to any permutation fPðiÞgi¼1 of figi¼1 , that is,
XN N N
RLL ¼ r r , where fri gi¼1 ¼ rPðiÞ i¼1 . This implies that RLL is an
T
i¼1 PðiÞ PðiÞ
intrasample correlation matrix, not an intersample spatial correlation matrix. In this
case, we can formulate the problem of determining the value of nVD as follows.
To determine the value of nVD, Harsanyi et al. (1994a, b) formulated the problem
of determining the value of nVD as a binary hypothesis problem as follows:
H 0 : zl ¼ ^λ l λl ¼ 0
versus for l ¼ 1, 2, . . . , L; ð4:1Þ
H 1 : zl ¼ ^λ l λl > 0
where the null hypothesis H0 and the alternative hypothesis H1 represent the case
that the correlation eigenvalue is equal to its corresponding covariance eigenvalue
and the case that the correlation eigenvalue is greater than its corresponding
covariance eigenvalue, respectively. In other words, when H1 is true (i.e., H0
fails), it implies that there is a hyperspectral signature contributing to the correlation
eigenvalue in terms of first-order statistics since the noise energy represented by the
82 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
where the likelihood ratio test Λ(zl) is given by Λðzl Þ ¼ p1 ðzl Þ=p0 ðzl Þ, with p0(zl) and
p1(zl) given by (4.1). Thus, a case of ^λ l λl > τl indicates that δNP(zl) in (4.2) fails
the test, in which case signal energy is assumed to contribute to the eigenvalue, ^λ l ,
in the lth spectral dimension. It should be noted that the test for (4.1) must be
performed for each of the L spectral dimensions. Therefore, for each pair of ^λ l λl ,
the threshold τl varies and should be spectral dimension-dependent. Using (4.2), the
value of nVD can be determined by calculating
XL
HFC ðPF Þ ¼
VDNP l¼1
δNP ðzl Þ ; ð4:3Þ
The MOCA, a recent approach developed by Kuybeda et al. (2007), has attracted
considerable interest and been further studied by Chang et al. (2011a, b, c, d, 2014a, b).
It was designed to find a signal subspace and its rank dimensionality. If we interpret
the spectrally distinct signatures defined by VD as rare signal sources, MOCA can
also be considered a VD estimation technique for HOS-characterized target signal
sources (Chang et al. 2014a, b).
N
Let fri gi¼1 be data sample vectors. Assume that for each l, which is supposed to
N
be the rank of a signal subspace Sl, the data sample vectors fri gi¼1 are divided into
4.3 Eigen-Analysis-Based VD 83
two index classes, one for the target signal class, IT(l ), the other for the background
class, IB(l ). Now, we define
2
νl ¼ maxi2IB ðlÞ P⊥Sl r i
; ð4:4Þ
⊥ 2
ξl ¼ maxi2IT ðlÞ PSl ri ; ð4:5Þ
ηl ¼ maxfξl ; νl g: ð4:6Þ
where the null hypothesis H0 represents that the maximum residual is from the
background signal sources, while the alternative hypothesis H1 represents that the
maximum residual is from the target signal sources. To make (4.9) work, we need to
find PDs governed by each of the two hypotheses. Because the orthogonal comple-
ment subspace projections of data sample vectors P⊥ Sl ri under H0 are supposed to be
noise sample vectors, it is reasonable for MOCA to assume that the vector P⊥ Sl r i
under H0 behaves as independent identically Gaussian random variables. Moreover,
ηl represents the maximum residuals of orthogonal projection obtained in < Sl >⊥
under H0. By virtue of extreme value theory (Leadbetter 1987), ηl can be modeled
as a Gumbel distribution, that is, Fvl ðηl Þ is the cumulative distribution function (cdf)
of vl given by
( h i)
ð2logN Þ1=2 p ffiffiffiffiffiffiffi
xσ 2 ðLlÞ
ð2logN Þ1=2 þ12ð2logN Þ1=2 ðloglogNþlog4π Þ
Fvl ðxÞ exp e σ 2 2ðLlÞ
: ð4:10Þ
Furthermore, MOCA made another assumption about the ηl under the alternative
hypothesis H1. That is, for each pixel ri, i 2 IT ðlÞ, the maximum of residuals of
orthogonal complement subspace projections can be modeled as a random variable
ξl with the PD pξl ðηl Þ, which is assumed to be uniformly distributed on
84 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
[0, ηl1]. Since there is no prior knowledge available about the distribution of target
pixels, assuming ηl under H1, uniform distribution seems most reasonable
according to the maximum entropy principle in Shannon’s information theory.
Under these two assumptions, we can obtain
pðH0 ; ηl Þ ¼ pνl ðηl ÞFξl ðηl Þ ¼ pνl ðηl Þðηl =ηl1 Þ; ð4:11Þ
pðH 1 ; ηl Þ ¼ Fνl ðηl Þpξl ðηl Þ ¼ Fνl ðηl Þð1=ηl1 Þ: ð4:12Þ
Since pηl ðηl Þ ¼ pðH 0 ; ηl Þ þ pðH 1 ; ηl Þ ¼ ð1=ηl1 Þ ηl pνl ðηl Þ þ Fνl ðηl Þ , we can obtain
an a posteriori PD of p(H0|ηl) given by
η l pν l ð η l Þ
p H 0 η l ¼ ð4:13Þ
η l p ν l ð η l Þ þ Fν l ð η l Þ
Fνl ðηl Þ
p H1 ηl ¼ : ð4:14Þ
ηl pνl ðηl Þ þ Fνl ðηl Þ
By virtue of (4.13) and (4.14), the rank of the desired signal subspace can be
obtained by
l* ¼ arg min1lL p H 0 ηl p H 1 ηl : ð4:15Þ
VDMOCA ¼ l* : ð4:16Þ
n o
of eigenvalues, λ1 λ2 λl* p* corresponding to background signal sources
n o
specified by SVD-generated singular vectors, λl* p* þ1 λl* p* þ2 λl*
corresponding
to rare signal sources
specified by SVD-generated singular vectors,
and λl* þ1 λl* þ2 λL corresponding to noise signal sources specified by
SVD-generated singular vectors. Nevertheless, the main theme in this chapter is to
determine the number of target signal vectors for the value of nVD, not to find the
number of rare signal sources. Thus, only MOCA is of major interest in finding l* rather
than MX-SVD in finding p*. In this case, the MX-SVD is not considered part
of MOCA.
While both the HFC method and MOCA take advantage of binary hypothesis
testing theory, several crucial differences between the HFC method and MOCA
need to be noted.
1. From a detection point of view, the HFC method is developed based on
Neyman–Pearson detection theory, as opposed to MOCA, which was derived
from Bayesian or maximum likelihood detection theory. As a result, the
HFC-estimated nVD is a function of the false alarm probability, PF, compared
to a single fixed value-VD produced by MOCA.
2. For each spectral component the HFC method designs an NPD to determine
whether a spectrally distinct signature is present in the lth spectral component by
testing the difference between the correlation eigenvalue and the corresponding
covariance eigenvalue calculated from this particular spectral component,
^λ l λl . By contrast, MOCA does not use singular values resulting from the
SVD in the same way as the HFC method uses eigenvalues. Instead, it replaces
^λ l λl with the maximum l2 norm of singular vectors to perform binary
hypothesis testing to determine whether a given singular vector is a rare signal
source vector. Therefore, technically speaking, the HFC method can be consid-
ered a first-order statistics method since it detects ^λ l λl resulting from the
sample mean of the lth spectral component. On the other hand, MOCA computes
the signal energy strength of a singular vector measured by the l2 norm. Accord-
ingly, MOCA can be considered a 2OS method because singular vectors are
obtained by second-order statistics. For this reason, MOCA is more accurately
expressed as SVD-MOCA.
3. Interestingly, the HFC method can also be implemented in the same framework
as MOCA, with ^λ l λl being replaced by the lth eigenvector using Neyman–
Pearson detection theory. The resulting HFC method, referred to as the principal
components analysis (PCA)-HFC method, is a 2OS HFC method because the
86 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
eigenvectors are obtained from PCA, which is formed by the 2OS sample
covariance matrix and can be considered a counterpart of MOCA.
The HFC method makes no assumptions about the data set compared to MOCA,
which assumes data sample vectors represented by a linear form from which a
sequence of linear signal subspaces can be determined via orthogonal subspace
projection (OSP). Thus, basically, MOCA can be considered as belonging to the
same category of signal subspace estimation (SSE) (Bioucas-Dias and Nascimento
2005), hyperspectral signal identification by minimum error (HySime) (Bioucas-
Dias and Nascimento 2008), and LSMA-based methods in Chang et al. (2010a, b),
all of which take advantage of various data linear representations to derive VD
estimation techniques via OSP. As a consequence, all of these methods, including
MOCA, can be considered data representation-driven techniques. In contrast, the
HFC method only deals with data statistics and has nothing to do with data
representation. In this case, the HFC method can be considered a data
characterization-driven technique. More details regarding such taxonomy can be
found in Chang (2013).
While using binary hypothesis testing to determine the value of nVD is an elegant
idea, the signals considered under each hypothesis are not real signals. For example,
the HFC method and MOCA respectively use eigenvalues and the vector length of
eigenvectors/singular vectors to determine the value of nVD, but none of them is a
real signal. As noted in Section 4.3 the HFC method and MOCA may underestimate
VD due to the fact that the first very large eigenvalues may dominate the subsequent
eigenvalues and their corresponding eigenvectors may not be able to use one
eigenvector to specify one real target signal. On some occasions different
eigenvectors may produce the same signal sources, thus determine the same value
of nVD. The concept of using target pixels to estimate the value of nVD was further
investigated in Chang et al. (2014a, b) to extend the HFC method to HOS-based
HFC methods. This approach has proven to be very promising and effective at
unifying various VD estimation techniques in the same problem setting and formu-
lation. The TSVD presented in this chapter is developed to address this issue. In
what follows we will characterize targets according to their spectral statistics into
two categories, 2OS and HOS.
anomalies. Targets of this type are generally unknown and very difficult to locate by
visual inspection owing to their limited spatial extent, or sometimes no spatial
information resulting from their presence at the subpixel level. To address this
issue, a new concept of spectral targets was recently introduced in Chang
et al. (2010a, b, 2011a, b, c, d) to differentiate spatial targets, which are generally
identified by their spatial properties such as size, shape, and texture. In traditional
image processing there are no spectral bands involved. Targets of interest are
generally defined and identified by their spatial properties such as size, shape, and
texture. Accordingly, the targets of this type are considered spatial targets. The
techniques developed to recognize such spatial targets are referred to as spatial
domain-based image processing techniques. On the other hand, owing to the use of
spectral bands specified by a range of wavelengths, a multispectral or hyperspectral
data sample is actually a vector expressed as a column vector, a sample of which in
a spectral band is produced by a particular wavelength. As a consequence, a single
hyperspectral sample vector already contains abundant spectral information pro-
vided by hundreds of contiguous spectral bands that can be used for data exploita-
tion. Such spectral information within a single data sample vector is called IBSI,
defined earlier. By virtue of such IBSI, two single data sample vectors can be
discriminated, classified, and identified via a spectral similarity measure, such as
the Spectral Angle Mapper (SAM). A target is called a spectral target if it is
analyzed based on its spectral properties characterized by IBSI(S), as opposed to
a spatial target, which is analyzed by interpixel spatial information provided by the
spatial correlation among sample pixels.
As an illustrative example, let S ¼ fri gi¼1 N
be a set of N data sample vectors,
T
where ri ¼ ðr i1 ; r i2 ; . . . ; r iL Þ is the ith data sample vector in S and L is the total
number of spectral bands. The spectral correlation across all the spectral bands
within ri is defined and referred to as the IBSI of signature ri, denoted by IBSI(ri),
with S ¼ fri g. That is, IBSI(ri) is provided by the spectral correlation among the
L
L spectral values, r ij j¼1 , across spectral bands within the single data sample
XL
vector ri, such as 2OS provided by IBSI(ri), the autocorrelation of ri, r 2 , and
j¼1 ij
XL
cross correlation of ri, r r . However, what we are really interested in
j¼1, k¼1, j6¼k ij ik
is the sample statistics provided IBSI(ri) among a set of data sample vectors,
N
S ¼ fri gi¼1 , denoted by IBSI(S), specifically, 2OS of IBSI(S) such as a sample
XN
autocorrelation matrix of S, r r T , and a sample cross-correlation matrix of S,
i¼1 i i
XN
r r T . It should be noted that IBSI(S) is independent of the intersample
i¼1, j¼1, i6¼j i i
spatial correlation because IBSI(S) remains the same after the samples in S used to
calculate IBSI(S) are reshuffled. This type of spectral information is the opposite of
the sample spatial statistics commonly used in traditional image processing, which
takes into account the spatial correlation among a set of data sample vectors where
reshuffling data sample vectors in a sample spatial correlation matrix can result in a
88 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
different sample spatial correlation matrix because it alters the spatial correlation
when the data sample vectors are rearranged in different spatial coordinates.
Two types of spectral target are generally of interest based on sample spectral
statistics, one characterized by second-order sample IBSI and the other by sample
IBSI of order higher than two, referred to as high-order IBSI. It should be noted that
the term IBSI(S) is defined as a sample spectral correlation resulting from a set of
data sample vectors specified by S. It is actually the size of S closely related to how
to define second-order IBSI and high-order IBSI. In hyperspectral image analysis,
the spectral targets of interest in hyperspectral data exploitation are those that either
occur with low probability or have small populations when they are present. Such
spectral targets generally appear in small populations and occur with low probabil-
ities, for example, special spices in agriculture and ecology, toxic wastes in
environmental monitoring, rare minerals in geology, drug/smuggler trafficking in
law enforcement, combat vehicles on the battlefield, landmines in war zones,
chemical/biological agents in bioterrorism, weapon concealment, and mass graves.
As a result, the sample size of data sample vectors specified by such a spectral
target, S, is relatively small and can generally be considered as being composed of
insignificant objects because of their very limited spatial information, but they are
actually critical and crucial for defense and intelligence analysis owing to the fact
that they are generally hard to identify by visual inspection. From a statistical point
of view, they are insignificant compared to targets with large sample pools and the
spectral information statistics of such special targets can be captured not by second-
order IBSI but rather by high-order IBSI.
Since most VD estimation techniques reported in the literature are developed based
on eigendecomposition, one common drawback of these methods is the value of
nVD, which is generally determined by eigenvalues or eigenvectors/singular vec-
tors, none of which represents real target signal sources or energies. In this case,
even though the number of spectrally distinct signatures can be estimated by VD,
we still do not know what these spectrally distinct signatures are. In this section we
present two techniques that generate 2OS target signal sources to be used to
estimate the value of nVD: (1) the LSMA-based techniques developed in Chang
et al. (2010a, b) and Paylor (2014), where particular algorithms are involved in
finding real signal sources used to unmix data in the LSE sense, and (2) the ATGP/
MOCA developed in Chang et al. (2011a, b, c, d), where the ATGP developed in
Ren and Chang (2003) was used to generate real target sample vectors to replace the
singular vectors used in MOCA.
4.4 Finding Targets of Interest 89
Assume that m1, m2, . . ., mp are spectral signatures used to unmix the data sample
vectors. Let L be the number of spectral bands and r be an L-dimensional data
sample vector which can be modeled as a linear combination of m1, m2, . . ., mp
with appropriate abundance fractions specified by α1, α2, . . . , αp. More precisely,
r is an L 1 column vector and M be an L p target spectral signature matrix,
denoted by [m1 m2 . . . mp], where mj is an L 1 column vector represented by the
spectral signature of the j-th target resident in the pixel vector r. Let
T
α ¼ α1 ; α2 ; . . . ; αp be a p 1 abundance column vector associated with
r where αj denotes the fraction of the j-th target signature mj present in the pixel
vector r. A classical approach to solving a mixed pixel classification problem is
linear unmixing which assumes that the spectral signature of the pixel vector r is
linearly mixed by m1, m2, . . ., mp, referred to as an LMM as follows:
r ¼ Mα þ n; ð4:17Þ
r ¼ dαp þ Uγ þ n ð4:18Þ
where d ¼ mp is the desired spectral signature of mp and U ¼ m1 m2 . . . mp1 is
the undesired target spectral signature matrix made up of m1 , m2 , . . . , mp1 which
are the spectral signatures of the remaining p 1 undesired targets. Using (4.18) we
can design an OSP to annihilate U from the pixel vector r prior to detection of tp.
One of such desired orthogonal subspace projectors was OSP derived by Harsanyi
and Chang (1994) and given by
P⊥
U ¼ I UU ;
#
ð4:19Þ
1 ⊥
where U# ¼ UT U UT is the pseudo-inverse of U. The notation ⊥
U in PU indicates
⊥
that the projector PU maps the observed pixel vector r into the orthogonal
complement of < U >, denoted by < U>⊥ . Applying P⊥
U to (4.19) yields
90 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
T ⊥ T ⊥
⊥ y =a p w PU d + w PU n
x = PU r T
y =w x
Fig. 4.1 Diagram of linear detection system δLDS(r) designed for (4.20)
P⊥ ⊥ ⊥ ⊥ ⊥ ⊥
U r ¼ PU dαp þ PU Uγ þ PU n ¼ αp PU d þ PU n: ð4:20Þ
2 ⊥ T
where EðÞ is expectation, P⊥U ¼ P⊥U and PU ¼ P⊥U.
There are two approaches to maximizing (4.21) over w. One is from a pattern
classification point of view by formulating (4.21) as a generalized eigenvalue
problem as follows (Duda and Hart 1973):
!
α2p wT P⊥ T ⊥
U dd PU w
λðwÞ ¼ SNRðwÞ ¼ ⊥ ; ð4:22Þ
σ2 wT PU w
which turns out to be the well-known Fisher’s ratio or Raleigh quotient, a criterion
commonly used for linear Fisher discriminant analysis (Duda and Hart 1973). As a
result, solving (4.22) is equivalent to solving
1
P⊥
U P⊥ T ⊥
U dd PU w ¼ λðwÞw: ð4:23Þ
by δOSP(r), was developed in Harsanyi and Chang (1994) as well as Chang (2003a, b)
and given by
δOSP ðrÞ ¼ dT P⊥
U r: ð4:24Þ
where
e w d
e ¼ κP⊥
equality , w e ¼
Ud , w ¼ κP⊥
P⊥ w 1=2 U
P⊥ w1=2 : ð4:27Þ
U U
, w ¼ kP⊥
Ud
w* ¼ κP⊥
U d: ð4:28Þ
Using w* obtained in (4.28) to specify the system δLDS(r) in Fig. 4.1 yields the
maximal SNR given by
!
α2p T ⊥
SNR w *
¼ d PU d ; ð4:29Þ
σ2
92 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
which is exactly the same as the OSP given by (4.24). Nevertheless, it should be
noted that (4.28) gives the same maximal SNR in (4.29) regardless of the scaling
constant k because (4.26) is independent of k.
The idea of (4.19) provides a clue as to how to design the following unsupervised
OSP (UOSP) algorithm to find targets of interest through successive operations of
(4.19) on a growing undesired target signal matrix U (Wang et al. 2002). It should
also be noted that the ATGP developed by Ren and Chang (2003) works by the
same principle as UOSP.
UOSP Algorithm
1. Initial condition:
Let ε be a prescribed error threshold
and
t0 a data sample vector with maximal
vector length, that is, t0 ¼ arg maxr r , where r is a data sample vector and
2
r ¼ rT r. Set k ¼ 0, and use (4.19) to define P⊥ ¼ I t0 t T t0 1 t T , with
t0 0 0
U0 ¼ ½t0 .
T 1 T
2. Let k k þ 1, apply P⊥ Uk1 ¼ I Uk1 Uk1 Uk1 Uk1 via (4.19), with
Uk1 ¼ ½t0 t1 . . . tk1 , to all image pixels r in the image, and find the kth target
tk generated at the kth stage that has the maximum orthogonal projection as
follows:
T
⊥ ⊥
tk ¼ arg maxr PUk1 r PUk1 r : ð4:31Þ
3. If mOSP ðtk Þ > ε, where mOSP ðÞ is a well-defined OSP measure, then go to step
2. Otherwise, the algorithm is terminated. At this point, all the generated target
pixels t0 , t1 , . . . , tk1 are considered the desired targets.
Originally, the OSP specified by (4.20) was designed to detect the abundance
fraction of the desired signature d, not to unmix the d since it made use of SNR
to perform signal detection, not signal estimation. To address this issue, Settle
(1996) derived a maximum likelihood estimation of αp, α^ p ðrÞ, corresponding to the
signature mp ¼ d, which was also shown later by Tu et al. (1997) to be a least-
squares OSP (LSOSP) given by
4.4 Finding Targets of Interest 93
dT P⊥
Ur
δLSOSP ðrÞ ¼ T ⊥ ; ð4:32Þ
d PU d
1
which is also identical to (4.30) and includes a constant dT P⊥
Ud to account for
the estimation error caused by the OSP. By virtue of (4.32), we can develop an
unsupervised LSOSP (ULSOSP) algorithm to find abundance-constrained least-
squares-specified targets.
ULSOSP Algorithm
1. Initial condition:
Select ε to be a prescribed error threshold, let t0 ¼ argfmaxr ½rT r g, where r is
run over all image pixel vectors, and set k ¼ 0.
T
ð1Þ ð1Þ
2. Let LSEð0Þ ðrÞ ¼ r α^ 0 ðrÞt0 rα^ 0 ðrÞt0 , and check whether
maxr LSEð0Þ ðrÞ < ε. If so, the algorithm is terminated; otherwise, continue.
3. Let k k þ 1 and find tk ¼ arg maxr LSEðk1Þ ðrÞ .
4. Apply LSOSP method with the signature matrix MðkÞ ¼ ½t0 t1 . . . tk1 to estimate
ðkÞ ðkÞ ðk Þ
^ 1 ðrÞ, α
the abundance fraction of t0 , t1 , . . . , tk1 , α ^ 2 ðrÞ, . . . , α
^ k1 ðrÞ.
5. Find the kth maximum LSE defined by
n o Xk1 ðkÞ
T Xk1 ðkÞ
ðk Þ
maxr LSE ðrÞ ¼ maxr r j¼1
^ j tj
α r j¼1
^ j tj : ð4:33Þ
α
Xp
the abundance sum-to-one constraint (ASC), j¼1
αj ¼ 1, and the abundance
nonnegativity constraint (ANC), αj 0 for all 1 j p, can be imposed on
(4.17). When only ANC is imposed on (4.17), LSOSP can be extended to
nonnegativity constrained least-squares (NCLS) method developed by Chang and
Heinz (2000) (see details in Sect. 9.2.2). Moreover, to use NCLS to find targets of
interest in an unsupervised manner, an unsupervised version, called unsupervised
NCLS (UNCLS), was further developed in Chang (2013) and Chang (2016),
described as follows.
UNCLS Algorithm
1. Initial condition:
Select ε to be a prescribed error threshold, let t0 ¼ argfmaxr ½rT r g, where r is
run over all image pixel vectors, and set k ¼ 0.
T
ð1Þ ð1Þ
2. Let LSEð0Þ ðrÞ ¼ r α^ 0 ðrÞt0 rα^ 0 ðrÞt0 , and check whether
maxr LSEð0Þ ðrÞ < ε. If so, the algorithm is terminated; otherwise, continue.
3. Let k k þ 1, and find tk ¼ arg maxr LSEðk1Þ ðrÞ .
4. Apply the NCLS method with the signature matrix MðkÞ ¼ ½t0 t1 . . . tk1 to
ðkÞ ðkÞ ðk Þ
^ 1 ðrÞ, α
estimate the abundance fraction of t0 , t1 , . . . , tk1 , α ^ 2 ðrÞ, . . . , α
^ k1 ðrÞ.
5. Find the kth maximum LSE defined by
n o Xk1 ðkÞ
T Xk1 ðkÞ
maxr LSEðkÞ ðrÞ ¼ maxr r j¼1
^
α j tj r j¼1
^ j tj : ð4:33Þ
α
UFCLS Algorithm
1. Initial condition:
Select ε to be a prescribed error threshold, let t0 ¼ argfmaxr ½rT r g, where r is
run over all image pixel vectors, and let k ¼ 0.
T
ð1Þ ð1Þ
2. Let LSEð0Þ ðrÞ ¼ r α^ 0 ðrÞt0 rα^ 0 ðrÞt0 , and check whether
4.4 Finding Targets of Interest 95
maxr LSEð0Þ ðrÞ < ε. If so, the algorithm is terminated; otherwise, continue.
3. Let k k þ 1, and find tk ¼ arg maxr LSEðk1Þ ðrÞ .
4. Apply the FCLS method with the signature matrix MðkÞ ¼ ½t0 t1 . . . tk1 to
ðkÞ ðkÞ ðk Þ
^ 1 ðrÞ, α
estimate the abundance fraction of t0 , t1 , . . . , tk1 , α ^ 2 ðrÞ, . . . , α
^ k1 ðrÞ.
5. Find the kth maximum LSE defined by
n o Xk1 ðkÞ
T Xk1 ðkÞ
ðk Þ
maxr LSE ðrÞ ¼ maxr r j¼1
^ j tj
α r j¼1
^ j tj : ð4:33Þ
α
Despite the fact that the OSP specified by (4.20) cannot unmix signatures accu-
rately, it can indeed be used to produce 2OS targets. ATGP is such an algorithm that
uses the OSP concept to find LS targets.
N
Let X be a data matrix formed by data sample vectors, fri gi¼1 , that is,
X ¼ ½r1 r2 rN . Define the norm of the data matrix X by
X ¼ max1iN ri ; ð4:35Þ
where ri is the length of the vector ri ¼ ðr i1 ; r i2 ; . . . ; r iL ÞT defined by
2 X L 2
ri ¼ r . Assume that i* ¼ arg max1iN ri . The norm of matrix
l¼1 il
X in (4.34) can be further expressed by
X ¼ r * ; ð4:36Þ
i
which is exactly the brightest pixel ri* whose norm has the maximum vector length.
That is, the maximum l2 norm defined by (4.35) is indeed the maximum pixel vector
length corresponding to the brightest pixel vector in the data set.
Using (4.36) ATGP performs a sequence of OSPs by repeatedly using orthog-
onal projectors P⊥ U ¼ I UU
#
specified by (4.19). Thus, if X is the original
96 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
ATGP
hyperspectral image cube, ATGP firstselectsan initial
target pixel t0 that yields
the norm of the space X, denoted by t0 ATGP ¼ X via (4.35), and then projects
ATGP
the space X into a subspace orthogonal to < t0ATGP > via P⊥ U0 with U0 ¼ t0 . Let
the resulting subspace be denoted by X1 ¼< t0ATGP >⊥ . Next, the ATGP a
ATGP selects
first target pixel that yields the norm of the space X1, denoted by t1 ¼ X1
via (4.35), and then projects the space X into a subspace orthogonal to < t0ATGP ,
ATGP ATGP
t1ATGP > via P⊥ U1 with U1 ¼ t0 t1 , and the resulting subspace is denoted by
ATGP ATGP ⊥
X2 ¼< t0 , t1 > . The same procedure is repeated until a stopping rule is
satisfied, that is, the number of target pixels that ATGP must extract. The details of
its algorithmic implementation are given in what follows.
3. If mðtk1 ; tk Þ > ε, where mð; Þ can be any target discrimination measure, then go
to step 2. Otherwise, the algorithm is terminated. At this point, all the generated
target pixels t0 , t1 , . . . , tk1 are considered the desired targets. It should be
noted that in Ren and Chang (2003) the mð; Þ was called the Orthogonal
Projection Correlation Index (OPCI) defined as
OPCI ¼ t0T P⊥
Uk1 t0 ; ð4:38Þ
where Uk1 ¼ ½t0 t1 . . . tk1 is the space linearly spanned by k targets, ftl gk1
l¼0 :
T
⊥ 2
P⊥ P⊥ ¼ r T P⊥ P r , then finding tk in (4.39) is
where Uk1 r Uk1 r Uk1 r ¼ Uk1
equivalent to finding a data sample vector that yields the maximal leakage into the
hyperplane specified by P⊥ Uk1 , which turns out to be the same tk as well. This
implies that ATGP can be terminated by estimating the VD with target signal
sources specified by ATGP-generated targets {tk}.
for each i ¼ 1, 2, . . . , N, where the μ and σ in (4.40) are the mean and standard
deviation of random variable zi, respectively. Since small targets can be character-
ized by data sample vectors that cause a maximum magnitude of asymmetry and
ripples of a distribution, finding a PV w that maximizes (4.40) is equivalent to
finding the direction with which these pixels are most likely aligned. By projecting
N
all data samples fxi gi¼1 on the PV w, the desired small targets can be detected by
those pixels that yield the largest projection along the direction of w.
Following the same treatment developed in Ren et al. (2006), an iterative
algorithm can be designed to find a projector that solves the following constrained
optimization problem with the optimal criterion specified by (4.40):
( ) ( )
1X N
1X N T k2 T
max zk ¼ max T
w yi yi w yi w subject to wT w ¼ 1; ð4:41Þ
w N i¼1 i w N i¼1
98 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
where zi is the projection resulting from the sphered data sample yi via the PV w.
The constraint wT w ¼ 1 is used for normalization such that the kth central moment
of the resulting data after the projection will not be affected by the magnitude of w.
Using the Lagrange multiplier to solve (4.41) results in solving the equation
h k2 T i
E yi yiT w yi λ0 I w ¼ 0; ð4:42Þ
h k2 T i
which is the eigenvector of E yi yiT w yi . Using the eigendecomposition,
(4.42) can be reduced to
h k2 T i
wT E yi yiT w yi w ¼ λ0 ð4:43Þ
T
which turns out to be the kth central moment of z ¼ w* Y. For example, (4.42) is
reduced to
T T
E yi yi wyi λ0 I w ¼ 0 for k ¼ 3; ð4:45Þ
T
E yi yi wwT yi yiT λ0 I w ¼ 0 for k ¼ 4: ð4:46Þ
It should be noted that a single PV w* solving (4.42) can only find one PV. To find
more PVs, a sequence of OSPs must be performed. In doing so, when a PV w* is found,
the decorrelated data Y are then mapped into the linear subspace < w* >⊥ orthogonal
to <w* >, which is the space linearly spanned by w*. The next PV w* is then found by
solving (4.42) in the space < w* >⊥ . The same process is continued until a stop
criterion is satisfied such as a predetermined number of projections required to be
generated. This resulting procedure is called the kth central moment projection
vector generation algorithm (PVGA) and can be described in detail as follows.
which is a combination of the third and fourth orders of statistics for ζ j in Hyvarinen
et al. [2001, Eq. (5.35), p. 115] and has been shown to approximate mutual
information that is generally used as a criterion to measure statistical independence
between two signal sources.
100 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
4.5 Target-Specified VD
Once targets of interest are found, a follow-up task is to use these targets as signal
sources under a binary composite hypothesis testing problem for TSVD.
H0 : zlk ¼ 0
versus for l ¼ 1, 2, . . . , L ; ð4:48Þ
H1 : zlk > 0
where zkl is the maximum of vector residuals given by zlk ¼ max1iN P perðk1Þ ri 2
Ul1
and P perðk1Þ ri is the orthogonal projection of each pixel vector ri, with 1 i N and
Ul1
h i
T 1
T
PV ðkÞ PV ðkÞ per ðkÞ ðk Þ ðk Þ ðk Þ
Ul1 ¼ t1
k1
. . . tl1 and nl with PUk ¼ Ul1 Ul1 Ul1 Ul1 and
l1
by the signals in this direction under H1, whereas in the noise direction under H0, it
is a low value that is governed by the maximum-norm noise residual. Under the null
hypothesis H0 with the white Gaussian noise assumption, the cdf of zkl is a Gumbel
distribution given by
( h i)
ð2logN Þ1=2 pffiffiffiffiffiffiffi
zσ 2 ðLlÞ
ð2logN Þ1=2 þ12ð2logN Þ1=2 ðloglogNþlog4π Þ
F0 ðzÞ exp e σ 2 2ðLlÞ
: ð4:49Þ
where p0(zkl ) is the probability density function (pdf) zkl under H0.
There are two detectors that can be derived to determine the value of nVD. One is
a Bayesian detector using an a posteriori probability, called the kth central moment
MOCA, denoted by k-MOCA, as
1, if Λzlk > 1,
δMOCA zlk ¼ ð4:52Þ
0, if Λ zlk 1;
where Λ zlk ¼ p H 1 zlk =p H 0 zlk . Using (4.52), the value of nVD can be deter-
mined by calculating
XL
VDk-MOCA ¼ l¼1
δMOCA zlk for k > 2: ð4:53Þ
ð
with the false alarm probability given by PF δNP
PVðkÞ ¼ p0 zlk dzlk :
Λðzlk Þτ
102 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
Finally, if the kth central moment PVGA-generated PVs in (4.53) are replaced
by FastICA-generated PVs, the resulting HFC used to estimate the value of nVD is
called Fast ICA-HFC, with VD defined by
XL 1
Fast ICA ðPF Þ ¼
VDHFC l¼1
δNP
Fast ICA zl ð4:55Þ
with the superscript “k,” as in kth central moment, replaced by the 3rd and 4th
moments, 1, and subscript “PV(k)” replaced by “ICA.”
Now, if we interpret SVD and PCA as the second central moment transformation
with k ¼ 2, then SVD-MOCA/PCA-MOCA and PCA-HFC can be considered 2OS
methods: 2-MOCA (variance-MOCA) and 2-HFC (variance-HFC), respectively.
By including the kth-order-statistics methods using the kth central moment for 3
k < 1 specified by (4.43)–(4.44) joined with the HFC method as a first-order-
statistics method, as well as negentropy-based ICA as a combination of 3rd and 4th
orders of statistics method specified by (4.54), a complete unified theory for VD can
be derived using spectral target statistics of kth order for any positive integer k; a
whole family of spectral target statistics-based VD estimation techniques are
tabulated in Table 4.2, where ATGP-HFC/ATGP-MOCA and ULSMA-HFC/
ULSMA-MOCA can be considered as joint first- and second-order-statistics-
based methods.
The HFC method and MOCA discussed in Sect. 4.3.2, along with their related
methods, are all eigenanalysis based. More specifically, the signal sources under
hypotheses are either eigenvalues in (4.1) or eigenvectors/singular vectors in (4.9).
Since eigenvalues/eigenvectors are obtained by PCA, they are not real target signal
sources. As expected, the VD estimated by the eigenanalysis-based methods is well
below the number of real target signal sources. Unfortunately, this issue was not
addressed until the work reported in Chang et al. (2014a, b), where the real target
signal sources were generated by specific algorithms.
4.5.2 ATGP-Specified VD Using ηl ¼ tlATGP 2
One particular algorithm of interest is ATGP, which has been widely used in
hyperspectral target detection to find unsupervised
SVD 2 target signal sources.
Replacing
tSVD in (4.7) with t ATGP
in (4.39) and ηl ¼ t in (4.8) with ηl ¼ t ATGP 2 , we
l l l l
can reformulate (4.9) as a new hypothesis testing problem:
H 0 : ηl ¼ tlATGP 2 p ηl H 0 ¼ p0 ðηl Þ
versus for l ¼ 1, 2, , L ; ð4:56Þ
H 1 : ηl ¼ tlATGP 2 p ηl H 1 ¼ p1 ðηl Þ
where the signal sources under the hypotheses are now real target signal sources
rather than eigenvectors in (4.9) or eigenvalues in (4.1). Then we can follow the
same treatment as what MOCA did to derive
l*, ATGP ¼ arg min1lL p H0 ηl p H 1 ηl ð4:57Þ
and ATGP-specified VD as
The problem with solving (4.60) is not straightforward since MOCA cannot be
pffiffiffiffi
directly applied to finding the pdfs of ηl under each of the hypotheses in (4.60).
Nevertheless, the same extreme theory in Leadbetter (1987) can be used to modify
MOCA for this purpose as follows.
Assume that fXi gi¼1 N
, where Xi ¼ ðXi1 ; Xi2 ; ; XiL ÞT is an L-dimensional
random vector, of which each component, Xil, is a Gaussian random variable with
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
XL ffi
zero mean and variance σ 2. Let Y i ¼ l¼1
X2 ¼ Xi . Then Yi follows a
il
chi-squared distribution of degree L and σ and its probability density function is
pffiffiffi 1L:=2 L1 pffiffi
given by f Y ðyÞ ¼ 2σ 2ΓðL=2yÞσ L ey =2σ , with mean E½Y ¼ 2σΓΓðððL=2
Lþ1Þ=2Þ
2 2
Þ and variance
2
2 2σ 2 ΓðL=2þ1Þ
var½Y ¼ Lσ ðE½Y Þ calculated by E Y ¼ ΓðL=2Þ ¼ Lσ .
2 2
As L ! 1 Y i ¼ Xi χ ðLÞ N ðE½Y i , var½Y i Þ.
Theorem 4.1 (Kuybeda et al. 2007) MN ¼ max1iN N i , where Ni is a standard
Gaussian random variable. Then
P aN MN bN u ¼ GðuÞ ¼ expðexpðuÞÞ
and
FV ðvÞ ¼ P
ðV vÞ ¼ Pðσ Y max1iN N i þ μ Y vÞ
v μY
¼ P max1iN N i
σY
v μY v μY ð4:63Þ
¼ P MN ¼ G an bn
σY σY
v ð μ Y þ bN σ Y Þ v ð μ Y þ bN σ Y Þ
¼G ¼ exp exp :
σ Y =aN σ Y =aN
Now if we replace (4.49) with (4.62) and (4.55) with (4.60), we can perform the
binary composite hypothesis testing problem specified by (4.60) as follows:
pffiffiffiffiffiffiffiffiffi pffiffiffiffi pffiffiffiffi
l*, ATGP
¼ arg min1lL p H 0 ηl p H 1 ηl ð4:65Þ
Also, by virtue of (4.60) and (4.61), we can also derive a new version of the
pffiffiffiffi
MLD-MOCA using ηl
pffiffiffiffiffiffiffiffiffiffiffi pffiffiffiffi pffiffiffiffi
l*, MOCA
¼ arg min1lL p H 0 ηl p H 1 ηl ð4:69Þ
pffiffiffiffiffiffiffiffiffiffiffi pffiffiffiffiffiffiffiffiffiffiffi
VD MOCA ¼ l*, MOCA ; ð4:70Þ
4.5.4 Discussions
The synthetic image simulated in Fig. 1.15 is shown in Fig. 4.2 with five panels in
each row simulated by the same mineral signature and five panels in each column
having the same size.
Among the 25 panels are five 4 4 pure pixel panels for each row in the first
column and five 2 2 pure pixel panels for each row in the second column, five
2 2 mixed pixel panels for each row in the third column and five 1 1 subpixel
panels for each row in the fourth and fifth columns, where the mixed and subpanel
pixels were simulated according to the legends in Fig. 4.2. Thus, a total of 100 pure
pixels (80 in the first column and 20 in the second column), referred to as
endmember pixels, were simulated in the data by the five endmembers A, B, C,
K, and M. The area marked “BKG” in the upper right corner of Fig. 1.14a was
selected to find its sample mean, that is, the average of all pixel vectors within the
“BKG” area, denoted by b and plotted in Fig. 1.14b, to be used to simulate the
background (BKG) for the image scene with a size of 200 200 pixels in Fig. 4.2.
The reason for this background selection is empirical since the selected “BKG” area
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
seemed more homogeneous than other regions. Nevertheless, other areas can also
be selected for the same purpose. This b-simulated image background was further
corrupted by an additive noise to achieve a certain SNR, which was defined as a
50 % signature (i.e., reflectance/radiance) divided by the standard deviation of the
noise in Harsanyi and Chang (1994). Once target pixels and background are
simulated, two types of target insertion can be designed to simulate experiments
for various applications.
The first type of target insertion is target implantation (TI), which can be
simulated by inserting clean target panels into a noisy BKG image by replacing
their corresponding BKG pixels, where the SNR is empirically set at 20:1. That is,
TI implants clean target panel pixels into the noise-corrupted BKG image with
SNR ¼ 20:1, in which case there are 100 pure panel pixels in the first and second
columns.
A second type of target insertion is target embeddedness (TE), which is simu-
lated by embedding clean target panels into a noisy BKG image by superimposing
target panel pixels over the BKG pixels, where the SNR is empirically set to 20:1.
That is, TE embeds clean target panel pixels into noise-corrupted BKG image with
SNR ¼ 20:1, in which case none of the 100 pure panel pixels in the first and second
columns is pure any longer. In other words, a salient difference between TI and TE
is worth mentioning. Since TE inserts targets by adding target pixels to and
superimposing over background pixels instead of replacing background pixels the
way TI does for target insertion, the abundance fraction of the pixel into which a
target pixel is embedded does not sum to one.
Table 4.3 tabulates the values of nVD estimated for TI and TE by HFC and
NWHFC, where NWHFC produced nVD ¼ 6 for all false alarm probabilities versus
HFC, which produced various values for nVD. However, both HFC and NWHFC
produced the same value of nVD ¼ 6 with PF 103 . Also included in Table 4.3 are
results produced by SSE and HySime, where their results were very close to that
produced by NWHFC.
For TSVD, Table 4.4 tabulates the values of nTSVD estimated using 2OS
specified by ATGP methods, ATGP-MOCA, and ATGP-HFC, where n2OSVD ¼ 8
and 9 for TI and n2OSVD ¼ 7 for TE.
Table 4.5 also tabulates the values of nTSVD estimated by various methods using
HOS-specified targets, where nHOSVD ¼ 5 for TI and nHOSVD ¼ 6 for TE.
A comparison of Tables 4.4 and 4.5 reveals that the results make sense. The
reason the values of VD, n2OS-VD, estimated by MOCA and HFC using ATGP-
specified targets produced higher values than that produced by methods using
Table 4.3 nVD estimated for synthetic images by HFC and NWHFC
Data SSE/HySime PF 101 102 103 104 105
TI 5/5 HFC 18 13 12 8 8
NWHFC 6 6 6 6 6
TE 6/5 HFC 14 9 6 6 6
NWHFC 6 6 6 6 6
4.6 Synthetic Image Experiments 109
Table 4.4 nVD estimated by 2OS-VD for synthetic images by ATGP-MOCA and ATGP-HFC
Data ATGP-MOCA PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
TI 9 9 9 9 8 8
TE 7 7 7 7 7 7
Table 4.6 nTSVD estimated for TI and TE by MOCA and ATGP using signal energy and strength
MLD NPD NPD NPD NPD NPD
l* PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
⊥
TI P r MOCA 6 6 6 6 6 6
U
ATGP 7 8 7 7 7 7
⊥ 2
P r MOCA 6 6 6 6 6 6
U
ATGP 8 9 8 8 7 7
⊥
TE P r MOCA 6 6 6 6 6 5
U
ATGP 6 6 6 6 6 5
⊥ 2
P r MOCA 6 6 6 6 6 6
U
ATGP 6 6 6 6 6 6
Although many real hyperspectral image scenes can be used for experiment, for
comparative analysis, here we used two real image scenes that have been studied
extensively in the literature.
The image scene shown in Fig. 4.3 (also shown in Fig. 1.10a) was used for
experiments. It was acquired by the airborne HYperspectral Digital Imagery Col-
lection Experiment (HYDICE). It has a size of 64 64 pixel vectors with 15 panels
in the scene and the ground truth map in Fig. 4.3b (Fig. 1.10b).
It is worth noting that panel pixel p212, marked yellow in Fig. 4.3b, is of
particular interest. Based on the ground truth, this panel pixel is not a pure panel
pixel and is marked yellow as a boundary panel pixel. However, with our extensive
and comprehensive experiments, this yellow panel pixel is always extracted as the
one with the most spectrally distinct signature compared to the R panel pixels in
row 2. This indicates that a signature of spectral purity is not equivalent to a
signature of spectral distinction. As a matter of fact, in many cases, panel pixel
p212 instead of panel pixel p221 is the one extracted by EFAs to represent the panel
signature in row 2. Also, because of such ambiguity, the panel signature
representing panel pixels in the second row is either p221 or p212, which is always
the last one found by EFAs. This implies that the ground truth of R panel pixels in
the second row in Fig. 4.3b may not be as pure as was thought.
Table 4.7 tabulates the values of nVD estimated for HYDICE by four VD
estimation methods, HFC/NWHFC and SSE/HySime, and the values of nVD esti-
mated by various VD estimation methods, ATGP-MOCA and ATGP-HFC, using
2OS-specified by the ATGP and HOS-based method using HOS-specified targets.
a b
p11, p12, p13,
Fig. 4.3 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels
4.7 Real Image Experiments 111
Table 4.8 nTSVD estimated for HYDICE by MOCA and ATGP using signal energy
Table 4.9 nTSVD estimated for HYDICE by MOCA and ATGP using signal strength
MLD NPD NPD NPD NPD NPD
pffiffiffiffi ATGP
ηl ¼ t l l* PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
MOCA 32 33 32 31 31 30
ATGP 41 43 41 40 39 36
much higher than their corresponding values of VD, nVD, in Table 4.7 and
MOCA in Tables 4.8 and 4.9. This makes sense because some spectrally distinct
real target signal sources may be projected onto the same eigenvectors produced
by MOCA.
A second real image scene used for experiments is a well-known Airborne Visible
InfraRed Imaging Spectrometer (AVIRIS) image scene, Cuprite, shown in Fig. 4.4
(see also Fig. 1.4a) available at the USGS Web site http://aviris.jpl.nasa.gov/. This
scene is a 224-band image with a size of 350 350 pixels; it was collected at the
Cuprite Mining District site in Nevada in 1991. It is one of the most widely used
hyperspectral image scenes available in the public domain and has 20 m spatial
resolution and 10 nm spectral resolution in a range of 0.4–2.5 μm. Since it is well
understood mineralogically and has reliable ground truth, this scene has been
studied extensively. Two data sets for this scene, reflectance and radiance data,
are also available for study. Five pure pixels in Fig. 4.4a, b can be identified as
corresponding to five different minerals, alunite (A), buddingtonite (B), calcite (C),
kaolinite (K), and muscovite (M) labeled A, B, C, K, and M in Fig. 4.4b, along with
their respective spectral signatures plotted in Fig. 4.4c, d.
Although the data set contains more than five minerals, the ground truth avail-
able for this region only provides the locations of the pure pixels: alunite (A),
a b
Fig. 4.4 (a) Cuprite AVIRIS image scene; (b) spatial positions of five pure pixels corresponding
to minerals: A, B, C, K, and M; (c) five mineral reflectance spectra; (d) five mineral radiance
spectra
4.7 Real Image Experiments 113
c
7000
Muscovite Alunite
Buddingtonite
6000 Calcite
Alunite Kaolinite Kaolinite
Muscovite
5000
Reflectance
4000 Buddingtonite
3000
Calcite
2000
1000
0
400 600 800 1000 1200 1400 1600 1800 2000 2200 2400
Wavelength(nm)
d
12000
Alunite
Muscovite Kaolinite Buddingtonite
Calcite
10000
Muscovite
Kaolinite
8000 Alunite
Radiance
6000
4000
Calcite
Buddingtonite
2000
0
400 600 800 1000 1200 1400 1600 1800 2000 2200 2400
Wavelength(nm)
Fig. 4.4 (continued)
buddingtonite (B), calcite (C), kaolinite (K), and muscovite (M) shown in Fig. 4.4b
with their reflectance and radiance spectra plotted in Fig. 4.4a, b.
Tables 4.10 and 4.11 tabulate the values of VD estimated for cuprite reflectance
and radiance data by 2OS- and HOS-based methods, respectively.
Since background is generally characterized by 2OS and targets are generally
specified by HOS, it is expected that the VD values estimated by 2OS-based
114 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
Table 4.10 nTSVD estimated for Curpite reflectance data by various VD estimation techniques
SSE/HySime 27/16
2OS-based MOCA PCA 31
ATGP 51
HOS-MOCA Skewness 21
Kurtosis 24
Fifth moment 30
ICA 23
NPD-based VD PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
2OS-based HFC 34 30 24 22 20
HFC NWHFC 32 29 25 23 21
PCA-HFC 31 31 30 30 30
ATGP-HFC 51 51 50 48 47
HOS-HFC Skewness- 22 21 21 21 21
HFC
Kurtosis- 24 24 22 22 22
HFC
Fifth-HFC 30 30 29 28 28
ICA-HFC 25 23 23 23 22
Table 4.11 nTSVD estimated for Curpite radiance data by various VD estimation techniques
SSE/HySime 18/17
Second order-MOCA PCA 46
ATGP 51
HOS-MOCA Skewness 25
Kurtosis 23
Fifth moment 28
Fast ICA 25
NPD-based VD PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
2OS -based HFC 29 18 17 15 15
HFC NWHFC 23 20 20 19 19
PCA-HFC 47 46 44 43 43
ATGP-HFC 52 51 51 51 48
HOS-HFC Skewness- 22 21 21 21 21
HFC
Kurtosis- 24 24 22 22 22
HFC
Fifth-HFC 30 30 29 28 28
Fast ICA- 25 23 23 23 22
HFC
4.7 Real Image Experiments 115
a
vegetation
cinders
shade
anomaly
rhyolite
dry lake
b
1800
anomaly
1600 cinders
dry lake
1400 rhyolite
shade
1200 vegetation
Radiance
1000
800
600
400
200
0
0.40 µm 2.5 µm
Wavelength
Fig. 4.5 (a) AVIRIS LCVF scene. (b) Spectra of anomaly, cinders, dry lake, rhyolite, shade, and
vegetation
methods will be higher than those estimated by HOS-based methods, where back-
ground signatures extracted by 2OS-based methods were not extracted by
HOS-based methods. The results in Tables 4.8 and 4.9 confirm this conjecture.
A third AVIRIS image scene used for experiments is shown in Fig. 4.5 (also
Fig. 1.8a). It is the Lunar Crater Volcanic Field (LCVF) located in Northern Nye
County, Nevada, which is one of the earliest hyperspectral image scenes studied in
the literature. Atmospheric water bands and low-SNR bands were removed from
the data, reducing the image cube from 224 to 158 bands. The LCVF image has a
10 nm spectral resolution and 20 m spatial resolution.
116 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
Unlike the Cuprite data in Fig. 4.4, which is a mineralogical scene containing
many pure mineral signatures this image scene in Fig. 4.5a has five targets of
interest, red oxidized basaltic cinders, rhyolite, playa (dry lake), vegetation, and
shade whose radiance spectra are plotted in Fig. 4.5b. According to the scene, these
three target areas, except vegtation, cover very large regions, and it is anticipated
that these three regions will be heavily mixed. Thus, it is expected that there will not
be too many pure signatures as endmembers. However, owing to a volanic explo-
sion, the crater lake in the lower right section may actually contain pure mineral-
ogical signatures as endmembers. In addition to these five target signatures, there is
also an interesting target, an anomaly two pixels in size located in the upper left dry
lake, with its spectral signature also plotted in Fig. 4.5b. Thus, if endmembers are
present in this scene, they will very likely be in areas of the dry lake and vegetation,
along with anomalies. Table 4.12 tabulates the values of VD estimated by 2OS- and
HOS-based methods. As we see from the table, the results make sense and can be
categorized into two classes, values around 11 1 estimated by 2OS-based
methods, which include background signatures, and values around 3 or 4 estimated
by HOS-based methods, which only specify targets by HOS.
In analogy with HYDICE experiments, Table 4.13 also tabulates the value of
TSVD, nTSVD, estimated by MOCA and ATGP-specified target signal energy and
signal strength, respectively, where MLD is the maximum likelihood detector used
by the MOCA and ATGP to replace the NPD.
As we can see from Table 4.13, conclusions similar to those drawn from
Tables 4.7 to 4.9 also apply to Tables 4.11, 4.12, and 4.13, where the values of
TSVD, nTSVD, estimated by ATGP-specified target signal energy and strength in
Table 4.12 nTSVD estimated for LCVF data by various VD estimation techniques
SSE/HySime 12/11
Second order-MOCA PCA 4
ATGP 12
HOS-MOCA Skewness 3
Kurtosis 3
Fifth moment 5
Fast ICA 3
NPD-based VD PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
2OS -based HFC 8 6 4 4 4
HFC NWHFC 20 13 11 11 10
PCA-HFC 4 4 4 4 4
ATGP-HFC 12 11 10 10 9
HOS-HFC Skewness- 4 3 3 3 3
HFC
Kurtosis- 3 3 3 3 3
HFC
Fifth-HFC 5 4 4 4 4
Fast ICA- 3 3 3 3 3
HFC
Table 4.13 nTSVD estimated by MOCA and ATGP using signal energy and strength
4.7 Real Image Experiments
MLD l* NPD PF ¼ 101 NPD PF ¼ 102 NPD PF ¼ 103 NPD PF ¼ 104 NPD PF ¼ 105
⊥
Cuprite reflectance P r
U
MOCA 40 41 40 39 38 37
ATGP 67 71 67 64 58 53
⊥ 2
P r
U
MOCA 43 45 43 40 40 39
ATGP 74 80 73 69 67 65
⊥
Cuprite radiance P r
U
MOCA 43 45 43 41 39 38
ATGP 66 74 66 61 59 54
⊥ 2
P r
U
MOCA 47 48 45 45 42 41
ATGP 74 81 74 68 63 61
⊥
LCVF P r
U
MOCA 7 7 7 7 7 7
ATGP 9 12 11 9 8 8
⊥ 2
P r
U
MOCA 8 9 8 7 7 7
ATGP 12 12 12 11 9 9
117
118 4 Target-Specified Virtual Dimensionality for Hyperspectral Imagery
Table 4.13 were much higher than the corresponding values of VD, nVD, in
Tables 4.10, 4.11, and 4.12 and of nTSVD by MOCA in Table 4.13. This is true
since some spectrally distinct real target signal sources may be projected onto the
same eigenvectors produced by MOCA.
4.8 Conclusions
This chapter revisits the concept of VD and presents a unified theory to expand the
applicability of VD from a broader point of view, where spectrally distinct signa-
tures defined by VD can be interpreted as various types of targets characterized by
data statistics explored in Chang (2013). Within this context many works reported
in the literature can be considered VD estimation methods for hyperspectral targets
specified by the first two orders of data statistics. Depending on how spectrally
distinct signatures are defined as desired targets of interest, VD emerges as a
versatile concept that can be used to adapt various applications. For example, if
hyperspectral targets of interest are specified by first-order statistics, then the HFC
method can be used for this case. On the other hand, if hyperspectral targets of
interest exhibit 2OS, then 2OS-based techniques, such as least squares methods
(Paylor, 2014; Paylor and Chang, 2013, 2014) PCA and signal subspace methods,
may be applicable to estimating the value of nVD. On the other hand, when target
signatures of major interest are rare, signal sources such as endmembers, anomalies,
and their data statistics generally go beyond the first two orders of data statistics.
In this case, they can be better characterized by HOS. To the authors’ best
knowledge the HOS-based approach to finding the value of nVD developed by
Chang et al. (2014a, b) was developed to resolve this issue and deviate from
conventional wisdom by finding specific HOS-characterized feature vectors that
can be used to generate desired target signal sources. These feature vectors are then
used to perform Neyman–Pearson detection to determine whether a test feature
vector is a target signal source. The novelty of the HOS methods is to integrate all
key ideas—Neyman–Pearson detection formulations in the HFC method (Harsanyi
et al., 1994a), HOS target generation algorithms in Ren et al. (2006), and finding
PDs under binary hypotheses via the extreme value theory in Kuybeda
et al. (2007)—into a new aspect of HOS-based binary hypothesis testing problems
that can be used to solve VD estimation problems for HOS hyperspectral target
signal sources. Table 4.14 summarizes various types of VD estimation techniques
according to their categorization in Chang (2013a, b), along with TSVD as
presented in this chapter.
4.8 Conclusions 119
5.1 Introduction
detect the desired signal source d so as to accomplish target detection. Details about
their relationships can be found in Chap. 5 in Chang (2016).
Since finding complete background knowledge is nearly impossible, CEM has
emerged as a very attractive technique because it requires no prior knowledge other
than about targets of interest. In particular, no background knowledge is needed to
process CEM. This advantage makes CEM perform subpixel detection very effec-
tively, specifically when a known target signature is identified and targets specified
by this target signature need to be found in various locations, such as in rescue and
search or finding combat vehicles in an unknown battlefield. Moreover, many such
targets of interest may also be moving objects that can be detected only by ongoing
processing in real time, in which case the capability to process in real time becomes
essential to implementing CEM. Most importantly, there may also be certain subtle
targets that can be only revealed and captured by real-time processing but may be
missed and overwhelmed by targets detected by CEM with no real-time processing.
Despite the fact that many real-time processing algorithms have been proposed
in the literature for target detection and classification, technically speaking, none of
these are true real-time processing algorithms; they are, rather, fast computational
algorithms. Theoretically, a true real-time processing algorithm must produce its
output at the same time as input comes in. However, in reality this is impossible
since there is always a time delay caused by data processing. With this interpreta-
tion all claimed real-time processing algorithms can only be called near real time,
with the assumption that the data processing time is negligible. Examples include
constrained linear discriminant analysis (CLDA) (Du et al. 2003a; Du and Nevovei
2005, 2009) and parallel processing algorithms such as anomaly detection using a
multivariate normal mixture model and graphic processing unit (GPU) processing
(Tarabalka et al. 2009). Nevertheless, from a practical point of view, such a time
delay is actually determined by the specific application. For instance, in surveil-
lance and reconnaissance applications, finding moving targets such as missiles is
urgent, and the response time must be instantaneous. In this case, very little time
delay should be allowed. As another example, for applications in fire damage
management/assessment, the time to respond can be minutes or hours, in which
case the allowable time delay can be longer. Thus, as long as an algorithm can meet
a required time constraint, it can be considered a real-time processing algorithm.
In particular, this chapter looks into issues of how to take full advantage of CEM
for subpixel target detection. As noted earlier, one of the great advantages of CEM
is that it suppresses background variation with sample vectors. It designs and
develops “causal processing” for real-time implementation via the concept of
innovations information, discussed in Chap. 3 as well as in Poor (1994, footnote
p. 78), which is defined as new information of the current input sample that cannot
be predicted from past events. It was originally proposed by Kailath (1968) who
redeveloped a Kalman filter and showed it to be a very promising and effective
means of updating data in a causal and real-time manner. To investigate these
issues, this chapter develops a progressive version of CEM, called progressive
CEM (P-CEM), that processes data in a causal manner sample by sample,
progressively and in real time. Since P-CEM requires repeatedly calculating the
126 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
sample-varying correlation matrix R1 ðrn Þ with the current data sample vector rn,
its computing time is expected to be very high. To further reduce the computational
complexity, the notion of innovation information is further used to derive recursive
equations to update R1 ðrn Þ without recalculating all previously visited data sample
vectors except rn. As a result, a recursive version of P-CEM can therefore be
developed, called recursive CEM (R-CEM), which can not only process data
sample by sample progressively but also update data processing results recursively
based only on the data sample vector rn currently being processed and previously
processed data in the same way that a Kalman filter does. Such R-CEM is indeed the
final version of real-time processing of CEM, to be called real-time CEM
(RT-CEM) in the rest of this chapter.
Note that the idea of the proposed causal processing arises in updating needed
information only through the data sample vector currently being processed and the
information generated by processing previous data sample vectors. Although it is
similar to Kalman filtering (Chap. 3), there are several notable differences between
the two. First, since CEM is performed on a single data sample basis, there is no
counterpart of a state equation as used by a Kalman filter found in CEM. Second,
the measurement equation used by CEM is quite different from that used by a
Kalman filter, where a noise term involved in a Kalman filter is not present in CEM.
Third, CEM usually requires the inversion of a sample correlation/covariance
matrix. To implement CEM in real time, the matrix inversion must be updated
sample by sample. Such an update is not found in the measurement equation in a
Kalman filter, which is updated from the state equation. Thus, a direct application
of real-time Kalman filtering to CEM is not feasible due to the lack of a state
equation that can be derived for CEM. To resolve this issue, an alternative approach
is to use Woodbury’s identity in Appendix A to derive causal innovation informa-
tion update equations for PC-CEM in the same way that is used to derive a Kalman
filter so that only the current data sample vector and the processed information
obtained by previous data sample vectors can be used to generate the innovation
information to update data processing results sample by sample.
This section briefly reviews CEM developed by Harsanyi in his dissertation (1993).
It assumes that a desired signature specified by d is used for active detection. CEM
uses this designated signature d to constrain signals passed through a custom-
designed finite impulse filter (FIR) while minimizing the least-squares error caused
by data samples that fail to pass the filter due to that fact that they are not spectrally
related to the desired signature d and can be considered interfering signal sources.
Accordingly, a major strength resulting from CEM is that it requires no prior
knowledge about the data to be processed other than the desired signature d to
achieve background suppression. This advantage is very significant and unique
5.2 Constrained Energy Minimization 127
minw wT Rw subject to dT w ¼ wT d ¼ 1: ð5:3Þ
Differentiating J(w) in (5.4) with respect to w and setting it to zero we can find the
optimal solution w* by
∂J ðwÞ
¼ 0 ) 2Rw* þ λd ¼ 0 ) w* ¼ ð1=2ÞλR1 d: ð5:5Þ
∂w w*
R1 d
w* ¼ : ð5:7Þ
dT R1 d
dT R1 r
δCEM ðrÞ ¼ : ð5:8Þ
dT R1 d
To explore the real-time capability of CEM, we need to understand how data are
acquired. Based on data acquisition, three data formats have been widely used by
hyperspectral imaging sensors (Schowengerdt 1997). One is the band-sequence
(BSQ) format, which collects data band by band. The other two are BIP/BIS, which
collects data sample by sample with full band information, and BIL, which collects
data line by line with full band information as well. Obviously, for real-time
processing, BIP/BIS and BIL are more appropriate compared to BSQ, which is
mainly focused on individual band processing rather than data sample processing,
where the former is referred to as hyperspectral sample processing and the latter as
hyperspectral band processing. Interestingly, while CEM using the BSQ format was
recently investigated in Chang et al. (2015a, b, c, d, e) and will also be discussed in
Chap. 13, the real-time sample processing of CEM using BIP/BIS and BIL has not
been addressed and reported in the literature, at least to the best of my knowledge.
Furthermore, although BIP/BIS and BIL look similar, their real-time
implementations are quite different owing to their use of causal sample correlation
5.2 Constrained Energy Minimization 129
matrices, that is, one is updated by samples and the other by data lines. As a result,
two different recursive equations are derived for CEM using BIP/BIS and BIL
respectively for real-time implementation.
Basically, between BIL and BIP/BIS, the latter is the only format that can be
used to perform CEM in real time. However, since hyperspectral imaging sensors
currently being used for data acquisition are carried out in a push-broom fashion,
their data are collected line by line. Accordingly, it makes more sense for CEM to
be implemented in such a manner as a compromise. As noted, the main problem
with the real-time implementation of CEM is its calculation of the global data
sample correlation matrix R, which requires entire data samples. To make CEM a
real-time processing algorithm, the R must be able to adapt to the data samples
currently being processed. In addition, it must also be causal in the sense that it can
only use data samples it has already visited and processed; no future data samples
should be allowed in the data processing. To deal with this causal issue, two types
of causal correlation matrix are defined to implement BIP/BIS and BIL formats. For
the BIP/BIS format, data samples are processed pixel by pixel/sample by sample. In
this case, a BIP/BIS-based CSCRM specified by R(n) can only be formed by data
samples up to the pixel/sample currently being processed, the nth data sample rn.
More specifically, let fri gn1
i¼1 be a set of data samples already processed by CEM.
Then a causal Xn sample correlation used for BIP/BIS should be formed by
RðnÞ ¼ ð1=nÞ i¼1 ri ri . For the BIL format, data samples are processed line by
T
line. Thus, a causal line correlation matrix (CLCRM) specified by R(L(n)) should
be formed by only those data lines up to the data line currently being processed,
denoted by the nth data line, Ln, where L(n) comprises all data lines up to Ln, that is,
LðnÞ ¼ [l¼1n
Ll, where Ll is the lth data line. If we further assume that the number of
N
samples in each line is N, then the set of data samples, xin i¼1 , forms the nth data
line, Ln, which is represented by a data matrix Xn ¼ x1n x2n xNn . Thus, RðLn Þ
XN T
¼ ð1=N ÞXn XnT ¼ ð1=N Þ i¼1 xin xin is a CLCRM formed only by the nth data
line. By virtue of R(Ln), we can define the BIL-based CSCRM for BIL up to the nth
Xn Xn, N T
data line Ln as RðLðnÞÞ ¼ ð1=nÞ l¼1 RðLl Þ ¼ ð1=nN Þ l¼1, i¼1 xil xil . By com-
paring BIL-based causal R(L(n)) to R(n) used by BIP/BIS, it is expected that
developing real-time CEM using the BIL format is more sophisticated than CEM
using BIP/BIS, as will be shown subsequently in derivations.
There are several advantages of implementing RT-CEM over CEM. First is its
ability to perform CEM in such a manner that various levels of background
suppression for each data sample or data line can be observed visually and analyzed
for detection. Second, these various background suppressions provide users with
sample-varying or data line-varying detection maps. For example, when weak
targets are present in earlier stages, the profiles of sample-varying or data line-
varying background suppressions provided by RT-CEM will help extract such
targets before they are overwhelmed by subsequently detected strong targets.
Third, RT-CEM offers a means of implementing CEM in real time. In fact, the
130 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
CEM currently in use is not a true real-time processing algorithm, although its
processing time is very fast and it can be implemented in a near-real-time fashion.
This is because causality is a prerequisite for real-time processing algorithms, and
CEM does not satisfy causality. Finally, recursive innovation information update
equations developed for RT-CEM paves the way to future hardware implementa-
tion for chip design, such as a field-programmable gate array (FPGA).
Now, assume that rn is the data sample vector currently being visited. Then fri gn1
i¼1
represents the set of data samples already visited and processed. Then the concept
of CSCRM, denoted by R(n) up to rn, is defined by
Xn
RðnÞ ¼ ð1=nÞ r rT:
i¼1 i i
ð5:9Þ
As a result, for a given current data sample rn a causal CEM (C-CEM) can be
rederived from (5.8) via R(n), defined in (5.9) as follows:
dT R1 ðnÞrn
δC-CEM ðrn Þ ¼ : ð5:10Þ
dT R1 ðnÞd
It should be noted that the causal sample autocorrelation matrix R1(n) defined
1
in (5.9) can be reexpressed as R1 ðnÞ ¼ ððn 1Þ=nÞRðn 1Þ þ ð1=nÞrn rnT . By
virtue of (5.11), we can derive a causal update equation for δRC-CEM(rn) in a manner
similar to how a Kalman filter is derived in Poor (1994). More specifically, the
derived causal update equation is obtained by dictating the difference between the
pixel currently being processed, rn, and the processed information obtained by the
1
previous n 1 data sample vectors fri gn1
i¼1 , R (n 1), that is, the information that
is contained in rn but that cannot be obtained and predicted from previously visited
5.3 RT-CEM Using BIP/BIS 131
Fig. 5.1 Diagram showing implementation of recursive equation specified by (5.12) to find R1
(n)
R1
n ðnÞ ¼ ½ð1 1=nÞRðn 1Þp1 on o
ffiffiffi pffiffiffi
½ð1 1=nÞRðn 1Þ1 ð1= nÞrn ð1= nÞrnT ½ð1 1=nÞRðn 1Þ1
pffiffiffi pffiffiffi :
1 þ ð1= nÞrnT ½ð1 1=nÞRðn 1Þ1 ð1= nÞrn
ð5:12Þ
e ðn 1Þ and e pffiffiffi
Furthermore, let ððn 1Þ=nÞRðn 1ÞR r n ¼ ð1= nÞrn in (5.12); then
e 1 ðn 1Þ ¼ ½ð1 1=nÞRðn 1Þ1 . Figure 5.1 presents a diagram implementing
R
(5.12), where D is one time unit delay. In particular, it shows how R e ðnÞ can be
updated recursively by the previously calculated data sample vector, R e ðn 1Þ, and
the data sample currently being processed, e r n , via (5.12). This diagram was also
derived in Chen et al. (2014a, b, c) since both anomaly detection and CEM use the
same causal correlation matrix R(n) to account for spectral statistics provided by
causal data sample vectors.
Now, using (5.11) in conjunction with the causal recursive update equation
specified by (5.12) CEM can be further implemented as a recursive version of
CEM, R-CEM. Interestingly, according to Fig. 5.1, C-CEM can also be
implemented as a real-time version of CEM, referred to as real-time CEM
(RT-CEM), δRT-CEM ðrn Þ, as follows:
RT-CEM using the BIL format is quite different from RT-CEM using the BIP
format because the former must be implemented line by line, in which case data
sample vectors are only available after a line of data sample vectors is completed.
Thus, the update of the sample correlation matrix must be done line by line instead
of sample by sample. Also, assume that the size of a data matrix is M N, where
there are M data lines and each line has N pixels. Let Ll be the lth data line and L(n)
N , n
the data comprising all of the first n data lines, that is, LðnÞ ¼ [l¼1
n
Ll ¼ ril i¼1, l¼1 .
XN T
Then RðLl Þ ¼ ð1=N Þ i¼1 ril ril , where the number of pixels in each line is N and
Xn Xn, N T
RðLðnÞÞ ¼ ð1=nÞ l¼1 RðLl Þ ¼ ðnN Þ1 l¼1, i¼1 ril ril . To drive a real-time
version of CEM using the BIL format, we use the following matrix identity in
Appendix A:
1
ðA þ BCDÞ1 ¼ A1 A1 B DA1 B þ C1 DA1 : ð5:14Þ
matrix, then
ð5:17Þ
I
(n-1)-1 -1 D
n/(n-1)
R-1(L(n-1))
Xn Ln R(Ln)
short of deriving recursive equations for all the algorithms, which is actually crucial
for them to be implemented in real time. It is our belief that without these recursive
equations, their implementation will not be considered as a true real-time process.
As will be demonstrated in what follows, deriving such recursive equations is not a
trivial matter.
Let Xn ¼ r1n r2n rNn be the nth data line, where rni is the ith pixel and N is the
total number of pixels in each line. If CEM uses the desired signature d and operates
on the correlation matrix formed by the (n 1)st data lines, then, by virtue of (5.17),
CEM using the desired signature d and the nth line data matrix Xn is given by
In this case, CEM using the desired signature d and the nth line data matrix Xn can
be derived by the following Eq. (5.20) as follows:
dT R1 ðLðnÞÞX(
n )
1
T 1
¼ ðn=ðn 1ÞÞ d R ðLðn 1ÞÞXn d R R þI T 1
ðR ðL ðn 1 ÞÞ Þ X n ;
n n1 n n1
ð5:19Þ
1
d R ðLðnÞÞ (d
T
)
1
T 1
¼ ðn=ðn 1ÞÞ d R ðLðn 1ÞÞd d R R þI 1
ðRðLðn 1ÞÞÞ d :
T
n n1 n n1
ð5:20Þ
RT -CEM
T 1 1 T 1
δn1
ðXn Þ ¼ d R ðLðn 1ÞÞd d R ðLðn 1ÞÞXn
1 1
¼ dT R1 ðLðn 1ÞÞd dT R1 ðLðn 1ÞÞr1n , . . . , dT R1 ðLðn 1ÞÞd dT R1
T
ðLðn 1ÞÞrNn
ð5:21Þ
and
8 T 1
9
>
> d R ðLðn 1ÞÞd RT -CEM >
>
>
< δ ðXn Þ >
=
dT R1 ðLðnÞÞd " n1 #
¼ ðn=ðn 1ÞÞ
1 :
>
> T 1
1 1 >
>
>
: ð ð Þ Þd T
þ ð ð ÞÞX n ;>
nn1 nn1
d R L n d R R I R L n 1
ð5:22Þ
• Initial condition: R1 ðLðn0 ÞÞ with n0 chosen to be the smallest number of data
lines that satisfy n0 N L to avoid the singularity in calculating the matrix
inverse where L is the total number of spectral bands.
Note that, unlike δRT ‐ CEM(rn) in (5.13), which processes data rn sample by
sample, δRT -CEM ðX Þ in (5.21) and δRT - CEM (X ) in (5.22) process data line by data
n1 n n n
line in the sense that they will not process data line Xn until the nth line data matrix
T 1
is completed. Thus, initial conditions must start with δRT -CEM ðX Þ ¼ d R ðL1 ÞX1 and
1 1 dT R1 ðL1 Þd
-CEM ðX Þ dT R1 ðL1 ÞX2
δRT
1 2 ¼ .
That is, we need to wait for the first line to come in and
dT R1 ðL1 Þd
1
be processed to calculate R ðL1 Þ prior to δ1RT - CEM (X1) and then further wait for the
second line to be processed to calculate δ1RT - CEM (X2). This is because both
δRT - CEM (X ) and δRT - CEM (X ) use the same initial sample correlation matrix R1
1 1 1 2
ðL1 Þ to complete their operations. As a consequence, for n 2, (5.22) is repeatedly
implemented recursively by
5.6 Real Image Experiments 137
Two real hyperspectral image scenes were used for experiments to conduct a
performance evaluation of anomaly detection.
The image scene shown in Fig. 5.5 (and Fig. 1.9a) was used for experiments. It was
acquired by the airborne HYperspectral Digital Imagery Collection Experiment
(HYDICE). It has a size of 64 64 pixel vectors with 15 panels in the scene and the
ground truth map in Fig. 5.5b (Fig. 1.9b), where the ith panel signature, denoted by
pi, was generated by averaging the red panel center pixels in row i, as shown in
Fig. 5.5c (and Fig. 1.10) These panel signatures will be used to represent target
knowledge of the panels in each row.
The panel signatures in Fig. 5.5c were used as prior target knowledge required
by CEM for the desired target signatures. Figure 5.6 shows five detection maps
produced by CEM using the five panel signatures p1, p2, p3, p4, and p5 in Fig. 5.5c
as the desired target signatures.
In accordance with the ground truth, the panels in rows 2 and 3 were made by the
same materials with slightly different colors of paint, a light olive parachute and a
dark olive parachute. As a consequence, detecting panels in row 2 would also be
able to detect panels in row 3 and vice versa. This fact was confirmed in Fig. 5.7b, c
and is reproduced in Fig. 5.6 in decibel (db). Similarly, it was also applied to panels
in rows 4 and 5, which were also made by the same materials with slightly different
paint colors than in Fig. 5.7d, e.
Using the results of Fig. 5.6 for a benchmark comparison, RT-CEM was
implemented in two formats, BIP/BIS and BIL, using the same five panel signatures
p1, p2, p3, p4, and p5 in Fig. 5.5c as desired target signatures. Figures 5.8, 5.9, 5.10,
5.11, and 5.12 show color sample-varying CEM-detection maps in db of panel
138 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
RT CEM
d (n) (-1)(.)-1 n (Xn )
Xn D Xn-1
I
R-1(L(n-1))
a b
c
7000
P1
6000 P2
P3
P4
5000 P5
Radiance
4000
3000
2000
1000
0
0 20 40 60 80 100 120 140 160 180
Band
Fig. 5.5 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels; (c) spectra of p1, p2, p3, p4, and p5
Fig. 5.6 Detection maps of CEM using (a) p1, (b) p2, (c) p3, (d) p4, and (e) p5 as desired target
signature; (a) row 1, (b) row 2, (c) row 3, (d) row 4, and (e) row 5
5.6 Real Image Experiments 141
Fig. 5.7 Detection maps in db of CEM using (a) p1, (b) p2, (c) p3, (d) p4, and (e) p5 as desired
target signature. (a) Panel pixels in row 1, (b) panel pixels in row 2, (c) panel pixels in row 3, (d)
panel pixels in row 4, and (e) panel pixels in row 5
Fig. 5.8 Detection maps in db of RT-CEM using BIP/BIS with p1 used as desired target signature
Fig. 5.9 Detection maps in db of RT-CEM using BIP/BIS with p2 used as desired target signature
142 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
Fig. 5.10 Detection maps in db of RT-CEM using BIP/BIS with p3 used as desired target
signature
Fig. 5.11 Detection maps in db of RT-CEM using BIP/BIS with p4 used as desired target
signature
pixels in each of the five rows produced by RT-CEM as processed in real time
according to the BIP/BIS format. As in Fig. 5.7, panels in rows 2 and 3 were
detected by Figs. 5.9 and 5.10 and panels in rows 4 and 5 were also detected in
Figs. 5.11 and 5.12. Most importantly, the color sample varying CEM-detection
maps produced by RT-CEM using the BIP/BIS format yielded an intriguing
finding. Initially, there were no pixels matched with the desired signature d, the
detection maps showed completely random abundance fractions in Figs. 5.8a–5.12a
until the first panel pixel that matched the d was detected, in which case, the
5.6 Real Image Experiments 143
Fig. 5.12 Detection maps in db of RT-CEM using BIP/BIS with p5 used as desired target
signature
background was suddenly suppressed in Figs. 5.8b–5.12b and the detected panel
pixels stood out as bright pixels with very high-intensity values in Figs. 5.8c–5.12c.
Then the background was continuously suppressed until the process was completed
as shown in Figs. 5.8d, f to 5.12d, f. Of particular interest are detection maps of
panels in row 3 in Fig. 5.12c–e and detection maps of panels in row 5 in Fig. 5.12c–
e, where the background was not supposed to be suppressed until the process
reached the matching panel pixels in Figs. 5.10e and 5.12e. However, as noted,
the signatures of panels in rows 2 and 3 are very similar due to the fact that they
were made by the same parachutes with light and dark olive paints, respectively.
Similarly, this is also true of panels in rows 4 and 5. Thus, when RT-CEM
processed the panels in row 2 in Fig. 5.10c and panels in row 4 in Fig. 5.12c,
their signatures were close to the matching signatures p3 and p5, respectively. As a
consequence, these panels were picked up by RT-CEM. When RT-CEM reached
the panels in row 3 in Fig. 5.10d and the panels in Fig. 5.12e that matched the
desired signatures, the background was further suppressed once again. Obviously,
such a phenomenon cannot be provided by Fig. 5.7, which represents results
obtained by implementing CEM in a one-shot operation, that is, the use of global
correlation matrix R, not causal correlation matrix R(n). Comparing Figs. 5.8, 5.9,
5.10, 5.11, and 5.12 to Fig. 5.7, it is clear that RT-CEM seemed to have better
background suppression than CEM did in Fig. 5.7. Nevertheless, final results
produced by RT-CEM using BIP/BIS were comparable to those produced by
CEM, with little appreciable difference between them.
Following similar experiments performed by RT-CEM using BIP/BIS,
Figs. 5.13, 5.14, 5.15, 5.16, and 5.17 show the time-varying progressive detection
maps of panels in five rows produced by RT-CEM according to the BIL format,
144 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
Fig. 5.13 Detection maps in db of RT-CEM using BIL with p1 used as desired target signature
Fig. 5.14 Detection maps in db of RT-CEM using BIL with p2 used as desired target signature
where the detected amounts of panel pixels are plotted in db for better visual
assessment.
Experimental results are the same as those obtained by BIP/BIS, except for
several very interesting findings.
1. A significant improvement in the detection of panels in row 1 compared to
Figs. 5.8 and 5.13, where the three panel pixels in row 1 were clearly detected in
Fig. 5.13, where the subpixel panel p13 in row 1 were barely detected in Fig. 5.8
and were invisible in Fig. 5.15;
5.6 Real Image Experiments 145
Fig. 5.15 Detection maps in db of RT-CEM using BIL with p3 used as desired target signature
Fig. 5.16 Detection maps in db of RT-CEM using BIL with p4 used as desired target signature
2. In detecting panels in row 4, RT-CEM using BIL also detected panels in row
1 prior to their detection. Interestingly, once the panels in row 4 were detected,
the detected panels in the row faded away immediately, and eventually small
amounts of abundance fractions of panels in row 1 were picked up in the final
detection map. Based on the spectral signature analysis conducted in Chap. 2 in
Chang (2003a, b), the spectral signature of p1 is indeed very similar to that of p4
and p5. Thus, before panels in row 4 was detected, the panels in row 1 were
considered to be those close to p4, in which case they were actually picked up as
146 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
Fig. 5.17 Detection maps in db of RT-CEM using BIL with p5 used as desired target signature
if they were panels in row 4. However, once p411 was detected, the detected
abundance fractions of three panel pixels, p11, p12, and p13, were quickly
diminished. Interestingly, this phenomenon was not observed in Figs. 5.7d and
5.16;
3. Similarly, the foregoing observations also applied to the detection of panels in
row 5. These experiments demonstrated that RT-CEM using BIL did have an
advantage in terms of finding similar signatures while proceeding line by line;
4. As noted, BIL was implemented as a compromise between RT-CEM using R(n)
according to the BIP/BIS format and CEM using the global correlation matrix R.
It turned out that it did have some advantages, such as those mentioned earlier,
and computational savings, to be described in the following section.
Since we have the ground truth of the 19 R panel pixels, we can further calculate
detection rates of RT-CEM using BIP/BIS and BIL formats via a receiver operating
characteristic (ROC) analysis for performance evaluation, that is, the detection of
panel pixels p11, p12, and p13 in row 1 using p1, the detection of panel pixels p211,
p221, p22, and p23 in row 2 using panel signature p2, the detection of panel pixels
p311, p312, p32, and p33 in row 3 using panel signature p3, the detection of panel
pixels p411, p412, p42, and p43 in row 4 using panel signature p4, and the detection
of panel pixels p511, p521, p52, and p53 in row 5 using panel signature p5. To
include the number of pixels and data lines as a parameter to see the various
detection performances of RT-CEM on these 19 R panel pixels sample by sample
and line by line, we do not follow the traditional approach to plotting an ROC
curve of detection probability, PD, versus false alarm probability, PF. Instead, we
choose the common practice used in medical diagnosis to calculate the area under
an ROC curve, referred to as area under curve (AUC) (Metz 1978). Using such an
AUC, an alternative ROC curve can be plotted in terms of AUC on the y-axis
5.6 Real Image Experiments 147
a b
1 1
0.98 BIP BIP
BIL 0.98 BIL
0.96
0.94 0.96
AUC
AUC
0.92 0.94
0.9
0.92
0.88
0.9
0.86
0.84 0.88
0 5 19 32 45 58 64 0 5 19 32 45 58 64
data line being processed data line being processed
c d
1 1
BIP BIP
0.9996 BIL BIL
0.9995
0.9994
0.999
0.9992
AUC
AUC
0.999 0.9985
0.9988
0.998
0.9986
0.9975
0.9984
0.997
0 5 19 32 45 58 64 0 5 19 32 45 58 64
data line being processed data line being processed
e
1
BIP
0.9999
BIL
0.9998
0.9997
0.9996
AUC
0.9995
0.9994
0.9993
0.9992
0.9991
0 5 19 32 45 58 64
data line being processed
Fig. 5.18 AUC values of 19 R panel pixels detected by RT-CEM: (a) panel pixels in row 1;
(b) panel pixels in row 2; (c) panel pixels in row 3; (d) panel pixels in row 4; and (e) panel pixels in
row 5
versus pixels or data lines on the x-axis. Figure 5.18 plots the values of AUC for
real-time performance of RT-CEM in detection of the 19 R panel pixels using five
panel signatures in Fig. 5.5c as desired target signatures where data lines instead
of pixels are used as a parameter on the x-axis to be able to conduct a comparative
analysis between BIP/BIS and BIL, and plots were not started until the data line
148 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
that contains panel pixels to be detected was reached. For example, for detecting
panel pixels in row 5, the plot in Fig. 5.18e was not started until it reached the 58th
data line.
As demonstrated in Fig. 5.18, it is interesting to note that RT-CEM using
BIP/BIS produced higher values of AUC than did RT-CEM using BIL for the
detection of panels in rows 2 and 4. However, the conclusion is reversed for the
detection of panels in rows 1, 3, and 5. Moreover, there were significant discrep-
ancies between RT-CEM using BIP/BIS and BIL in detecting panels in rows 1 and
2 compared to small differences in detecting panels in rows 3–5. This is mainly
due to the fact that the spectral signatures in rows 1 and 2 are closer to those in
rows 3–5 according to analysis (Chang 2003a, b, Chap. 2) where sample-varying
RT-CEM could perform better detection than data line-varying RT-CEM did. In
addition, comparison of Figs. 5.17 and 5.12 makes it very clear that RT-CEM
using BIP outperformed BIL in terms of false alarms prior to the detection of the
first panel pixel in row 5 at the 58th data line since RT-CEM using BIP did not
pick up the panel pixels in row 1 as RT-CEM using BIL did. However, this fact
was not reflected in the ROC curve analysis in Fig. 5.18e, where RT-CEM using
BIL performed slightly better than RT-CEM using BIP/BIS. Thus, visual inspec-
tion of detection maps in Figs. 5.12 and 5.17 reveals that RT-CEM using BIP/BIS
performed better than RT-CEM using BIL. On the other hand, examination of the
ROC analysis in Fig. 5.18e leads to the opposite conclusion. Similarly, compar-
ison of Figs. 5.15 and 5.10 showed the same phenomenon in Fig. 5.18c. These
simple examples demonstrated that without implementing CEM in real time, an
incorrect final conclusion could be drawn if one looked only at the final results
produced by CEM. This is because ROC analysis is performed solely based on the
performance of detection probability relative to false alarm probability, not on
both individually. In other words, ROC analysis does not deal with the back-
ground suppression issue (Wang et al. 2013a, b, c). To address this issue, a 3D
ROC analysis recently developed by Chang (2010) and Chang (2013, Chap. 3) can
be used for this purpose.
One major advantage of implementing RT-CEM is that it makes use of the most
recent incoming data samples, a pixel from BIP/BIS and a line from BIL, to update
currently available information via recursive equations without reprocessing past
information. This can reduce costs by minimizing the memory required for
processing very large hyperspectral image cubes. Specifically, a typical image
cube is near 1 GB in size, while RT-CEM inverting a correlation matrix combined
with newly received data only needs less than 100 KB. Figure 5.19 plots the
processing times required for RT-CEM using BIP/BIS and BIL to complete the
HYDICE scene where the y-axis is the processing time in seconds and the x-axis is
the value of n corresponding to the data line, Xn, being processed. As we can see
5.6 Real Image Experiments 149
Time (sec)
0.015
0.01
0.005
0
0 10 20 30 40 50 60 64
after initial conditions, the processing times for both formats are nearly constant. As
also shown in the plots, the processing time required for RT-CEM using BIL is
much less than that required for RT-CEM using BIP/BIS. This is due to the fact that
the processing time for RT-CEM using BIP/BIS must include time to process each
individual pixels compared to RT-CEM using BIL, which only processes one data
line at a time, in which case it only needs to invert one single correlation matrix to
update the current information.
Despite the fact that RT-CEM has advantages over CEM, RT-CEM does have
some disadvantages as well. The primary disadvantage is that, with the exception of
the pixels/data lines used to generate the initial inverse correlation matrix, each
pixel or line must be processed using a different and new causal sample correlation
matrix. As a result, it increases the computational complexity. Nevertheless,
RT-CEM has been shown to have significant advantages over CEM in many
respects. Specifically, the ability of RT-CEM to process data in real time without
waiting for a complete data set opens up opportunities for applications such as
moving target detection, an advantage not be offered by the typical CEM algorithm.
800
5
600
400
1
200
0
0 20 40 60 80 100 120 140 160
Fig. 5.20 Five desired targets and two anomalies shown in (a) with corresponding spectra of
signatures in (b)
Fig. 5.21 Detection maps in db of dry lake by RT-CEM using BIP/BIS. (a) BIP/BIS. (b) BIL
5.6 Real Image Experiments 151
Fig. 5.22 Detection maps in db of cinders by RT-CEM using BIP/BIS. (a) BIP/BIS. (b) BIL
5.24, 5.25, and 5.26 show real-time CEM-detection maps for five signatures: dry
lake, cinders, vegetation, shade, and rhyolite plus anomaly by RT-CEM using
BIP/BIS and BIL, where the desired signatures were obtained by averaging pixels
within a 3 3 window according to the ground truth shown in Fig. 5.20a, b. For
better visual assessment the results in Figs. 5.21, 5.22, 5.23, 5.24, 5.25, and 5.26 are
presented in db so that the background suppression can be brought up more visibly
for inspection.
As we can see from these progressive real-time detection maps, RT-CEM using
BIL always had better background suppression than RT-CEM using BIP/BIS at
various stages of detection when areas are large such as cinders, shade, rhyolite, and
the middle stages of dry lake. Of particular interest is Fig. 5.23, which shows that
152 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
Fig. 5.23 Detection maps in db of vegetation by RT-CEM using BIP/BIS. (a) BIP/BIS. (b) BIL
Fig. 5.24 Detection maps in db of shade by RT-CEM using BIP/BIS. (a) BIP/BIS. (b) BIL
Finally, Fig. 5.27 also plots the computer processing time required by RT-CEM
using BIP/BIS and BIL. As in Fig. 5.19, RT-CEM using BIL required much less
time than RT-CEM using BIP/BIS, as expected.
154 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
Fig. 5.25 Detection maps in db of rhyolite by RT-CEM using BIP/BIS. (a) BIP/BIS. (b) BIL
5.6 Real Image Experiments 155
Fig. 5.26 Detection maps in db of anomaly by RT-CEM using BIP/BIS. (a) BIP/BIS. (b) BIL
156 5 Real-Time RHSP for Active Target Detection: Constrained Energy Minimization
0.05
0.045 BIP
BIL
0.04
0.035
0.03
Time (sec)
0.025
0.02
0.015
0.01
0.005
0
0 50 100 150 200
data line being processed
Fig. 5.27 comparison between computer processing time required by RT-CEM using BIP/BIS
and BIL for LCVF image scene
5.7 Conclusions
This chapter developed a real-time processing of CEM with several new ideas. First
and foremost was the introduction of a new concept of causality into CEM so that
RT-CEM can be implemented in real time. Second, a new real-time CEM using
BIP/BIS via a causal sample-varying correlation matrix for performing sample-by-
sample subpixel detection was developed. Third, a new real-time CEM using BIL
via a causal line-varying correlation matrix for performing subpixel detection line
by line was also developed, where the derivations for RT-CEM using BIL were
completely new and only recently reported in Chang et al. (2015d). It is by no
means trivial, but instead, much more involved than that deriving RT-CEM using
BIP/BIS, as shown in its computational complexity in Sect. 5.4. Finally, for
RT-CEM to deal with nonstationarity and improve its computational complexity,
RT-CEM is further extended to recursive RT-CEM, where recursive equations are
derived in a fashion similar to how a Kalman filter is carried out for RT-CEM using
both BIP/BIS and BIL. Accordingly, only the most recently received data sample
vector will be used to update results recursively sample by sample and line by line
similarly to how a Kalman filter is carried out. As a result, RT-CEM can extract
targets of interest during its ongoing processing on a timely basis.
Chapter 6
Real-Time Recursive Hyperspectral Sample
Processing for Passive Target
Detection: Anomaly Detection
6.1 Introduction
processing does not have to be done in real time. On the other hand, real-time
processing must be causal. Unfortunately, this causal issue has never been
addressed by many reported real-time processing algorithms. Following an approach
similar to that treated in Chap. 5, two commonly used anomaly detectors, K-AD and
R-AD, can be first extended to their causal versions, CK-AD and CR-AD, which can
process data in a causal manner sample by sample. Then CK-AD and CR-AD are
further extended to their respective recursive versions, referred to as recursive
CK-AD/CR-AD (R-CK-AD/R-CR-AD), which can update data information via
recursive equations without reprocessing all past data sample vectors as the causal
sample correlation/covariance matrix (CSCRM/CSCVM) varies with the data sample
vectors being processed. Since R-CK-AD and R-CR-AD are eventually realized by
real-time processing, they are actually real-time CK-AD (RT-CK-AD) and real-time
CR-AD (RT-CR-AD).
One of the most widely used anomaly detectors is the algorithm developed by Reed
and Yu, the K-Anomaly Detector (K-AD), which makes use of the global sample
covariance matrix, K, to account for spectral statistics among data sample vectors.
Since its introduction, many different K-AD-like anomaly detectors have been
proposed. Among them, of particular interest is causal AD (CR-AD) developed in
Chang and Chiang (2002) and Chang (2003a).
In what follows, we briefly describe these two anomaly detectors. Assume that
N
fri gi¼1 , where N is the total number of entire data sample vectors in the data and
ri ¼ ðr i1 ; r i2 ; ; r iL ÞT is the ith data sample vector, where L is the total number of
spectral bands.
6.2.1 K-AD/R-AD
The K-AD, denoted by δK-AD(r), is also known as RXD, developed by Reed and Yu
(1990), and is given by
where μ is the sample mean and K is the sample data covariance matrix. The form
of δK-AD(r) in (6.1) is actually the well-known Mahalanobis distance. However,
from a detection point of view, the use of K1 can be interpreted as a whitening
process to suppress image background. By replacing the global covariance matrix
K and (r μ) with the global correlation matrix R and the data sample vector r,
6.2 Anomaly Detection 161
Chang and Chiang (2002) further modified K-AD by developing a new correlation
matrix, an R-based anomaly detector, referred to as R-AD, which is given by
where the correlation matrix R is used to handle first- and second-order statistics by
including the sample mean μ, compared to K-AD, which uses the covariance matrix
K to account for only second-order statistics.
N
Let fri gi¼1 be a set of data sample vectors to be processed. The causal R-AD
(CR-AD), denoted by δCR-AD(r), is specified by
n
Assume that fri gi¼1 is a set of input data sample vectors, where ri ¼ ðr i1 ; r i2 ; ;
T
r iL Þ is the ith data sample vector and L is the total number of spectral bands.
Suppose that rn is the most recent input data sample vector and μðnÞ ¼ ð1=nÞ
Xn n
r is the sample mean of fri gi¼1
i¼1 i
. Then CSCRM and CSCVM formed by
n
Xn
fri gi¼1 , R(n), and K(n) are defined by RðnÞ ¼ ð1=nÞ i¼1 ri riT and KðnÞ ¼ ð1=nÞ
Xn T
i¼1
ðri μðnÞÞðri μðnÞÞT respectively and the relationship between R(n) and
K(n) can be derived as follows:
Xn
KðnÞ ¼ ð1=nÞ i¼1 ðri μðnÞÞðri μðnÞÞT
Xn Xn Xn
¼ ð1=nÞ i¼1 ri riT ð1=nÞ i¼1 ri μT ðnÞ ð1=nÞ i¼1 riT μðnÞ
Xn
þ μðnÞμT ðnÞ þ ð1=nÞ i¼1 μðnÞμT ðnÞ, ¼ RðnÞ
Xn Xn ð6:5Þ
ð1=nÞ i¼1 riT μðnÞ ð1=nÞ i¼1 r μT ðnÞ
The form specified by (6.5) for R(n) and K(n) can be expressed by separating the
most recent input data sample vector rn from its previous input data sample vectors,
n
fri gi¼1 , to derive a form to be used for real-time processing as follows:
Xn Xn1
RðnÞ ¼ ð1=nÞ i¼1 ri riT ¼ ð1=nÞ i¼1 ri riT þ ð1=nÞrn rnT
Xn1
¼ ððn 1Þ=nÞð1=ðn 1ÞÞ i¼1 ri riT þ ð1=nÞrn rnT ð6:6Þ
and
Xn
KðnÞ ¼ ð1=nÞ ðri μðnÞÞðri μðnÞÞT
i¼1
ð6:7Þ
¼ ððn 1Þ=nÞKðn 1Þ þ ðrn μðnÞÞðrn μðnÞÞT ;
respectively. Xn Xn
Since μðnÞ ¼ ð1=nÞ i¼1 ri and KðnÞ ¼ ð1=nÞ i¼1 ðri μðnÞÞðri μðnÞÞT ,
we can derive
Xn
KðnÞ ¼ ð1=nÞ i¼1 ðri μðnÞÞðri μðnÞÞT
Xn Xn Xn
¼ ð1=nÞ i¼1 ri riT ð1=nÞ i¼1 ri μT ðnÞ ð1=nÞ i¼1 riT μðnÞ
Xn
þ μðnÞμT ðnÞ þ ð1=nÞ i¼1 μðnÞμT ðnÞ
Xn Xn
¼ RðnÞ ð1=nÞ i¼1 riT μðnÞ ð1=nÞ i¼1 r μT ðnÞ þ μðnÞμT ðnÞ
Thus,
eX
R ð nÞ ¼ X eT: ð6:11Þ
T
KðnÞ ¼ XX : ð6:12Þ
Xn pffiffiffi
e ðnÞ ¼ ð1=nÞ
Now, if we further define μ e
r ¼ ð1= nÞμðnÞ as the mean of data
i¼1 i
sample vectors fe n
r i gi¼1 , then ri ¼ ri μn
ðnÞ
pffiffi ¼ priffiffi
n
μpðnffiffinÞ ¼ e e ðnÞ and
ri μ
Xn Xn
KðnÞ ¼ ð1=nÞ i¼1 ðri μðnÞÞðri μðnÞÞT ¼ i¼1
ðe e ðnÞÞðe
ri μ e ðnÞÞT
ri μ
¼ RðnÞ μðnÞμT ðnÞ ¼ RðnÞ ðnÞe
μ ðnÞe
μ T ðnÞ:
ð6:13Þ
is very helpful.
1
Since R1 ðnÞ ¼ ððn 1Þ=nÞRðn 1Þ þ ð1=nÞrn rnT , from (6.10), we can
calculate R1(n) via (6.14) by setting A ¼ ððn 1Þ=nÞRðn 1Þ and
pffiffiffi
u ¼ v ¼ ð1= nÞrn as follows:
h pffiffiffiffiffiffiffiffiffiffiffi ihpffiffiffiffiffiffiffiffiffiffiffi i
fð1 1=nÞKðn 1Þg1 n 1=n ðrn μðn 1ÞÞ n 1=n ðrn μðn 1ÞÞT fð1 1=nÞKðn 1Þg1
hpffiffiffiffiffiffiffiffiffiffiffi pffiffiffiffiffiffiffiffiffiffiffi i
1þ n 1=n ðrn μðn 1ÞÞT fð1 1=nÞKðn 1Þg1 n 1=n ðrn μðn 1ÞÞ
ð6:17Þ
166 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
Assume that M is a matrix of size L p and can be written M ¼ ½U d, with U and
d being anL ðp 1Þ matrix and an L-dimensional vector. Then, according to a matrix
inverse identity derived in Settle (1996), the inverse of MTM can be expressed as
" #1 " 1 T #
T 1 UT U UT d UT U þ βU# ddT U# βU# d
M M ¼¼ ¼ T ;
dT U dT d βdT U# β
ð6:18Þ
1 n h 1 i o1
1
where U# ¼ UT U U and β ¼ dT I U UT U UT d ¼ dT P ⊥
U d .
Now, if we let U ¼ ½r1 rn2 rn1 ¼ Xn1
T
and d ¼ rn, then
M ¼ ½Ud ¼ ½r1 r2 rn1 rn ¼ Xn is an L n matrix and MTM becomes an L
L matrix with its inverse calculated by (6.18) as follows:
" #1
1 1 1
T
Xn1 Xn1 Xn1 rn
T
M M ¼ Xn XnT ¼ nR ðnÞ ¼
rnT Xn1 rnT rn
" 1 T #
T
Xn1 Xn1 þ βX#n1 ddT X#n1 βX#n1 rn
¼ T
βrnT X#n1 β
" # T #
1
ðn 1ÞR ðn 1Þ þ βXn1 dd Xn1
# T
βX#n1 rn
¼ T ;
βrnT X#n1 β
ð6:19Þ
1 T n h i o1
with X#n1 ¼ Xn1 Xn1
T
Xn1 and β ¼ rnT P⊥
X T rn .
n1
e¼e
Using (6.11) with d e n ¼ ½e
r i ¼ priffiffin and X r 1e
r 2 e
r n , (6.19) can be reexpressed as
" #1
h i e n1 X eT e rn
eT ¼
e nX X n1 X n1 e
R1 ðnÞ ¼ X n
e e n1
r nT X e
r nT e
rn
2 1 T 3
e n1 X
X eT þ β Xe# d edeT X e# β e# e
X r
6 n1 n1 n1 n1 n
7
¼4 T 5
βe r nT X e# β
n1
2 T 3
ðn 1Þ=nR1 ðn 1Þ þ βX e# d edeT Xe# β e# e
X r
6 n1 n1 n1 n
7
¼4 T 5;
βe r nT Xe# β
n1
ð6:20Þ
1
1
e# ¼ X
where X eT
e n1 X eT
X r n Pe T e
and β ¼ e T ⊥
rn .
n1 n1 n1
X n1
6.4 Real-Time Causal Anomaly Detection 167
Two types of real-time anomaly detectors are derived in this section, real-time
causal R-AD, which uses CSCRM, R(n), to account for the spectral correlation
characterized by the first two orders of spectral statistics correlation, and real-time
causal K-AD, which uses K(n) to account for the second-order statistics spectral
correlation.
From (6.10) the innovation information is provided by the pixel currently being
processed, rn, and R1(n 1) is the processed information obtained by previous
n 1 data sample vectors fri gn1
i¼1 . By means of (6.15) we can derive
e# e
þ2βrnT X n1 r n þ βrn rn :
T
168 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
Fig. 6.1 Flowchart showing implementation of recursive equation (6.7) to find R1(n)
Unlike the C-R-AD, which uses the autocorrelation matrix without calculating the
sample mean, finding a real-time causal version of K-AD is not trivial because it
requires knowing the causal sample mean prior to calculating the causal covariance
matrix. In this case, we reexpress
rn μðnÞ
Xn1
¼ rn ððn 1Þ=nÞð1=ðn 1ÞÞ ri þ ð1=nÞrn ð6:23Þ
i¼1
¼ rn ð1 1=nÞμðn 1Þ þ ð1=nÞrn :
Using (6.23) in conjunction with (6.17), we can further derive a real-time causal
version of K-AD (RT-CK-AD) in (6.24) as follows, which can be easily updated by
K1(n 1), μ(n 1) as well as the current input data sample vector rn, where no
inverse need be calculated once the initial calculation of K1(L ) is done:
6.4 Real-Time Causal Anomaly Detection 169
ðrn μðnÞÞ
¼ ð1 1=nÞ1 ðrn μðnÞÞT ½Kðn 1Þ1 ðrn μðnÞÞT
ð rn μ ð nÞ Þ T
pffiffiffiffiffiffiffiffiffiffiffi 2 h ih i
ð1 1=nÞ2 n 1=n fKðn 1Þg1 ðμðn 1Þ rn Þ ðμðn 1Þ rn ÞT fKðn 1Þg1
pffiffiffiffiffiffiffiffiffiffiffi 2 h i
1 þ ð1 1=nÞ1 n 1=n ðμðn 1Þ rn ÞT fKðn 1Þg1 ðμðn 1Þ rn Þ
ðrn μðnÞÞ:
ð6:24Þ
It should also be noted that there is no similar derivation of (6.21) and (6.24)
derived in Du and Nekovei (2009) since all the classifiers (Du and Nekovei 2009)
used the sample correlation matrix R, not the sample covariance matrix K. Also,
comparison of (6.24) and (6.21) shows that it is more complicated to calculate a
causal form of RT-CK-AD than of RT-CR-AD.
This section provides a detailed analysis of the complexity of calculating (6.21) and
(6.24). Since the computational complexity of (6.24) is the same as that of (6.21),
only (6.21) is discussed.
First, to avoid a singularity problem, the initial condition of calculating R1(n)
in (6.21) should involve collecting a sufficient number of pixels to ensure the
sample correlation/covariance matrix is of full rank. In this case, the causal
processing must begin with the Lth pixel, rL where L is the total number of spectral
bands. Also, according to (6.21), the R1(n) in the RT-CR-AD, δRT-CR-AD(rn), can
be easily updated using previously processed information R1(n 1) and the
innovation information provided by rn without actually calculating the matrix
inverse once the initial calculation of R1(L) is done. In other words, we only
need to calculate a matrix inverse once, which is R1(L ). However, in causal
processing the R1(n) must be recalculated for each incoming pixel rn. This may
lead to a belief that its computational complexity may be much higher than that
required for real-time causal processing using (6.21). As a matter of fact, it is not. In
causal processing there are three major operations involved for each data sample
170 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
vector. One is to calculate the outer product of the incoming pixel rn, which is rnrTn ,
to update R1(n 1) to R1(n), which requires L2 multiplications. Another is to
calculate the inverse of R1(n) with size L L. The third one is to calculate rnT
½RðnÞ1 rn , which requires (L + 1) inner products, each of which requires
L multiplications. As a result, the computational complexity of finding R1(n) is
the same as that required for R1(L ) and is constant for all n.
On the other hand, the real-time causal processing using (6.21) requires three
operations for calculating xTAy (i.e., two calculations of rnT ½Rðn 1Þ1 rn and one
n on o
calculation of rnT ½Rðn 1Þ1 rn rnT ½Rðn 1Þ1 rn), each of which requires L
+ 1 innernproducts, whichoresults
n in L(L + 1)o multiplications as well as one outer
product ½Rðn 1Þ1 rn rnT ½Rðn 1Þ1 , which requires L2 multiplications.
More specifically, at the nth data sample vector rn with n L þ 1, the calculation of
R1(n) requires calculations of L-dimensional inner products, that is, ½Rðn 1Þ1
rn and rnT ½Rðn 1Þ1 rn , which result in L2 multiplications and one outer product
n on o
Oðn 1Þ ¼ ½Rðn 1Þ1 rn rnT ½Rðn 1Þ1 , which carries out L2 multiplica-
tions plus another L-dimensional inner products to calculate rnT Oðn 1Þrn , which
amounts to L multiplications. As a result, a total of two L-dimensional scalar inner
products and one matrix outer product are required and give rise to L2 + L multi-
plications, which is exactly the same as that derived in Du and Nevovei (2009). But
the entire complexity to carry out a real-time causal anomaly detector in (6.21) is
actually 3(L2 + L) + L2 multiplications plus the complexity of calculating the initial
condition, R1 ðLÞ=K1 ðLÞ.
According to the foregoing analysis, the computational complexity of causal and
real-time anomaly detectors is determined by the computer processing time (CPT)
required by calculating three elements, an inverse of an L L matrix, an inner
product of two L-dimensional vectors, and an outer product of two L-dimensional
vectors. Most importantly, in both causal processing and real-time causal
processing the computational complexity is independent of the data sample vector
to be processed and is linearly increased with the number of data sample vectors. A
detailed study of computational complexity is provided in Table 6.1.
Table 6.1 tabulates the computational complexity of causal and real-time causal
versions of R-AD and K-AD in terms of the required number of multiplications “”
used for calculation, where in the second column under causal anomaly detectors an
additional number of nL2 multiplications is required to calculate the sample corre-
lation/covariance matrix R(n)/K(n) plus processing time of the matrix inversion of
R(n)/K(n) as n varies, neither of which is needed if a real-time causal algorithm in
the third column is used via (6.21) or (6.24).
To conclude this section, it is worth noting that the CK-AD and CR-AD can be
indeed implemented without real-time processing. In other words, the sample
covariance matrix K(n) and sample correlation matrix R(n) can be recalculated
without using the causal innovation update equations (6.21) and (6.24) for updating
6.5 Synthetic Image Experiments 171
Table 6.1 Comparative analysis of computational complexity between causal and real-time
causal versions of R-AD and K-AD
Computational
complexity CR-AD/CK-AD RT-CR-AD/RT-CK-AD
Initial condition N/A R1(L )/K1(L )
Input rn R1(n)/K1(n) for each input N/A
sample rn
Number of inner L+1 3(L + 1)
products/sample
Number of outer L (one for (rn(rn)T) L (one for
n on oT
products /sample
( ½Rðn 1Þ1 rn r½Rðn 1Þ1 rn )
Number of “”/ [(L + 1)L + L2 ¼ 2 L2 + L + c]/2 [3(L + 1)L + L2 ¼ 4 L2 + 3 L + c]/2
sample
Number of “”/ (N L )(2 L2 + L + c)/2 (N L )(4 L2 + 3 L + c)/2
data set
Complexity O(N ) O(N )
innovations information. As a result of not using (6.21) and (6.24), the following
experiments demonstrate that the computer processing time for CK-AD and
CR-AD is linearly increased with the number of pixels to be processed compared
to the computer processing time remaining constant when their real-time counter-
parts, RT-CK-AD and RT-CR-AD, are used.
The synthetic image simulated in Fig. 1.15 is shown in Fig. 6.2 with five panels in
each row simulated by the same mineral signature and five panels in each column
having the same size.
Among the 25 panels are five 4 4 pure pixel panels for each row in the first
column and five 2 2 pure pixel panels for each row in the second column, five
2 2 mixed pixel panels for each row in the third column, and five 1 1 subpixel
panels for each row in both the fourth and fifth columns, where the mixed and
subpanel pixels were simulated according to the legends in Fig. 6.2. Thus, a total of
100 pure pixels (80 in the first column and 20 in the second column), referred to as
endmember pixels, were simulated in the data by the five endmembers A, B, C, K,
and M. An area marked “BKG” in the upper right corner of Fig. 1.14a was selected
to find its sample mean, that is, the average of all pixel vectors within the “BKG”
area, denoted by b and plotted in Fig. 1.14b, to be used to simulate the background
(BKG) for the image scene with a size of 200 200 pixels in Fig. 6.2. The reason
for this background selection is empirical since the selected “BKG” area seemed
more homogeneous than other regions. Nevertheless, other areas could also be
selected for the same purpose. This b-simulated image background was further
172 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
The first type of target insertion is TI, in which the previously mentioned 130 panel
pixels are inserted into the image by replacing their corresponding background
pixels. Thus, the resulting synthetic image has clean panel pixels implanted in a
noisy background with an additive Gaussian noise of SNR ¼ 20:1 for this scenario,
as shown in Fig. 6.3a. Figure 6.3b–g shows the results of traditional R-AD and
K-AD, along with their causal and real-time causal versions in terms of detected
abundance fractions, where the value of x is represented by decibel (db), that is,
20 log10x with x being the detected abundance fractions to enhance visual assess-
ment. As we see, all versions of anomaly detector performed very comparably,
except for the different degrees of background suppression resulting from the use of
global and causal correlation matrices.
Since CR-AD and RT-CR-AD use the same updating autocorrelation matrix R
(n), they both produce exactly the same detection performance with the only
difference in CPT resulting from whether or not the causal innovation information
update equation (6.22) is implemented. Similarly, this is also true of CK-AD and
6.5 Synthetic Image Experiments 173
Fig. 6.3 Detection results for TI with detected abundance fractions in db: (a) 50th band of TI
scenario, (b) K-AD, (c) CK-AD, (d) RT-CK-AD, (e) R-AD, (f) CR-AD, and (g) RT-CR-AD
Fig. 6.4 CK-AD detection results for TI with detected abundance fractions in db: (a) no panels
detected, (b) row 1 panels detected, (c) row 2 panels detected, (d) row 3 panels detected, (e) row
4 panels detected, and (f) row 5 panels detected
RT-CK-AD. Thus, Figs. 6.4, 6.5, 6.6, and 6.7 show the detection results of the
correlation matrix K-based real-time causal anomaly detectors CK-AD and RT-
CK-AD and the covariance matrix R-based anomaly detectors, CR-AD and
RT-CR-AD, respectively.
174 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
Fig. 6.5 RT-CK-AD detection results for TI with detected abundance fractions in db: (a) no
panels detected, (b) row 1 panels detected, (c) row 2 panels detected, (d) row 3 panels detected, (e)
row 4 panels detected, and (f) row 5 panels detected
Fig. 6.6 CR-AD detection results for TI with detected abundance fractions in db: (a) no panels
detected, (b) row 1 panels detected, (c) row 2 panels detected, (d) row 3 panels detected, (e) row
4 panels detected, and (f) row 5 panels detected
The second type of target insertion is TE, which is the same as TI described earlier
except for the way panel pixels are inserted. Background pixels are not removed to
6.5 Synthetic Image Experiments 175
Fig. 6.7 RT-CR-AD detection results for TI with detected abundance fractions in db: (a) no
panels detected, (b) row 1 panels detected, (c) row 2 panels detected, (d) row 3 panels detected, (e)
row 4 panels detected, and (f) row 5 panels detected
Fig. 6.8 Detection results for TE with detected abundance fractions in db: (a) 50th band of TE
scenario, (b) K-AD, (c) CK-AD, (d) RT-CK-AD, (e) R-AD, (f) CR-AD, and (g) RT-CR-AD
accommodate the inserted panel pixels as in TI but are rather superimposed with the
inserted panel pixels. Thus, in this case, the resulting synthetic image shown in
Fig. 6.8a has clean panel pixels embedded in a noisy background with the same
additive Gaussian noise as TI. The same experiments conducted for TI in Sect. 6.5.1
were repeated for the TE scenario. Figure 6.8b–g shows the detection results
produced by K-AD and R-AD, along with their causal and real-time causal
176 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
Fig. 6.9 CK-AD detection results for TE with detected abundance fractions in db: (a) no panels
detected, (b) row 1 panels detected, (c) row 2 panels detected, (d) row 3 panels detected, (e) row
4 panels detected, and (f) row 5 panels detected
Fig. 6.10 RT-CK-AD detection results for TE with detected abundance fractions in db: (a) no
panels detected, (b) row 1 panels detected, (c) row 2 panels detected, (d) row 3 panels detected, (e)
row 4 panels detected, and (f) row 5 panels detected
versions, where their performances are close with various degrees of background
suppression due to the use of global and causal correlation matrices. Figures 6.9,
6.10, 6.11, and 6.12 show the detection results of correlation matrix K-based real-
time causal anomaly detectors CK-AD and RT-CK-AD and covariance matrix
R-based anomaly detectors CR-AD and RT-CR-AD, respectively.
6.5 Synthetic Image Experiments 177
Fig. 6.11 CR-AD detection results for TE with detected abundance fractions in db: (a) no panels
detected, (b) row 1 panels detected, (c) row 2 panels detected, (d) row 3 panels detected, (e) row
4 panels detected, and (f) row 5 panels detected
Fig. 6.12 RT-CR-AD detection results for TE with detected abundance fractions in db: (a) no
panels detected, (b) row 1 panels detected, (c) row 2 panels detected, (d) row 3 panels detected, (e)
row 4 panels detected, and (f) row 5 panels detected
Despite the fact that causal and real-time causal anomaly detectors produce exactly the
same detection results, their required CPTs are different. The computer environments
used for experiments were a 64-bit operating system on an Intel i5-2500 3.3 GHz CPU
178 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
T 1
Fig. 6.13 CPT of CR-AD, RT-CR-AD,
n CK-AD,
onRT-CK-AD required
o for TI: (a) rnrn , (b) R (n),
(c) rnT ½Rðn 1Þ1 rn , and (d) ½Rðn 1Þ1 rn rnT ½Rðn 1Þ1
and 8 GB RAM. Figures 6.13a–d and 6.14a–d plot the CPT of calculating R1(n), rnrTn ,
n on o
rnT ½Rðn 1Þ1 rn , and ½Rðn 1Þ1 rn rnT ½Rðn 1Þ1 required for per pixel
vector processing by running four anomaly detectors, CR-AD, RT CR-AD, K-AD, and
RT-CK-AD, on TI and TE scenarios respectively where the x- and y-axes represent the
order of pixels being processed, that is, the nth pixel vector with n varying from 189 to
4000 and the CPT required to process the nth pixel vector. Comparing Figs. 6.13a, c
and 6.14a, c to Figs. 6.13b, d and 6.14b, d reveals that the computational complexity
requiring computing a matrix inverse in 103 s is one order higher than computing an
inner product in 104 s.
As shown in Figs. 6.13 and 6.14, the processing time of running the causal
anomaly detector, CR-AD/CK-AD and RT-CR-AD/RT-K-AD, on both the TI and
TE scenarios is nearly constant for each pixel vector. Thus, if an anomaly detector
is run on the entire image data, the CPT must be calculated by processing all image
pixel vectors. In this case, the CPT required for a causal anomaly detector, C-AD or
6.5 Synthetic Image Experiments 179
X40000
CK-AD, is CPT(OP(L ))+ n¼189
½CPTðMIðnÞÞ þ ðn þ 1ÞCPTðIPðLÞÞ , where
CPT(MI(n)) and CPT(IP(L )) are the CPT required to process finding the matrix
inverse, R1(n)/K1(n), and inner product of two L-dimensional vectors respec-
tively and the CPT(OP(L )) is used to calculate the matrix outer product, rnrTn of the
new nth L-dimensional input vector rn. On the other hand, the CPT required for a
real-time causal anomaly detector, RT-CR-AD or RT-CK-AD, is ð40000 189Þ
ðL þ 1ÞCPTðIPðLÞÞ þ CPT OPðLÞ , plus the CPT of processing the initial condi-
tion, CPT(MI(L )) where CPT(OP(L ) is the CPT required to calculate the matrix
outer product of two L-dimensional vectors. The results of various CPTs in seconds
required by running CR-AD, CK-AD, RT-CR-AD, and RT-CK-AD on complete
images of scenarios TI and TE are tabulated in Table 6.2.
One final comment on the performance in Figs. 6.3 and 6.8 is worthwhile.
Plotting the areas under their receiver operating characteristic (ROC) curve (Poor
1994) shows that they are all nearly close to one. However, their detection maps are
180
Table 6.2 CPT for causal and real-time causal versions of R-AD and K-AD
CPT CR-AD/CK-AD RT-CR-AD/RT-CK-AD
Initial condition CPT(R(L )) ¼ LCPT(OP(L )) CPT(R(L ) ¼ LCPT(OP(L )) + CPT(MI(L )) (R1(L )/K1(L ))
TI TE TI TE
0.01375 s 0.02111 s 0.0150 s 0.02882 s
n on o
CPT/pixel n > L CPT (rn(rn)T) 1 T 1
CPT ½Rðn 1Þ rn rn ½Rðn 1Þ þ 3CPTðxT AyÞ
TI TE
0.0001075 s 0.00009723 s TI TE
CPT(R1(n)/K1(n)))
TI TE 0.0004898 s 0.0004650 s
0.001222 s 0.001215 s
XN
Total CPT CPTðOPðnÞÞ þ CPTðMIðnÞÞ (N L ){[3(L + 1)CPT(IP(L ))} + CPT(OP(L ))]}
n¼L
TI TE TI TE
54.0114 s 53.3096 s 20.8528 s 6.8661 s
6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
6.6 Real Image Experiments 181
Two real hyperspectral image scenes were specifically selected for experiments to
conduct a performance evaluation of AD.
An Airborne Visible Infrared Imaging Spectrometer (AVIRIS) image data set was
used for the experiments shown in Fig. 6.15, which is the Lunar Crater Volcanic
Field (LCVF) shown in Fig. 1.8 located in Northern Nye County, Nevada. Atmo-
spheric water bands and low SNR bands were removed from the data, reducing the
image cube from 224 to 158 bands. The image in Fig. 6.15 has a 10 nm spectral
resolution and 20 m spatial resolution. There are five targets of interest, the radiance
spectra of red oxidized basaltic cinders, rhyolite, playa (dry lake), vegetation, and
shade. This scene is of particular interest because there is a two-pixel-wide anomaly
located at the left top edge of the crater.
Figure 6.16 shows the final detection maps in db produced by six anomaly
detectors, K-AD, R-AD, CK-AD, CR-AD, RT-CK-AD, and RT-CR-AD, where
all of them were able to detect the two-pixel-wide anomaly.
Figures 6.17, 6.18, 6.19, and 6.20 show progressive real-time causal processing
of CK-AD, CR-AD, RT-CK-AD, and RT-CR-AD in six progressive stages where
vegetation
cinders
shade anomaly
rhyolite
dry lake
Fig. 6.16 Detection maps of LCVF with detected abundance fractions in db: (a) K-AD, (b)
CK-AD, (c) RT-CKAD, (d) R-AD, (e) CR-AD, and (f) RT-CR-AD
Fig. 6.17 CK-AD detection results with detected abundance fractions in db: (a) vegetation, (b)
cinders, (c) playa and anomaly detected, (d) shade, (e) rhyolite, and (f) RT-CK-AD
the detected abundance fraction maps are displayed in db for better visual
assessment.
Interestingly, as the detection process progresses (Figs. 6.17, 6.18, 6.19, and
6.20), different levels of background suppression could also be witnessed. This was
particularly evident when the background was significantly suppressed once the
process detected the anomaly. This was because the detected abundance fraction of
the anomaly was so strong that the previously detected background information was
6.6 Real Image Experiments 183
Fig. 6.18 CR-AD detection results with detected abundance fractions in db: (a) vegetation, (b)
cinders, (c) playa and anomaly detected, (d) shade, (e) rhyolite, and (f) CR-AD
Fig. 6.19 RT-CK-AD detection results with detected abundance fractions in db: (a) vegetation,
(b) cinders, (c) playa and anomaly detected, (d) shade, (e) rhyolite, and (f) RT-CK-AD
overwhelmed by the anomaly. This is a very good example demonstrating the issue
of background suppression in AD, which will be further discussed in Sect. 6.6.3.
To further evaluate computational complexity, Fig. 6.21 plots the averaged CPT
[104 s for (a, c) and103 s for (b, d)] of running CR-AD, RT-CR-AD, CK-AD,
and RT-CK-AD on LCVF five timesnto compute various on individual operations,
o
rnrTn , R1(n), rnT ½Rðn 1Þ1 rn , and ½Rðn 1Þ1 rn rnT ½Rðn 1Þ1 , where
the x- and y-axes represent the order of pixels being processed, that is, the nth pixel
184 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
Fig. 6.20 RT-CR-AD detection results with detected abundance fractions in db: (a) vegetation,
(b) cinders, (c) playa and anomaly detected, (d) shade, (e) rhyolite, and (f) RT-CR-AD
Fig. 6.21 CPT of CR-AD, RT-CR-AD, nCK-AD, RT-CK-AD on required foro LCVF: (a) rnrTn ,
(b) R (n), (c) rn ½Rðn 1Þ rn , and (d) ½Rðn 1Þ rn rn ½Rðn 1Þ1
1 T 1 1 T
6.6 Real Image Experiments 185
vector in the data and CPT required to process the nth pixel vector. It is clearly
shown that the CPT resulting from using the recursive update equations (6.6) and
(6.8) nearly constant for each image pixel vector. This implies that the CPT is
indeed linearly increased with the number of data sample vectors processed. Similar
to Figs. 6.13 and 6.14, it also required one order higher to invert a matrix (in 103 s)
than to compute the inner product (in 104 s).
The HYDICE image scene in Fig. 1.10 is used for experiments and reproduced in
Fig. 6.22a, b. It has a size of 200 74 pixel vectors shown in Fig. 6.22a, along with
its ground truth provided in Fig. 6.22b, where the center and boundary pixels of
objects are highlighted by red and yellow, respectively.
The upper part in Fig. 6.22b contains fabric panels with sizes of 3, 2, and 1 m2
from the first column to the third column. Since the spatial resolution of the data is
1.56 m2, the panels in the third column are considered subpixel anomalies. The
lower part in Fig. 6.22c contains different vehicles with sizes of 4 8 m (the first
four vehicles in the first column) and 6 3 m (the bottom vehicle in the first
column) and three objects in the second column (the first two have a size of
two pixels and the bottom one has a size of three pixels), respectively. In this
particular scene, there are three types of artificial targets with different sizes, small
targets (panels of three different sizes, 3, 2, and 1 m2) and large targets (vehicles of
two different sizes, 4 8 m and 6 3 m and three objects two pixels and three
pixels in size) to be used to validate and test AD performance.
There are several advantages of using this HYDICE image scene in Fig. 6.22a.
One is that the ground truth provides precise spatial locations of all artificial target
pixels, which allows us to evaluate the real-time processing performance of AD
pixel by pixel, a task that has not been explored before now. Second, the provided
ground truth enables us to perform ROC analysis for AD via ROC curves of the
detection rate versus false alarm rate. Third, the scene has objects of various sizes
that can be used to evaluate the ability of an anomaly detector to detect anomalies of
different sizes, an issue that has not received sufficient treatment in many reports.
Fourth, this scene can be processed by operating the same anomaly detector on
three different image sizes (Fig. 6.22a–c) (i.e., a 15-panel scene 64 64 pixel
vectors in size, marked by an upper rectangle, and a vehicles + objects scene of
size 1000 64 pixel vectors marked by a lower rectangle, and the entire scene
containing 15 panels and vehicles + objects) to evaluate the effectiveness of its
performance. Finally and most importantly, the clean natural background and
targets make it easier to see various degrees of background being suppressed by
an anomaly detector.
186 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
Fig. 6.22 HYDICE scene with ground truth map of spatial locations of 15 panels, 5 vehicles and
3 objects. (a) HYDICE scene with ground truth map of spatial locations of 15 panels, 5 vehicles
and 3 objects, (b) 15 panels, and (c) vehicles + object scene with ground truth map of 5 vehicles
and 3 objects
6.6 Real Image Experiments 187
To see how causal and real-time causal AD perform, Figs. 6.23, 6.24, 6.25, and 6.26
show the real-time causal processing of CK-AD, CR-AD, RT-CK-AD, and RT-CR-
AD on Fig. 6.22a with their detected abundance fractions in db, where each pass
shows a real-time detection map of different targets.
Since panels b and c in Fig. 6.22 are part of the scene in Fig. 6.22a, the results of
real-time processing of these two subscenes are not included here. Nevertheless,
their detection results will be discussed in detail in the following two subsections. It
is worth noting that to avoid the singularity problem of calculating the inverse of the
sample correlation/covariance matrix used by anomaly detectors, an anomaly
detector will not begin to operate until it collects a sufficient number of initial
data sample vectors, which is the total number of spectral bands of the image to be
processed. It should be noted that K-AD and R-AD are not included in the
experiments because they are neither causal nor real time. By visually inspecting
the results in Figs. 6.23, 6.24, 6.25, and 6.26, sample correlation matrix R-based
causal and real-time causal anomaly detectors seemed to perform slightly better that
than their K-based counterparts in terms of panel pixel detection. Interestingly, the
conclusion is reversed if the detection of vehicles is of major interest. This obser-
vation is further confirmed by the following ROC analysis. Since the CPT for this
scene is very similar to that in Fig. 6.21, its plots are not included here.
Using the ground truth provided by Fig. 6.22, we can perform a quantitative study
via ROC analysis. In doing so, an idea similar to that proposed in Chang and Ren
(2000) and Chang (2003) can be derived by converting real values to hard decisions
as follows.
Assume that δAD(r) is the detected abundance fraction obtained by operating an
anomaly detector on a data sample vector r. We then define a normalized detected
abundance fraction ^δ
AD
ðrÞ by
normalized
More specifically, ^δ normalized ðrÞ in (6.25) can be regarded as a probability vector that
AD
8
< 1, if ^δ AD a
normalized ðrÞ τ ¼ ,
χ a%APAC ðrÞ ¼ 100 ð6:26Þ
:
0, otherwise:
Fig. 6.27 3D ROC curve and its three corresponding 2D ROC curves. (a) 3D ROC curve of (PD,
PF,τ) for 15-panel, vehicle, and entire scenes; (b) 2D ROC curve of (PD,PF); (c) 2D ROC curve of
(PD,τ); and (d) 2D ROC curve of (PF,τ)
in Fig. 6.22a, c, the best anomaly detector was real-time or causal K-AD for the
vehicles scene in Table 6.5 according to the values of Az calculated for the 2D ROC
curves of (PD,PF) and (PD,τ). Interestingly, for all three image scenes, the best
anomaly detector to produce the smallest Az of (PF,τ) was real-time/causal K-AD/
R-AD. This indicates that a smaller Az of (PF,τ) implies less background suppres-
sion. Furthermore, a higher Az of (PD,PF) does not necessarily imply a higher Az of
(PF,τ), as shown in Tables 6.4 and 6.5. Unfortunately, two such pieces of informa-
tion are not provided by the traditional 2D ROC analysis, Az of (PD,PF). These
experiments demonstrate the utility of 3D ROC analysis via three 2D ROC curves
generated from a 3D ROC curve, where AD performance can be analyzed through
interrelationships among PD, PF, and the threshold τ via three 2D ROC curves
plotted based on three different pairs, (PD,PF), (PD,τ), and (PF,τ).
192 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
Figure 6.29 suggests that all six anomaly detectors performed comparably in the
detection of targets, but the global anomaly detectors K-AD and R-AD had better
background suppression than their real-time and causal counterparts in terms of
suppressing grass surrounding panels, vehicles, and objects. This certainly made
sense. Since a global anomaly detector utilizes the global spectral correlation
provided by the sample correlation/covariance matrix of all the image data, it
performs better background suppression than any local anomaly detector, as
expected. However, on many occasions, when no prior knowledge is available,
background information may help image analysts perform better data analysis
because background generally provides crucial information surrounding anomalies.
If background suppression is overdone, we may have no clue about anomalies. For
example, in Fig. 6.28, anomalies were detected with very clean background sup-
pression, in which case we have no idea what these anomalies are; all we know are
their spatial locations. But if we look closely at Fig. 6.29, the background has a tree
line along the left edge, panels were placed on grass, while vehicles were actually
6.6 Real Image Experiments 195
Table 6.3 Values of areas under three 2D ROC curves, Az, produced by six algorithms (entire
panels + vehicles scene)
Algorithm K-AD CK-AD RT-CK-AD R-AD CR-AD RT-CR-AD
Az of (PD,PF) 0.9886 0.9818 0.9819 0.9840 0.9747 0.9747
Az of (PD,τ) 0.2368 0.1372 0.1372 0.2349 0.1356 0.1356
Az of (PF,τ) 0.0193 0.0144 0.0144 0.0199 0.0145 0.0145
Table 6.4 Values of areas under three 2D ROC curves, Az, produced by six algorithms (15-panel
scene)
Algorithm K-AD CK-AD RT-CK-AD R-AD CR-AD RT-CR-AD
Az of (PD,PF) 0.9898 0.9680 0.9683 0.99 0.9691 0.9691
Az of (PD,τ) 0.3329 0.2590 0.2590 0.3342 0.2596 0.2596
Az of (PF,τ) 0.0428 0.0372 0.0372 0.0433 0.0377 0.0377
Table 6.5 Values of areas under three 2D ROC curves, Az, produced by six algorithms (vehicles
scene)
Algorithm K-AD CK-AD RT-CK-AD R-AD CR-AD RT-CR-AD
Az of (PD,PF) 0.9751 0.9776 0.9776 0.9669 0.9662 0.9662
Az of (PD,τ) 0.2172 0.1307 0.1307 0.2150 0.1294 0.1294
Az of (PF,τ) 0.0332 0.0221 0.0221 0.0333 0.0222 0.0222
This section calculated the computing time in seconds required by running two
causal anomaly detectors, CK-AD and CR-AD, and two real-time causal anomaly
detectors, RT-CK-AD and RT-CR-AD, on LCVF and three HYDICE scenes by
averaging five different runs where the computer environments used for experi-
ments was 64-bit operating system on an Intel i5-2500 3.3 GHz processor, , and
8 GB RAM. Their results are tabulated in Table 6.6 and show interesting findings.
The time required by causal AD to compute initial conditions for a vehicle scene
was greater than that for an entire scene that included a vehicle scene as a subscene.
Similarly, this is also true of the real-time causal AD, where the computing time for
the panel scene is greater than the scene with panels + vehicles. This is because
none of the causal and real-time causal anomaly detectors starts processing data
until sufficient data sample vectors are collected to compute initial conditions, such
as the total number of spectral bands, to avoid an ill-ranking issue in calculating
196 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
Fig. 6.28 Detection maps with detected abundance fractions: (a) K-AD, (b) R-AD, (c) CK-AD,
(d) CR-AD, (e) RT-CK-AD, and (f) RT-CR-AD
sample correlation/covariance matrix. In this case, all the algorithms use the same
set of 169 image pixels to calculate their initial conditions, and their algorithmic
structures determine the computational complexity.
6.6 Real Image Experiments 197
As we see from Table 6.6, a real-time causal anomaly detector generally runs
two or three times faster than its causal counterpart while showing the same
performance. Table 6.7 also tabulates the computing time required by two global
anomaly detectors, K-AD and R-AD, to process all the image data.
198
A comparison of Tables 6.7 and 6.6 may lead one to believe that the computa-
tional complexity required by a real-time causal anomaly detector is exceedingly
high. As a matter of fact, it is not if we consider that a real-time causal anomaly
detector must update and recalculate its sample correlation/covariance matrix every
time a new input data sample vector comes in, and its computational complexity is
linearly increased with the number of data sample vectors that must be processed, as
shown in Fig. 6.21, where the processing time for each pixel is nearly constant.
Thus, since a global anomaly detector only needs to calculate the sample correla-
tion/covariance matrix once for all data sample vectors, the computing time of a
real-time causal anomaly detector is supposed to be the computing time required by
a global anomaly detector multiplied by the total number of data sample vectors
being processed by it, to some extent. However, the computing time documented in
Table 6.6 is significantly less than that. This tremendous savings was mainly due to
the fact that a real-time causal anomaly detector makes use of a recursive causal
update equation specified by either (6.21) or (6.24), which uses only innovation
information provided by the current data sample vector to update the equation
without reprocessing already visited data sample vectors.
Since RT-AD using the sample correlation matrix R(n) to suppress the background
is carried out in the same way as RT-CEM using R(n), a theory similar to RT-CEM
using band-interleaved-line (BIL) in Sect. 5.4 can also be developed for RT-AD
using BIL as follows.
Suppose that the size of a data matrix is M N, where there are M data lines and
each line has N pixels. Let Ll be the lth data line and L(n) the data comprising all the
200 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
N , n
first n data lines, that is, LðnÞ ¼ [l¼1
n
Ll ¼ ril i¼1, l¼1 . Then RðLl Þ ¼ ð1=N Þ
XN T
r l r l , where the number of pixels in each line is N and RðLðnÞÞ ¼
i¼1 i i
Xn Xn, N T
ð1=nÞ l¼1 RðLl Þ ¼ ðnN Þ1 l¼1, i¼1 ril ril . To derive a real-time version of
AD using BIL, we need the following matrix identity in Appendix A
1
ðA þ BCDÞ1 ¼ A1 A1 B DA1 B þ C1 DA1 : ð6:27Þ
ð6:30Þ
I D
(n-1)-1 -1 n/(n-1)
R-1(L(n-1))
Xn Ln R(Ln)
According to the concept of causal processing and the recursive equation derived
in (6.26), a real-time R-AD, referred to as RT-R-AD, can be formulated as
ð6:31Þ
To implement RT-K-AD using BIL, we first express the sample covariance matrix
K(L(n)) of the first n data lines L(n) as
Xn XN l T
KðLðnÞÞ ¼ ð1=nN Þ1 l¼1 i¼1
r i
l
μð L ð n Þ Þ r i μ ð L ð n Þ Þ
Xn XN Xn XN
1 l l T 1
¼ ð1=nN Þ l¼1
r
i¼1 i i
r 2 ð 1=nN Þ l¼1
r
i¼1 i
l
μT ðLðnÞÞ
Using (6.31) and (6.32) we can derive RT-K-AD using BIL as follows:
n pffiffiffi on pffiffiffi o
½ð1 1=nÞRðLðn 1ÞÞ1 ð1= nÞr ð1= nÞrT ½ð1 1=nÞRðLðn 1ÞÞ1
ðr μðLðnÞÞÞT pffiffiffi pffiffiffi
1 þ ð1= nÞrT ½ð1 1=nÞRðLðn 1ÞÞ1 ð1= nÞr
ðXn μðLðnÞÞÞ:
ð6:33Þ
Unlike RT-R-AD, using BIL where δRT - R - AD (X ) can be updated by δRT -R-AD ðX Þ
n n n1 n
RT - K - AD
recursively via (6.31), δn (Xn) specified by (6.33) cannot be updated recur-
RT -K-AD
sively by δn1 ðXn Þ. This implies that RT-K-AD does not have the same
advantage as RT-R-AD in terms of being implemented as a Kalman filter.
202 6 Real-Time RHSP for Passive Target Detection: Anomaly Detection
In analogy with the experiments conducted for the 15-panel HYDICE scene in
Sect. 5.6.1, Fig. 6.31a, b shows the final anomaly detection maps resulting from
operating RT-K-AD using BIL and RT-R-AD using BIL on the 15-panel HYDICE
data in Fig. 6.26b respectively for comparison, where detection maps are shown in
grayscale maps, color maps, and color maps in db.
Despite the fact that both RT-K-AD and RT-R-AD produced nearly the same
result with no appreciable differences in grayscale detection maps in Fig. 6.31 by
inspection, there remain some visible differences in detection of panels in rows
4 and 5 shown in color detection maps due to different levels of background
suppression. As also shown in Fig. 6.31, the color detection maps suppressed
more background than the color detection map in dB did because dB closed the
gap in gray values between the detected panel pixels and the background. As
discussed in Chaps. 5 and 16 in Chang (2016), background suppression is a crucial
factor in performance assessment when no prior knowledge is available. To further
see this, Figs. 6.32 and 6.33 show a sequence of progressive color detection maps
produced by RT-K-AD using BIL and RT-R-AD using BIL respectively along their
respective results in dB shown in Figs. 6.34 and 6.35, where the color detection
maps suppressed more background than the color detection map in dB did.
Fig. 6.31 Final detection maps of RT-AD using BIL on 15-panel HYDICE data in Fig. 6.26b. (a)
δRT - K - AD (X ), (b) δRT - R - AD (X )
n n n n
6.7 RT-AD Using BIL 203
As we can from the foregoing results, progressive detection maps provide more
information than one-shot results (Fig. 6.31). For example, the panel pixels in row
1 were visibly detected in the first few passes and began to fade away when more
panel pixels were detected.
With regard to experiments for synthetic image data in Fig. 6.2, the interested
reader is referred to Li and Chang (2016), which has detailed experiments.
6.8 Conclusions
as a signal source and then use this signal source to formulate a binary hypothesis
testing problem. A desired NPD can be developed in a manner similar to how the
Harsanyi–Farrand–Chang (HFC) method was developed by Harsanyi et al. (1994)
to estimate the virtual dimensionality (VD), which is extended to target-specified
VD in Chap. 4, where target signatures to be specified can be generated by a
recursive process. The NPD determines whether or not a signal source specified
by the maximal OP leakage of a target signature fails the test, in which case the
considered target signature is declared to be a real target. Since the maximal OP
leakages calculated from progressively generated targets are monotonically
decreasing, the NPD generally fails in the beginning and the test is then continued
until it reaches the first time the NPD passes the test, in which case the NPD is
terminated. Because Neyman–Pearson detection is performed in conjunction with a
recursive process that generates each new target signature, it can be implemented in
real time while a new target is being generated at the same time.
Part III is mainly focused on RHSP. Chapter 7, “Recursive Hyperspectral
Sample Processing of the Automatic Target Generation Process,” extends a well-
known automatic target detection algorithm, the automatic target generation pro-
cess (ATGP), to RHSP-ATGP, which can implement ATGP recursively. Using an
idea similar to that used to derive RHSP-ATGP, Chap. 8, “Recursive Hyperspectral
Sample Processing of Orthogonal Subspace Projection,” also extends the well-
known orthogonal subspace projection (OSP) to RHSP-OSP. Similar to OP used
by both ATGP and OSP to derive recursive equations, a second application is
discussed in Chap. 9, “Recursive Hyperspectral Sample Processing of Linear
Spectral Mixture Analysis,” and Chap. 10, “Recursive Hyperspectral Sample
Processing of Maximum Likelihood Estimation,” use least-squares error (LSE) in
RHSP-LSMA and RHSP-MLE, respectively. Finally, growing simplex volume
analysis (GSVA) is considered as a third application in finding endmembers.
Chapter 11, “Recursive Hyperspectral Sample Processing of Growing Simplex
Volume Analysis” makes use of OP to derive a recursive version of SGA developed
in Chang et al. (2006) called RHSP OP-Based Simple Growing Algorithm (RHSP-
OPSGA), which not only quickly computes simplex volumes but also significantly
reduces computer processing time. With another new approach different from
RHSP-OPSGA, Chap. 12, “Recursive Hyperspectral Sample Processing of
Geometric Simplex Growing Algorithm” makes use of the Gram–Schmidt orthog-
onalization process (GSOP) to derive an RHSP-geometric SGA (RHSP-GSGA). It
turns out that RHSP-OPSGA and RHSP-GSGA are the best algorithms of all the
SGA-based variants reported in the literature in terms of computational complexity
and computing time.
Chapter 7
Recursive Hyperspectral Sample Processing
of Automatic Target Generation Process
7.1 Introduction
The automatic target generation process (ATGP), developed by Ren and Chang
(2003), was originally designed to find targets of interest in an unsupervised manner
when no prior knowledge is available. Due to its effectiveness and simple imple-
mentation, ATGP has become a versatile technique and has also proven very useful
in many applications. For example, it has been used for anomaly detection (Chang
2003a), endmember finding (Chang and Plaza 2006; Chang 2013, 2016),
unsupervised linear spectral mixture analysis (Chang et al. 2010a), and magnetic
resonance imaging (Wang et al. 2002). Its Field Programmable Gate Array (FPGA)
design (Bernabe et al. 2011) and graphics processing unit (GPU) implementation
(Bernabe et al. 2013) have also been studied. Most recently, it was shown that many
algorithms, such as vertex component analysis (VCA) developed by Nascimento
and Bioucas-Dias (2005) and the simplex growing algorithm (SGA) developed by
Chang et al. (2006a, b, c, d), can be considered as its variants (Chang et al. 2013),
and their underlying ideas are exactly the same as that of ATGP (Chang 2013,
Chap. 11, 2016, Chap. 13; Chang et al. 2016a, b).
Although ATGP has already been proven very effective, it still has room
for improvement from a practical point of view, specifically, its repeated imple-
mentation of orthogonal subspace projections (OSPs). Assume that t1 is its initial
target obtained
by
finding
a data sample vector with maximal vector length, that is,
t1 ¼ arg maxr r , to produce an initial target subspace hU1i generated by a
matrix U0 formed by U1 ¼ ½t1 . By means of U1 ATGP finds its second target, t2,
with the maximal vector length
in the subspace orthogonal to hU1i, that is,
T T 1 T
t2 ¼ arg maxr P⊥ U1 r P ⊥
U1 r , where P⊥ U 1 ¼ I t1 t1 t1 t1 . Suppose
that t1 , t2 , . . . , tp1 are p 1 targets already found by ATGP. Forming Up1 ¼
t1 t2 tp1 as a target subspace ATGP can extract its pth target by finding a data
sample vector with the maximal vector length in a subspace orthogonal to Up1 as
follows:
T
where
1
P⊥
Up1 ¼ I U p1 U T
U
p1 p1
T
Up1 ; ð7:2Þ
with Up1 defined as a target subspace linearly spanned by p 1 targets, t1 , t2 ,
. . . , tp1 , already found by ATGP. In this case, a sequence of nested embedded
target subspaces {hUpi}p can be produced as hU1 i hU2 i Up1 Up
hUk i , where target subspaces generated by targets previously found by
ATGP are embedded in and also part of the target subspaces produced by targets
subsequently found by ATGP.
7.2 Recursive Hyperspectral Sample Processing of ATGP 211
For ATGP to find a set of p targets, t1, . . ., tp it must repeatedly implement (7.1)
and (7.2) p times. It does not take full advantage of the fact
that {hUpi}p is a
sequence of nested embedded subspaces where Up1 in Up ¼ Up1 tp is fixed and
remains unchanged when ATGP finds all subsequent targets beyond the p 1
targets. Second, in Ren and Chang (2003), ATGP derived a stopping measure,
called the orthogonal projection correlation index (OPCI), defined by
1
where U#p1 ¼ Up1
T
Up1 T
Up1 and
1
1 n h i o1
β ¼ tpT I Up1 Up1
T
Up1 T
Up1 tp ¼ tpT P⊥
Up1 tp ; ð7:5Þ
p p1
212 7 Recursive Hyperspectral Sample Processing of Automatic Target Generation Process
2 T 3
1
6 Up1 þ βUp1 tp tp Up1 Up1 βUp1 tp tp 7
# # T # T # T
U#p ¼ UpT Up UpT ¼ 4 T 5; ð7:6Þ
βtpT U#p1 Up1 T
þ βtpT
1
Up U#p ¼ Up1 tp UpT Up UpT
T
¼ Up1 U#p1 þ βUp1 U#p1 tp tpT U#p1 Up1
T
βUp1 U#p1 tp tpT ð7:7Þ
T
βtp tpT U#p1 Up1
T
þ βtp tpT ;
P⊥ ⊥
Up ¼ I Up Up ¼ PUp1 β u
#
ep1 u
ep1
T
2e
u p1 tpT þ tp tpT
T ð7:8Þ
¼ P⊥Up1 β u ep1 tp u ep1 tp ;
T
r T P⊥ T ⊥
Up r ¼ r PUp1 r βr u
T e
p1 tp u ep1 tp r
T 2 ð7:9Þ
¼ r T P⊥
Up1 r β r u ep1 tp ;
where tp+1 can also be found by (7.1) and (7.9) is used to find the next tp+1 target.
1
As noted, P⊥ Up1 ¼ I Up1 U #
p1 ¼ I Up1 U T
p1 Up1
T
Up1 , and tp is
ep1 ¼ Up1 U#p1 tp and tpT u
obtained by (7.1), along with u ep1
ep1 ¼ u T
tp . It should
T e
T
also be pointed out that r u p1 tp ¼ u ep1 tp r and β are scalars, where
n h i o1
tpT u ep1
ep1 ¼ u T
tp and β ¼ tpT P⊥
Up1 tp are used to account for the correlation
between Up1 and tp.
According to (7.2), P⊥ Up1 implemented by ATGP can be updated by PUp2
⊥
1. Initial condition:
Find an initial target pixel vector, t1 ¼ argfmaxr rT rg. Set U1 ¼ ½t1 , and
calculate P⊥U1 .
7.2 Recursive Hyperspectral Sample Processing of ATGP 213
• Inner Loop
(a) Progressive process indexed by parameter i (running through all data
N
sample vectors fri gi¼1 ):
For p 2, find tp by maximizing riT P⊥ Up1 ri via (7.9) over all data
sample vectors, r.
T
It is worth noting that riT P⊥Up1 r i in (7.9) is identical to P ⊥
Up1 r i
T
PUp1 ri owing to the fact that PUp1 ¼ PUp1 and P⊥
⊥ ⊥ ⊥
Up1 is
T T
idempotent. That is, P⊥
Up1 ri P⊥Up1 ri ¼ ri PUp1
T ⊥
P⊥Up1 ri ¼
riT P⊥
Up1 ri .
(b) Recursive process by updating P⊥
Up :
Use (7.6) and (7.8) to update U#p and P⊥
Up via previously calculated
U#p1 and P⊥
Up1 .
To terminate RHSP-ATGP, a stopping rule was proposed in Ren and Chang (2003),
where (7.3) was introduced as an error threshold for convergence. Interestingly,
we can modify the OPCI by replacing t0 in (7.3) with the pth target, tp, obtained
by (7.1) to yield a new measure,
ηp ¼ tpT P⊥
Up1 tp ; ð7:10Þ
which turns out to be exactly the same signal source under the binary hypothesis
testing problem considered in the maximal orthogonal subspace projection
(MOSP)
in Chang et al. (2011c), where the energy of the pth target signal tp, tp 2 , was used
to determine the virtual dimensionality (VD) or the rank of rare signals by the
maximal orthogonal complement algorithm (MOCA) in Kuybeda et al. (2007). The
only difference between OPCI in (7.3) and ηp in (7.10) is that the former requires a
prescribed error ε to terminate ATGP, while the latter needs a predetermined false
alarm probability to terminate RHSP-ATGP.
Following the target-specified virtual dimensionality (TSVD) developed in
Sect. 4.5.1 of Chap. 4, we assume that for each 1 p L, U0 ¼ ∅, and Up1 is
n op1
the target space linearly spanned by the p 1 targets tR-ATGP j previously
j¼1
found by RHSP-ATGP, which is exactly the space Up1 described in the RHSP-
ATGP algorithm, Up1 ¼ Up1. Then for 1 p L we can find
n o
tpRHBP-ATGP ¼ arg maxr P⊥
Up1 r
ðwhich is the same as ð7:1ÞÞ; ð7:11Þ
2
ηpRHSP-ATGP ¼ P⊥ RHSP-ATGP 2
Up1 tp ¼ maxr P⊥
Up1 r ; ð7:12Þ
where ηRHSP - ATGP in (7.12) is the maximal residual of the pth target data sample,
p
RHSP - ATGP
tp found by RHSP-ATGP leaked from < Up1 > into < Up1 >⊥, which is
the complement space orthogonal to the space < Up1 >. This ηpRHSP - ATGP is
exactly the same as vkspecified by equation
(7.11) in Kuybeda et al. (2007).
It should be noted that P⊥ U0 r ¼ rT r in (7.11) when p ¼ 1. Since ηRHSP - ATGP
p
in (7.12) calculates the maximum residual of tpRHSP - ATGP leaked into < Up1 >⊥ , it
n oL
is this sequence, ηRHSP -ATGP , that will be used as the signal source in a binary
p
p¼1
216 7 Recursive Hyperspectral Sample Processing of Automatic Target Generation Process
T
Since P⊥
Up1 is symmetric and idempotent, that is, P ⊥
Up1 P⊥ ⊥
Up1 ¼ PUp1 , (7.13)
is reduced to
⊥ RHSP-ATGP 2
P ¼ tRHSP-ATGP TP⊥ RHSP-ATGP
¼ ηRHSP-ATGP ; ð7:14Þ
Up1 tp p Up1 tp p
which is exactly (7.12). This signal energy can be used as a signal source in a binary
composite hypothesis testing problem (Poor 1994) to determine whether the pth
potential target candidate tpRHSP - ATGP is a true target by a detector. That is, for p
¼ 1, 2, . . . , L,
H 0 : ηRHSP-ATGP governed by a pdf p ηRHSP-ATGP H ¼ p ηRHSP-ATGP
p p 0 0 p
versus
H : ηRHSP-ATGP governed by a pdf p ηRHSP-ATGP H ¼ p ηRHSP-ATGP ;
1 p p 1 1 p
ð7:15Þ
where the alternative hypothesis H1 and the null hypothesis H0 represent two cases
of tpRHSP - ATGP as a target signal source under H1 and not a target signal source under
H0, respectively, in the sense that H0 represents the maximum residual resulting
from the background signal sources, while H1 represents the maximum residual
leaked from the target signal sources. To make (7.15) work, we need to find
probability distributions under both hypotheses. The assumptions made on (7.15)
are derived from Kuybeda et al. (2007), where H0 represents a background that can
be characterized by a Gaussian distribution and H1 represents targets of interest that
are distributed uniformly. That is, according to Kuybeda et⊥ al. (2007),
the signal
P⊥ r can be modeled by a Gaussian distribution so that P r2 is a chi-squared
Up1 Up1
distribution that can be approximated
2 by a Gaussian distribution asymptotically.
Then the maximal value of P⊥ Up1 r over r, ηRHSP-ATGP ¼ maxr P⊥ r2 via
p Up1
(7.12), can be shown by the extreme value theory (Leadbetter 1987) to converge to
a Grumbel distribution as was done in Kuybeda et al. (2007). For details on deriving
probability distributions, we refer the interested reader to Kuybeda et al. (2007).
By virtue of the Neyman–Pearson detection theory, we can derive the following
Neyman–Pearson detector, which maximizes the detection power PD subject to the
false alarm probability PF, that in turn determines the threshold value τp:
7.4 Synthetic Image Experiments 217
8
>
> 1, if Λ ηRHSP-ATGP
> τp
>
>
p
<
RHSP-ATGP -ATGP ¼ τ ,
δNP
RHSP-ATGP ηp ¼ 1 with probability κ, if Λ ηRHSP p
>
>
p
>
>
: 0, if Λ ηRHSP-ATGP < τ ;
p p
ð7:16Þ
where the likelihood ratio test Λ(ηpRHSP - ATGP ) is given by Λ ηpRHSP-ATGP ¼ p1
ηpRHSP-ATGP =p0 ηpRHSP-ATGP , with p0 ηpRHSP-ATGP , and p1 ηpRHSP-ATGP ,
given by (7.15), and the constant κ is the probability of saying H1. Thus, a
case of ηpRHSP-ATGP ¼ tpT P⊥ RHSP - ATGP
Up1 tp > τp indicates that δRHSP - ATGP (ηp
NP
) in
RHSP - ATGP
(7.16) fails the test, in which case t p is assumed to be a desired
target. Note that the test for (7.16) must be performed for each of the
potential target candidates. Therefore, for a different value of p, the threshold
τp varies. Using (7.16) the VD can be determined by calculating
n h io
RHSP-ATGP
VDNP
RHSP-ATGP ð P F Þ ¼ arg max δNP
p RHSP-ATGP η p ¼ 1 ; ð7:17Þ
j k
RHSP-ATGP
where PF is a predetermined false alarm probability, δNPRHSP-ATGP ηp ¼ 1,
j k
RHSP-ATGP RHSP-ATGP
only if δNP RHSP-ATGP ηp ¼ 1 and δNP RHSP-ATGP ηp ¼ 0 if
RHSP-ATGP
RHSP-ATGP ηp
δNP < 1.
A comment on (7.17) is worthwhile. The sequence of {ηRHSP - ATGP } is mono-
p
RHSP - ATGP
tonically decreasing. Thus, the sequence {δNP RHSP - ATGP (ηp )} starts with a
RHSP-ATGP
failure of the test (7.16), that is, δRHSP-ATGP ηp
NP
¼ 1 and continues as
the value of p is increased
until preaches the value where the test is passed, in
RHSP-ATGP
which case δRHSP-ATGP ηp
NP
< 1. The largest value of p makes δNP RHSP-ATGP
RHSP-ATGP
ηp ¼ 1 the value of VDRHSP - ATGP (PF) according to (7.17). This unique
NP
property allows VDNP RHSP - ATGP (PF) to be implemented in real time as the process is
continued with increasing p.
The synthetic image data shown in Fig. 7.2 (also shown in Fig. 1.15) have two
scenarios, target implantation (TI) and target embeddedness (TE), that can be used
for experiments.
218 7 Recursive Hyperspectral Sample Processing of Automatic Target Generation Process
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
There are 25 panels with five 4 4 pure pixel panels for each row in the first
column, five 2 2 pure pixel panels for each row in the second column, five 2 2
mixed pixel panels for each row in the third column, and five 1 1 subpixel panels
for each row in both the fourth and fifth columns, where the mixed panel pixels and
subpanel pixels, all of which were simulated according to the legends in Fig. 7.2.
Thus, a total of 100 pure pixels (80 in the first column and 20 in the second column),
referred to as endmember pixels, were simulated in the data by the five
endmembers A, B, C, K, and M. An area marked “BKG” in the upper right corner
of Fig. 1.14a was selected to find its sample mean, that is, the average of all pixel
vectors within the “BKG” area, denoted by b and plotted in Fig. 1.14b, to be used to
simulate the background (BKG) with a signal-to-noise ratio (SNR) of 20:1, defined
in Harsanyi and Chang (1994) for an image scene with a size of 200 200 pixels in
Fig. 7.2. Once the target pixels and background are simulated, two types of target
insertion, TI and TE, can be designed to simulate experiments for various
applications.
Figure 7.3 shows 189 targets generated recursively by RHSP-ATGP for TI and
TE target by target, where numerals are used to indicate the orders in which these
targets were generated. Because different criteria were used, (7.3) for ATGP and
(7.10) for RHSP-ATGP, ATGP and RHSP-ATGP did not necessarily produce the
same targets as shown in Fig. 7.3. Nonetheless, sets of their extracted targets were
nearly identical.
Two real hyperspectral image scenes were used for experiments. The first image
data to be studied are the HYperspectral Digital Imagery Collection Experiment
(HYDICE) image scene shown in Fig. 7.4a (and Fig. 1.10a), with a size of 64 64
pixel vectors with 15 panels in the scene and the ground truth map in Fig. 7.4b
(Fig. 1.10b). It was acquired by 210 spectral bands with a spectral coverage from
0.4 to 2.5 μm. Low signal/high noise bands, bands 1–3 and 202–210, and water
vapor absorption bands, bands 101–112 and 137–153, were removed. Thus, a total
of 169 bands were used in the experiments. The spatial resolution and spectral
resolution of this image scene are 1.56 m and 10 nm, respectively.
7.4 Synthetic Image Experiments 219
Fig. 7.3 One hundred eighty-nine targets generated by RHSP-ATGP for TI and TE
a b
p11, p12, p13
Fig. 7.4 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels
Another data set is one of the most widely used hyperspectral image scenes
available in the public domain, Cuprite Mining District site in Nevada, shown in
Fig. 7.5a (and Fig. 1.6). It is an image scene with 20 m spatial resolution collected by
224 bands using 10 nm spectral resolution in the range of 0.4–2.5 μm. Since it is well
understood mineralogically and has reliable ground truth, this scene has been studied
extensively. There are five pure pixels in Fig. 7.5a, b, which can be identified as
corresponding to five different minerals, alunite (A), buddingtonite (B), calcite (C),
kaolinite (K), and muscovite (M), labeled A, B, C, K, and M in Fig. 7.5b.
Figure 7.6 shows targets generated recursively by RHSP-ATGP for TI and TE
target by target with numerals used to indicate the orders in which these targets
were generated. In general, both ATGP and RHSP-ATGP should not generate
identical sets of targets because of the different criteria used to generate targets,
(7.3) for ATGP and (7.10) for RHSP-ATGP. Interestingly, as shown in Fig. 7.6,
most of the targets extracted by ATGP and RHSP-ATGP were nearly the same
except in a few cases. In this case, if the targets generated by ATGP and
220 7 Recursive Hyperspectral Sample Processing of Automatic Target Generation Process
Fig. 7.5 (a) Spectral band number 170 of Cuprite AVIRIS image scene; (b) spatial positions of
five pure pixels corresponding to minerals alunite (A), buddingtonite (B), calcite (C), kaolinite (K),
and muscovite (M)
Fig. 7.6 Targets generated by ATGP and RHSP-ATGP for HYDICE and Cuprite data sets. (a)
HYDICE (169 targets) and (b) cuprite (189 targets)
RHSP-ATGP are the same, they are highlighted by yellow circles; otherwise,
targets from ATGP and RHSP-ATGP are highlighted by red circles and blue
circles, respectively, where four targets are different for the HYDICE data in
Fig. 7.6a and only one target is different for the Cuprite data in Fig. 7.6b.
7.5 Discussions on Stopping Rule for RHSP-ATGP 221
which is exactly the energy of the residual of tp, ηp, leaked into Up1. By means of
this ηp a recursive version of ATGP, RHSP-ATGP, can be derived and further used
to estimate ATGP-specified VD, nRHSP-ATGP, to determine how many targets are
required for RHSP-ATGP to generate via a binary hypothesis testing problem
(7.16), that is, nRHBP-ATGP ¼ VDNPRHSP-ATGP ðPF Þ, specified by (7.17).
Since noise whitening can affect VD estimation (Chang 2003a, b; Chang and Du
2004), we also ran RHSP-ATGP on two different data sets, the original data set and
a noise-whitened data set, where the noise covariance matrix was estimated using
the technique developed in Roger and Arnold (1996). Table 7.1 tabulates the values
n oL
of n RHSP-ATGP estimated by (7.17), where ηRHSP-ATGP
p in (7.12) was used as
p¼1
the signal source for the binary hypothesis testing problem specified by (7.15), with
a/b indicating with/without noise whitening.
As we can see, the estimated values of nRHSP-ATGP without noise whitening
was 19, which is very close to the 18 reported in Chang et al. (2010a, 2011b), that
is, the number of endmembers. But VD with noise whitening is three or four,
which indicates that many signals were considered noise due to their weak signal
energies.
Another legitimate criterion by which to terminate RHSP-ATGP or ATGP is to
use unmixed error as a stopping rule where the fully constrained least-squares
(FCLS) method developed by Heinz and Chang (2001) was used to perform linear
spectral data unmixing. In this case, FCLS was performed using the targets gener-
ated by RHSP-ATGP for TI and TE in Fig. 7.3 and for the HYDICE and Cuprite
222 7 Recursive Hyperspectral Sample Processing of Automatic Target Generation Process
a b
9 unmixing error vs the number targets 8 unmixing error vs the number targets
3 x 10 8 x 10
7
2.5
6
2
5
unmixing error
unmixing error
1.5 4
3
1
2
0.5
1
0 0
0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200
# of targets # of targets
c d
7 unmixing error vs the number targets 9 unmixing error vs the number targets
12 x 10 9 x 10
8
10
7
8 6
unmixing error
unmixing error
5
6
4
4 3
2
2
1
0 0
0 20 40 60 80 100 120 140 160 180 0 20 40 60 80 100 120 140 160 180 200
# of targets # of targets
Fig. 7.7 Unmixing error vs. number of targets for HYDICE and Cuprite data sets. (a) TI (189), (b)
TE (189), (c) HYDICE data (115), and (d) Cuprite data (186)
data in Fig. 7.6 for data unmixing. Figure 7.7a–e plots the FCLS-unmixed errors for
the TI, TE, HYDICE, and Cuprite data sets versus the number of target signatures,
where the number of targets in parentheses is the one that yields the minimal
unmixed errors.
For TI and TE this number was 189, which means that the unmixed errors
for both TI and TE are strictly decreasing curves as the number of target
signatures grows. For HYDICE and Cuprite data, this number was 115 and
186, respectively.
It is worth noting that the unmixing error plot is different from that in Gao and
Chang (2014), where 117 yielded the minimal unmixed error (Fig. 7.8). This is
because the number of bands went from 1 to 169 here, compared to Gao and Chang
(2014), where it went from 9 to 169.
7.6 Computational Complexity of RHSP-ATGP 223
Fig. 7.8 FCLS unmixed Total error vs N using ATGP for testing purpose
x 106
error using 169 RHSP- 6
ATGP-generated targets
5
Total error
3
0
0 20 40 60 80 100 120 140 160 180
N
Fig. 7.9 Comparison of cumulative running times required for running ATGP and RHSP-ATGP
on TI, TE, HYDICE, and Cuprite data sets as the number of targets from one up to numbers
indicated on x-axis
computing time between ATGP and RHSP-ATGP, where the y-axis represents their
computing time in seconds and the x-axis indicates the Nth individual target. As we
can see, for any given target the computing time required by RHSP-ATGP is nearly
constant regardless of how many signatures are used. By contrast, ATGP gradually
increases its computing time as the Nth signature grows.
According to Figs. 7.9 and 7.10, RHSP-ATGP seemed to gain more benefits on
the HYDICE data than the other data sets. Nevertheless, RHSP-ATGP does have a
significant benefit in the reduction of computational complexity, which is a major
advantage for simple and cheap hardware design. An effort to implement RHSP-
ATGP in a FPGA is currently being investigated, which will certainly improve the
work in Bernabe et al. (2011, 2013).
7.7 Conclusions 225
Fig. 7.10 Comparison of running times required for running ATGP and RHSP-ATGP on TI, TE,
HYDICE, and Cuprite data sets for each individual target as the number of targets increase from
one up to numbers indicated on x-axis
7.7 Conclusions
ATGP has shown great potential in many applications in hyperspectral data exploi-
tation, for example, automatic target recognition, supervised subpixel detection/
classification, and endmember extraction. However, when the number of targets it
produces is very large, which is indeed the case in hyperspectral imaging, ATGP
becomes slow as a result of inverting its growing large target matrices performed by
a sequence of orthogonal subspace projections. To resolve this issue, this chapter
developed RHSP-ATGP to update targets found by ATGP via specially derived
recursive equations. As a result of such recursions, RHSP-ATGP works as if it were
a Kalman filter, described in Chap. 3, to perform real-time processing by only
updating the innovation information. Most importantly, requires no matrix inver-
sion, just matrix multiplications and outer products of vectors. Another significant
advantage of RHSP-ATGP over ATGP is that RHSP-ATGP can be terminated by
226 7 Recursive Hyperspectral Sample Processing of Automatic Target Generation Process
an automatic stopping rule in real time rather than a prescribed error threshold, as in
Ren and Chang (2003), via OPCI, which must be determined a priori. This stopping
rule is carried out by an ATGP-specified Neyman–Pearson detector to determine
when RHSP-ATGP should stop finding new targets as the process proceeds in
real time.
Chapter 8
Recursive Hyperspectral Sample Processing
of Orthogonal Subspace Projection
8.1 Introduction
Orthogonal subspace projection (OSP) has proven very effective and useful in
many applications in hyperspectral data exploitation, for example, subpixel target
detection, mixed pixel classification/quantification, and dimensionality reduction.
It can be implemented in two stages in a cascade, annihilation of undesired/
unwanted signal sources in the first stage, followed by the extraction of desired
signal sources via a matched filter in the second stage. While the matched filter can
be implemented to enhance signal detectability, it is the unwanted signal source
annihilation performed in the first stage that plays a key role in the success of target
extraction. In OSP the first stage process is specially designed to accomplish this
task by using an OSP to eliminate unwanted signatures in such a way that the
background can be suppressed to improve target contrast against the background.
Consequently, the effectiveness of background suppression has a significant impact
on the performance of the follow-up matched filter.
To be more specific, assume that there are p signal signatures, m1, m2, . . . , mp
and tp is the desired signal signature of interest to be extracted without loss of
generality. To effectively extract mp we need to annihilate unwanted signal signa-
tures m1 , m2 , . . . , mp1 prior to extraction of mp. To achieve this goal, we first
form U as an unwanted signal source matrix made up of all the unwanted signal
signatures, m1 , m2 , . . . , mp1 , denoted by Up1 ¼ m1 m2 mp1 . Then we
design an unwanted signature annihilator, P⊥ U , defined in (4.19) as
P⊥
Up1 ¼ I Up1 Up1 ;
#
ð8:1Þ
1
with U#p1 being the pseudo-inverse of Up1 given by U#p1 ¼ Up1
T
Up1 T
Up1
to annihilate the interfering effect of m1 , m2 , . . . , mp1 on the detection of mp.
Once these unwanted signal signatures are annihilated, a matched filter using the
matched signal signature tp is then applied in the following stage to extract mp.
Apparently, for OSP to work effectively, finding an appropriate set of unwanted
signal signatures for P⊥ Up1 to perform unwanted signature annihilation is crucial.
Unfortunately, when OSP is implemented, the p signal signatures m1, m2, . . . , mp
are assumed to be known in advance. To address this issue, an unsupervised version
of OSP, called unsupervised OSP (UOSP), was developed by Wang et al. (2002);
it can be considered an earlier version of the automatic target generation process
(ATGP) developed by Ren and Chang (2003) and is also discussed in Sect. 4.4.2.3.
It allows OSP to repeatedly find unwanted signal sources in an unsupervised
manner so as to achieve better background suppression prior to the extraction of a
desired target signal signature. In this case, UOSP can process OSPs with growing
signatures, as opposed to the Harsanyi–Chang OSP, which requires prior knowl-
edge about signatures at a fixed number of signatures, p. Although UOSP can find
unknown signal sources, it also assumes that the number of unknown signal sources
is known and fixed, as OSP does. In many practical applications, this number is
generally not known a priori. To address this issue, this chapter further extends
UOSP to progressive OSP (P-OSP), which can progressively generate a growing set
of unknown signal sources one after another directly from the data to be processed
without prior knowledge of this number. To effectively determine how many
unwanted signal signatures should be used in Up1, a real-time automatic stopping
rule is further developed. In other words, P-OSP can also automatically determine
8.2 Difference Between OSP and Linear Spectral Mixture Analysis 229
Although OSP can be used to perform linear spectral mixture analysis (LSMA) via
a signal detection model, it is very important to realize its salient differences from
LSMA. The first prominent difference is that OSP is developed to detect the
abundance fraction of a particular specified desired signal signature while making
use of prior knowledge of undesired signal signatures to annihilate their interfering
effects on the detection of the desired signal signature. Thus, from an unmixing
point of view, OSP uses its detected abundance fractions as unmixed results with no
abundance constraints. In order for OSP to perform LSMA to unmix data accu-
rately, it must satisfy two abundance constraints, the abundance sum-to-one con-
straint (ASC) and abundance nonnegativity constraint (ANC). However, since OSP
only detects the abundance fraction of each individual signature present in a data
230 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
sample vector independently, it does not satisfy ASC. In addition, its detected
abundance fraction may also be negative, so it does not satisfy ANC either.
Accordingly, OSP should not be considered an abundance-constrained LSMA
technique. Nevertheless, OSP can be considered an abundance-unconstrained
1
LSMA if an error-correction term, dT P⊥ Ud in (4.21), is included in its detector
to become a least-squares OSP (LSOSP) specified by (4.32) as an abundance
estimator.
As also noted in Chang and Heinz (2000), a fully abundance-constrained LSMA
may not be as effective as OSP in terms of signal detection, where the nonnegativity
constrained least-squares (NCLS) method performed better than the fully
constrained least-squares (FCLS) method in detecting signals because the NCLS
method does not satisfy ASC. In other words, to comply with ASC, no abundance
fraction of any signature can exceed one, and the abundance fractions of all
signatures must be constrained to a range between 0 and 1. Consequently, the
more signatures, the more abundance fractions that need to be constrained to the
range from 0 to 1, thus the fewer their abundance fractions. As a result, the detected
abundance fractions of signatures are diminishing. By contrast, OSP is a detection
technique specified by (4.20) which uses the signal-to-noise ratio (SNR) as an
optimal criterion, as opposed to the least squares error (LSE) used by LSMA to
perform abundance fraction estimation via (4.21) as an estimation technique.
Accordingly, OSP can enhance signal strength as much as it can for target
detectability without complying with ASC, regardless of how many signatures
are considered. Therefore, they are completely different theories but their connec-
tion is bridged by (4.20) and (4.21).
The idea of RHSP-OSP arises from the need to update the unwanted signature
matrix U for signature annihilation to improve OSP performance as more unwanted
signatures are found and added to U. As noted in (8.1), the impact caused by the
effect of U significantly impairs the detectability of the desired signature d. When
the environment is unknown, the signature matrix M should be adapted and updated
as data processing progresses. This is specifically critical in real-time processing. In
this case, P⊥U in (8.1) must be repeatedly implemented as U varies. From an
algorithmic implementation point of view, this is not effective for target detection
or for real-time implementation. Thus, it is highly desirable to have OSP
implemented adaptively in an unsupervised manner so that only the new signature
m added to the signature matrix U needs to be processed without having to
reimplement U over and over again. This is because the new augmented undesired
signature matrix [U m] includes only the new signature m, while other signatures
in U remain unchanged. This section presents a new look of OSP, RHSP-OSP, to
address this issue.
8.3 Recursive Hyperspectral Sample Processing of OSP 231
For simplicity, we assume that mp is the pth undesired signal signature to be added
to Up1 to form a new matrix, Up, and d is the desired signal signaturenot in U p.
Assume that Up is a matrix of size L p and can be written as Up ¼ Up1 mp ,
with Up1 and mp being an L ðp 1Þ matrix and an L-dimensional vector.
Then, according to a matrix inverse identity derived using Eq. (12.25) in Chang
(2003a, b), the inverse of UTp Up can be expressed as
" #1
h i1 T
Up1 Up1 UpT mp
UpT Up ¼
mpT UpmpT mp
2 1 T 3
T
Up1 Up1 þ β
U#p1 mp
6 p p1 p p1 7
¼6
4 T
7;
5
β
mpT U#p1
p p1 p p1
ð8:2Þ
1
where U#p1 ¼ Up1
T
Up1 T
Up1 and
1 1
β
1
Up U#p ¼ Up UpT Up UpT ¼ Up1 mp
2 1 T 3
T
Up1 Up1 þ β
U#p1 mp " U T #
6 p p1 p p1 7 p1
6
4 T
7
5
β
mpT U#p1 β
mpT
p p1 p p1
1 T
¼ Up1 Up1
T
Up1 T
Up1 þ β
T
mp mpT U#p1 Up1 β
þ β
mp mpT :
p p1
ð8:4Þ
232 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
1
e p p1 ¼ Up1 U#p1 mp , then P⊥
U
If we define m Up ¼ I U U#
p p ¼ I Up U T
U
p p UpT
becomes
1
Up1 T
P⊥
Up1 Up1
¼ T T
β e e e þ T
p
p1
Up I Up1 U U
p1 p1 U p1 m p m p 2 m p m p m m
p p
T
¼ P⊥ β
e
Up1
e
Up1
p
p1
Up1 m p m p m p m p
T
¼ P⊥
U
e p p1 mp m
U
e p p1 ;
Up1 βp
p1 mp m
ð8:5Þ
^ OSP
α p ðrÞ ¼d T
P⊥
Up r
Up1 T
¼ d T P⊥
Up1
β e e
p
p1
Up1 m p m p m p m p r ð8:6Þ
T
¼ d T P⊥
U
e p p1 mp m
U
e p p1 r:
Up1 r βp
p1 d mp m
T
In case there is another new set of undesired signal signatures, Uq, instead of a
single signal signature found, mp, to be added to the signal signature matrix, Up, the
resulting unwanted signal signature matrix becomes a combined signal signature
matrix, Up+q. This situation often occurs when RHSP-OSP is implemented as
RHSP-OSP, where more undesired signal signatures, instead of a single signal
signature, are found and need to be included for undesired signal signature annihi-
lation. For example, assume that p and q are two values of undesired signal
signatures found by different algorithms, in which case we can operate RHSP-
OSP on Up and Uq independently and individually, then fuse their corresponding
results without reprocessing all p + q signal signatures, Up+q. In this case, (8.4) can
be extended as follows:
" # " T #
UpT Up Up UpT Uq
T
Upþq Upþq ¼ Up Uq ¼ ; ð8:7Þ
T
Uq UqT Up UqT Uq
" T #!1
1 Up Up UpT Uq
T
Upþq Upþq ¼
UqT Up UqT Uq
2 1 T 3
ð8:8Þ
U T
6 p p U þ U #
U
p q ΓU T
q Up
#
U#p Uq Γ 7
¼6
4 T
7;
5
Γ T UqT U#p Γ
2 1 T 3
" #
6 UpT Up þ U#p Uq ΓUqT U#p U#p Uq Γ 7 UpT
Upþq U#pþq ¼ Up Uq 6
4 T
7
5
Γ T
UqT U#p Γ UqT
2 1 T 3
T
6 Up Up Up þ Up Uq ΓUq Up Up Up Uq ΓUq 7
T # T # T # T
¼ Up Uq 6
4 T
7
5
Γ Uq Up Up þ ΓUq
T T # T T
1 T
¼ Up UpT Up UpT þ Up U#p Uq ΓUqT U#p UpT Up U#p Uq ΓUqT
T
Uq Γ T UqT U#p UpT þ Uq ΓUqT :
ð8:9Þ
dT P⊥
Upþq r ¼ d I P
T
Upþq r ¼ d r d PUpþq r
T T
¼ dT r dT Upþq U#pþq r
T
¼ dT P⊥Up r d T
U p U#
p Uq ΓU T
q Up
#
UpT r þ dT Up U#p Uq ΓUqT r
T
þ dT Uq Γ T UqT U#p UpT r dT Uq ΓUqT r
¼ dT P⊥
Up r d PUp Uq ΓUq PUp r þ d PUp Uq ΓUq r
T T T T ð8:10Þ
þ dT Uq Γ T UqT PUp r dT Uq ΓUqT r
¼ dT P⊥
Up r þ d PUp Uq ΓUq I PUp r d Uq ΓUq I PUp r
T T T T
¼ dT P⊥
Up r þ d P
T
U ΓU T P⊥ r dT U ΓU T P⊥ r
T ⊥
Up q q Up T ⊥ q q Up
¼ d PUp r d I PUp Uq ΓUq PUp r
T
¼ dT P⊥ T ⊥ T ⊥
Up r d PUp Uq ΓUq PUp r:
As a special case, let Uq ¼ mp+1, and let d be the desired signal signature not in
Uq. Then Γ becomes β, (8.9) is reduced to (8.4), and (8.10) is reduced to (8.6):
T ⊥ T ⊥ T ⊥ ⊥
^ OSP
α pþ1 ðrÞ ¼ d PUpþ1 r ¼ d PUp r βd PUp mpþ1 mpþ1 PUp r
T
T ð8:11Þ
¼ dT P⊥ β
dT m T P⊥ T ⊥
pþ1
p
Up r pþ1 Up m P
pþ1 Up r;
1
where P⊥ Up ¼ I U U
p p
#
¼ I U p U T
U
p p UpT . It is worth noting that β
in
p p1
n h i o1
(8.3) can be further reexpressed as β
¼ mpþ1 T
P⊥
Up mpþ1 ¼
pþ1 p
h i T h i
1
P⊥
Up m pþ1 P ⊥
Up m pþ1 because P⊥ Up is symmetric and idempotent.
234 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
According to (8.5), P⊥ ⊥
Upþ1 can be updated via PUp by including a correction term of
T
β
m
Up Up
e pþ1 mpþ1 m e pþ1 mpþ1 recursively.
pþ1 p
Several comments on (8.11) are worthwhile.
U
1. me pþ1
p
¼ Up U#p mpþ1 is an innovation vector representing the leakage of the pth
⊥
Up
signal signature mp into Up1 . In this case, mpþ1 m e pþ1
T
Up
mpþ1 me pþ1 in (8.5), which is an outer product of the vector
U
mpþ1 m e pþ1
p
, can be considered an innovation information matrix. Thus, if
⊥
there is no leakage into Up , then U#p is the inverse of Up and Up U#p ¼ I is an
¼ mpþ1 . As a result, P⊥ ⊥
U
identity matrix, which results in m e pþ1
p
Upþ1 ¼ PUp
according to (8.5).
⊥
2. Since P⊥ Up projects the leakage of U p into its complement subspace Up ,
⊥
⊥
^ OSP
Using (8.12), α p ðrÞ in (8.6) can actually be calculated by
T
^ OSP
α T ⊥
dT P⊥ mpþ1 P⊥ mpþ1 r
pþ1 ðrÞ ¼ d PUp r β
pþ1
p Up Up
ð8:13Þ
¼α OSP
^ p ðrÞ β
d PUp mpþ1 mpþ1 P⊥
T ⊥ T
pþ1 p Up r;
T
where P⊥ ⊥
Up is idempotent and PUp ¼ P⊥
Up .
1. Initial condition:
p ¼ total number of signatures used to form a linear mixing model. Technically
speaking, its value can be determined at one’s discretion.
Start with the first signal signature, denoted by m1. Set U1 ¼ ½m1 , and calculate
P⊥U1 . Let j ¼ 2. 1
2. At the jth recursion use (8.2) and (8.5) to calculate UjT Uj and P⊥ Uj .
3. Stopping rule:
If a stopping rule is satisfied (to be discussed in Sect. 8.4.2), RUA is terminated.
Otherwise, let p p þ 1, and go to step 2.
8.4 Signature Generation by RHSP-OSP 235
^ OSP
Table 8.1 Computational complexity of (8.6) and (8.13) to calculate α p ð rÞ
1
Initial m1 m1T m1 , m1mT1 One inner product (L multiplications) plus one outer product of an
conditions L-dimensional vector, m1 (L2 multiplications)
T 1
P⊥
U1 P⊥ m1 m1T L subtractions and L multiplications
U1 ¼ I m1 m1
T ⊥ (L + 1) inner products of two L-dimensional vectors (L + L2
^ OSP
α 1 ð rÞ ^ OSP
α 1 ðrÞ ¼ d PU1 r
multiplications)
Available information: U L2 subtractions and L inner products of two L-dimensional vectors
m
e p p1 ¼ Up1 U#p1 mp
Up1 , U#p1 , mp, P⊥
Up1 U (L2 multiplications)
mp m
e p p1 ¼ P⊥
Up1 mp L inner product of L-dimensional vectors (L2 multiplications)
h i T h i 1
Innovation informa- L inner products of two L-dimensional vectors and one inner product
⊥
tion: βp|p1 by (8.3) β
¼ P⊥
Up1 mp PUp1 mp (L2 + L multiplications)
p p1
T U
Update P⊥
Up by (8.5) ⊥
U U One outer product of an L-dimensional vector mp me p p1
P⊥ e p p1 mp m
Up ¼ PUp1 β p
p1 mp m e p p1 2 2 2
(L multiplications) and L subtractions plus L multiplications
T ⊥ T ⊥ U 2
Update α ^ OSP
p ð rÞ ^ OSP
α ^ OSP
p ð rÞ ¼ α p1 ðrÞ βp
p1 d PUp1 mp mp PUp1 r One outer product of mp m e p p1 ¼ P⊥
Up1 mp (L multiplications) + L
by (8.6) inner products of two L-dimensional vectors and one inner product (L2 + L
multiplications) plus one subtraction and one multiplication
8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
8.4 Signature Generation by RHSP-OSP 237
However, RHSP-OSP also comes with two challenging issues: (1) how many target
signal sources does RHSP-OSP have to generate? (2) how does one find these signal
signatures? Each of these issues is discussed in what follows.
in the derivation of RHSP-OSP in (8.11). Now, for general purposes, we use target
signal sources t1, t2, . . ., tp to replace the unwanted or undesired signal signatures,
m1, m2, . . ., mp used in previous derivations. To find the next target signal source
tp+1, t1, t2, . . .,tp must be excluded from consideration. In this case,
Up ¼ t1 t2 tp forms an unwanted target signal matrix.
According to (8.11), the most unwanted target signal source that will be added to
create a new ( p + 1)-unwanted signature matrix Up+1 is the one that yields the
maximum innovation information. Since the matrix in (8.11) is the outer product of
the vector tpþ1 et pþ1
Up
, its rank is one, with only one nonzero eigenvalue given by its
inner product,
T
tpþ1 et pþ1 tpþ1 et pþ1
Up Up
: ð8:15Þ
In this case, (8.15) represents the innovation information provided by the correla-
tion between a new target signal source tp and previously generated target signal
sources, t1, t2, . . ., tp. Using (8.11), (8.15) can be further written as
T T
tpþ1 et pþ1 tpþ1 et pþ1 ¼ P⊥ P⊥ ⊥
Up Up
Up tpþ1 ¼ tpþ1 PUp tpþ1 ; ð8:16Þ
T
Up tpþ1
which suggests that the most unwanted new target signal source should be the one
that maximizes
2
r Up Þ ¼ r e
r Up
T
ρ r; Up ¼ ðr er U p Þ ðr e
T ð8:17Þ
¼ P⊥Up r P⊥ T ⊥
Up r ¼ r PUp r
In light of (8.16) and (8.17), we can design a criterion to find a potential target
signal source, tp+1, which is
n o
tpþ1 ¼ arg maxr ρ r; Up ¼ arg maxr rT P⊥
Up r : ð8:18Þ
Once a sequence of unwanted target signal sources {tp} is generated by (8.18), the
next challenging issue for RHSP-OSP is how to terminate RHSP-OSP. Using (8.16)
and (8.18) we can define
2
ηp ¼ tRHSP-OSP T P⊥ tRHSP-OSP ¼
P⊥ RHSP-OSP
: ð8:19Þ
p Up1 p Up1 tp
This ηp turns out to be exactly the same signal source under the binary hypothesis
testing problem considered in the maximal orthogonal subspace projection (MOSP)
in Chang et al. (2011c), where the energy of the pth target signal source tp,
tp
2 ,
was used to determine VD or the maximal orthogonal complement algorithm
(MOCA) in Kuybeda et al. (2007) was used to determine the rank of rare signal
dimensionality.
Following Sect. 4.5.1 of Chap. 4, we assume that for each 1 p L, U0 ¼ I is
the identity matrix and Up1 is the target space linearly spanned by previously
n op1
found p 1 target signal sources tRHSP-OSP
j by RHSP-OSP. Then, for 1 p
j¼1
L, we can find tRHSP - OSP
via (8.18) and ηp via (8.19), where ηp in (8.19) is the
p
maximum residual of the pth target signal source, tRHSP - OSP found by RHSP-OSP
p
⊥
leaked from < Up1 > into < Up1 > , which is
the
complement
T
space
orthogonal
to the space < Up1 >. It should be noted that
P⊥ U0 r
r r
in (8.19) when
n oL
RHSP-OSP
p ¼ 1. Since the found tp obtained by RHSP-OSP may be highly
p¼1
correlated, ηp in (8.19) is used instead because ηp represents the maximum residuals
of tRHSP - OSP leaked into < U >⊥ . It is this sequence, η L , that will be used as
p p1 p p¼1
the signal source in a binary composite hypothesis testing problem formulated as
follows to determine whether or not the lth potential target candidate, tpRHSP - OSP is a
true target signal source by a detector.
8.4 Signature Generation by RHSP-OSP 239
H0 : ηp p ηp
H 0 ¼ p0 ηp
versus for p ¼ 1, 2, . . . , L ; ð8:20Þ
H1 : ηp p ηp
H 1 ¼ p1 ηp
where the alternative hypothesis H1 and the null hypothesis H0 represent two cases
of tpRHSP - OSP being an endmember under H1 and not an endmember under H0,
respectively, in the sense that H0 represents the maximum residual resulting from
the background signal sources, while H1 represents the maximum residual leaked
from the target signal sources. Note that when p ¼ 0, η0 is undefined in (8.20). To
make (8.20) work, we need to find probability distributions under both hypotheses.
The assumption made on (8.20) is that if a signal source is not an endmember under
H0, then it should be considered part of the background, which can be characterized
by a Gaussian distribution. On the other hand, if a signal source is indeed an
endmember, it should be uniformly distributed over [0, ηp1]. By virtue of extreme
value
theory
(Leadbetter 1987), ηp can be modeled as a Gumbel distribution, that is,
Fvp ηp is the cumulative distribution function (cdf) of vp given by
( h i)
ð2logN Þ1=2 pffiffiffiffiffiffiffiffi
xσ 2 ðLpÞ
ð2logN Þ1=2 þ12ð2logN Þ1=2 ðloglogNþlog4π Þ
Fvp ðxÞ exp e σ 2 2ðLpÞ
: ð8:21Þ
h i
Since pνp ηp ¼ p H 0 ; ηp þ p H 1 ; ηp ¼ 1=ηp1 ηp pνp ηp þ Fνp ηp , we
can obtain an a posteriori probability distribution of p(H0|ηp) given by
ηp pνp ηp
p H 0
ηp ¼ ð8:24Þ
η p pν p η p þ Fν p η p
240 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
assumed to be a desired target signal source. Note that the test for (8.26) must be
performed for each of L potential target signal candidates. Therefore, for a different
value of p the threshold τp varies. Using (8.26) we can derive the OSP-specified VD,
nRHSP-ATGP, by calculating
n h io
nRHSP-OSP ¼ VDNP
RHSP-OSP ðP F Þ ¼ arg max p δ NP
RHSP-OSP η RHSPOSP
p ¼ 1 ;
ð8:27Þ
where PF is a predetermined false alarm probability, δNP - ηp ¼ 1 only if
NP
RHSP OSP
δRHSP-OSP ηp ¼ 1 and δRHSP-OSP ηp ¼ 0 if δRHSP-OSP ηp < 1.
NP NP
Note that the sequence of {ηRHSP - OSP } is monotonically decreasing. Thus, the
p
sequence {δNP (ηRHSP - OSP )} starts with a failure of the test (8.26), that is,
RHSP - OSP p
RHSP-OSP
RHSP-OSP ηp
δNP ¼ 1 , and continues as the value of p is increased until
p reaches the value where the test is passed, in which case
RHSP-OSP RHSP-OSP
δRHSP-OSP ηp
NP
< 1. The largest value of p makes δRHSP-OSP ηp
NP
¼ 1 the value of VDNP RHSP - OSP (PF) according to (8.27). This unique property allows
VDNP (P
RHSP - OSP F ) to be implemented in real time as the process continues with
increasing p.
8.4 Signature Generation by RHSP-OSP 241
1. Initial condition:
Find an initial target pixel vector m1 ¼ argfmaxr rT rg. Set U1 ¼ ½m1 , and
calculate P⊥U1 .
• Inner Loop
(a) Progressive process indexed by the parameter i (running through all
N
data sample vectors fri gi¼1 ):
For p 2, find mp by maximizing riT P⊥ Up1 ri via (8.19) over all data
n h i o1
sample vectors, ri, that is, β
¼ mpþ1 T
P⊥
Up mpþ1 ¼
pþ1 p
h i T h i
1
P⊥Up m pþ1 P ⊥
Up m pþ1 .
^ OSP
(b) Recursive process by updating α p ðrÞ:
^ OSP
Use (8.5) to update α ^ OSP
p ðrÞ via previously calculated α p1 ðrÞ, βp|p1,
U
e p p1 .
and m
• End (Inner Loop)
2. Stopping rule:
If a stopping rule is satisfied by (8.26) and (8.27), RHSP-OSP is terminated.
Otherwise, let p p þ 1, and go to step 1(a).
• End (Outer Loop)
Figure 8.1 presents a flowchart diagram illustrating the implementation of
RHSP-OSP, which consists of three processes, a progressive process used to
process data sample vectors sample by sample progressively to calculate maximal
value of riT P⊥ # ⊥
Up1 ri to find mp, a recursive process used to update Up and PUp , and a
L
successive process to grow target signal sources, mp p¼1, where a stopping rule is
the OSP-specified VD, nRHSP-ATGP estimated by the Neyman–Pearson detector in
(8.26), and VDNP RHSP - OSP (PF) in (8.27).
Since the previously described RHSP-OSP can be implemented in two loops, the
inner loop for data sample vectors in a progressive manner and the outer loop for
242 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
signal signatures in a successive manner, we can also interchange these two loops
by implementing the signal signatures in the inner loop and the data sample vectors
in the outer loop. The resulting RHSP-OSP is called causal RHSP-OSP (CRHSP-
OSP) and is described in what follows.
8.4 Signature Generation by RHSP-OSP 243
Causal RHSP-OSP
1. Initial condition:
N
Assume that there are N data sample vectors, fri gi¼1 .
N
2. Outer loop: iterate fri gi¼L by index k.
n oL
3. Inner loop: iterate tR-OSP ðr Þ
j i by index p.
j¼1
Use RHSP-OSP to iterate signal sources tjR - OSP (ri) from j ¼ 1 to pk, where pk is
k
determined by nRHSP-OSP via VDNP RHSP - OSP (PF) based on fri gi¼1 in (8.27).
The foregoing CRHSP-OSP has only theoretical interest but is not practical, for
two reasons. One is that the signal signature tj found by CRHSP-OSP in the inner
loop depends on the data sample vector ri. The other is that pk is determined by a
k
growing data sample pool, fri gi¼1 . Thus, the RHSP-OSP described in Fig. 8.1 is a
more practical version. Nevertheless, RHSP-OSP and CRHSP-OSP have their own
merits. For example, CRHSP-OSP iterates signal signatures in an inner loop that
can produce the appropriate number of signal signatures required for each data
sample to achieve its best possible performance in terms of detecting target signal
signatures of interest. It also utilizes an outer loop to find the number of signal
k
signatures for growing data sample pools fri gi¼1 to determine the best possible
number of signal signatures. An advantage of CRHSP-OSP is that it provides users
with a means of seeing different levels of detecting a particular target signature of
interest as each new data sample vector is fed in. On the other hand, RHSP-OSP
reverses two loops implemented in CRHSP-OSP by iterating all the data sample
vectors in an inner loop while iterating signal signatures in an outer loop. In this
case, RHSP-OSP allows users to see progressive abundance fraction detection maps
by adding one signal signature at a time. This is particularly useful because users
can see the detection performance progressively improved by adding more signal
signatures to the unwanted/undesired signal source matrix U.
A final note on RHSP-OSP is worthwhile. The quantity of ηp specified by (8.19)
is actually the same criterion specified by
T
tp ¼ arg maxr P⊥
Up1 r P ⊥
Up1 r ; ð8:28Þ
with U ¼ t1 t2 tp1 , called the orthogonal projection correlation index
(OPCI) by the automatic target detection and classification algorithm (ATDCA)
that was developed by Ren and Chang (2003) to find a new target signal source tp,
where ATDCA was subsequently renamed ATGP. With this interpretation RHSP-
OSP becomes RHSP-ATGP recently developed in Chap. 7. But it should also be
noted that the OPCI used by ATGP in Ren and Chang (2003) is exactly the same as
the ηp derived in (8.19). Figure 8.2 shows a flow diagram depicting relationships
among all variations derived from OSP.
244 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
OSP
ATGP
RHSP-ATGP RHSP-OSP
CRHSP-OSP
a b
p11, p12, p13
Fig. 8.3 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels
The image data to be studied are the HYperspectral Digital Imagery Collection
Experiment (HYDICE) image scene in Fig. 8.3a (and Fig. 1.10a), which has a size
of 64 64 pixel vectors with 15 panels in the scene and the ground truth map in
Fig. 8.3b (Fig. 1.10b).
It was acquired by 210 spectral bands with a spectral coverage from 0.4 to 2.5
μm. Low signal/high noise bands, bands 1–3 and 202–210, and water vapor
absorption bands, bands 101–112 and 137–153, were removed. Thus, a total of
169 bands were used in the experiments. The spatial resolution and spectral
resolution of this image scene are 1.56 m and 10 nm, respectively.
8.5 HYDICE Experiments 245
ηp ¼
tp
Table 8.2 tabulates the values of VDNP RHSP - OSP (PF), denoted by nRHSP-OSP, with
various false alarm probabilities, PF. Also included in Table 8.2 are the VD results
in Chang et al. (2014a) for comparison.
Interestingly, since ηp in (8.19) represents the maximal residual leakage of the
pth signal tpRHSP - OSP into the subspace P⊥ Up1 and ηp in Chang et al. (2014a) is indeed
R - OSP
the energy of the pth signal t p , they produced two ranges of values, from 19 to
21 and from 34 to 37 (Table 8.2), respectively, both of which have been used in the
literature for various applications. For example, it was shown in Heinz and Chang
(2001) and Chang (2003a, b) that 34 was an appropriate value to estimate the
number of signatures with target pixels found using the unsupervised FCLS
n o34
(UFCLS) method, tpUFCLS , for spectral unmixing. On the other hand, it was
p¼1
also shown in Chang et al. (2010a, 2011b) that a value of 18 was an appropriate
estimate for the number of signatures to use in unsupervised linear spectral
mixture analysis. These two sets
of
by ηp ¼ tp PUp1 tp and ηp ¼ tp
PF 102 for ηp ¼
tp
no
impact
on unmixed results. In addition, the values produced by means of ηp ¼
tp
2, to be
used as signal sources for (8.20), were in a range of 34–37, with PF 102 , which
are also indeed very close to the value of 34 shown in Heinz and Chang (2001).
Thus, in this case, values of 34–37 were used for experiments.
246 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
Since p13, p23, p33, p43, and p53 in the third column in Fig. 8.3b are subtarget
pixels, detecting these targets in an unsupervised fashion without prior knowledge
generally presents a challenge because their presence cannot be inspected visually.
Thus, our experiments were conducted for the detection of these five subpixel
panels. To see the causality detection of RHSP-OSP in finding abundance fractions
of subtarget pixels p13, p23, p33, p43, and p53 as the number of unwanted signals
sources, p, increases, Fig. 8.5a plots the abundance fractions of these five subtarget
n op
panels detected by CRHSP-OSP using targets tRHSP-OSP j ranging from 1 to
j¼1
p ¼ 169 found by RHSP-OSP; in addition, Fig. 8.5b is a zoomed-in plot of Fig. 8.5a
from p ¼ 1–37 for better visual assessment.
8.5 HYDICE Experiments 247
Detected abundances
abundance fractions of 6
CRHSP-OSP with
p ¼ 1–169; (b) zoomed-in 5
plots of detected abundance 4
fractions of CRHSP-OSP
with p ¼ 1–169 3
2
1
0
-1
0 20 40 60 80 100 120 140 160 180
Number of signal sources
b
106
9
p13 p43
8
p23 p53
7 p33
6
Detected fraction
5
4
3
2
1
0
-1
0 5 10 15 20 25 30 35
Number of signal sources
According to Fig. 8.5a, b, the detected abundance fractions of all subpixel panels
at p ¼ 6, 9, 18 showed three clear cuts for sudden drops, where the dashed lines are
used to highlight p ¼ 6, 9, 18, 19. Since the HYDICE data have five distinct panel
signatures plus a background signature; the first cut at p ¼ 6 is a prior estimate using
the ground truth. Without using prior knowledge, this scene was estimated by VD
as nine calculated by the HFC method (Harsanyi et al. 1994a, b; Chang 2003a, b),
248 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
which corresponds to the second cut. The value of nRHSP-OSP ¼ 19 was estimated by
an orthogonal complement maximal residual via (8.19), which is very close to the
third cut at 18. This shows that nRHSP-OSP ¼ 19 is indeed a very good estimate for
VD since after the third cut all the detected abundance fractions of the five subpixel
panels were saturated and flat. Figure 8.6a, b shows the plots of the abundance
fractions of all five subpixel panels estimated by CRHSP-LSOSP with p ¼ 1–169
and zoomed-in plots with p ¼ 1–37. Note that, unlike the very high RHSP-OSP-
detected abundance fractions produced in Fig. 8.5a, b, to enhance signal detectabil-
ity, Fig. 8.6a, b tried to estimate the abundance fractions, in which case the
abundance fractions were relatively small within a range between 0.1 and 0.5.
The implemented CRHSP-LSOSP is simply the same as RHSP-OSP, except that
OSP, α ^ OSP
p ðrÞ, in (4.20) used for abundance detection was replaced by LSOSP,
^ pLSOSP ðrÞ , in (4.21) and used for abundance estimation, which includes
α
T ⊥ 1
d PU d to account for estimation error.
Like Figs. 8.5a, b, 8.6a, b also showed clear cuts at p ¼ 6, 9, and 18. But unlike
Fig. 8.5a, b, where the abundance fractions approached zero, the estimated abun-
dance fractions in Fig. 8.6a, b fluctuated after p ¼ 18. It has been shown that for
supervised LSMA for this HYDICE scene, p ¼ 9 was estimated by VD in Chang
(2003a, b, 2013) and Chang and Du (2004), p ¼ 18 was estimated for unsupervised
LSMA (Chang et al. 2010a, 2011b), p ¼ 19 was estimated by nRHSP-OSP, p ¼ 29 was
estimated using an approach called RHSP-specified VD in Chang et al. (2014b),
and p ¼ 34 was estimated for the UFCLS method in Heinz and Chang (2001). In this
case, Fig. 8.6b dictates changes in the estimated abundance fractions, where the
dashed lines are used to highlight p ¼ 9, 18, 19, 29, 34–37.
The results shown in Figs. 8.5 and 8.6 are RHSP-OSP-detected and RHSP-
LSOSP-estimated abundance fractions of five single subtarget panel pixels pro-
duced by CRHSP-OSP and CRHSP-LSOSP. The figures do not show RHSP-OSP-
detection maps of all panel pixels and their estimated abundance fractions. It is
interesting to see the progressive RHSP-OSP-detection maps produced by RHSP-
OSP as p varies from 1 to 169, along with their estimated abundance fractions, for
comparison. In this way, RHSP-OSP must generate 169 detection maps showing
progressive changes in the detection of abundance fractions. As noted in Figs. 8.5
and 8.6, the only cases of interest are p ¼ 9, 18, 19, 29, 34–37, and 169. Therefore,
we only need to show progressive detection maps and their estimated abundance
fractions for these cases. The recursive equations derived in the appendix allow us
to accomplish this task without going through varying p one by one. Figures 8.7 and
8.8 show the RHSP-OSP-detected and RHSP-LSOSP-estimated abundance frac-
tional maps of 19 R panel pixel targets produced by RHSP-OSP, with p ¼ 9 and
q ranging from 9 ( p + q ¼ 18) in Chang et al. (2010a, 2011b), to q ¼ 10 ( p
+ q ¼ 19 ¼ nRHSP-OSP), q ¼ 20 ( p + q ¼ 29) in Chen et al. (2014a, b, c), and q ¼ 25
( p + q ¼ 34) in Heinz and Chang (2001), and full bands, 169. Since the RHSP-OSP-
detected abundances were so high, their detection maps are plotted in decibel (db)
defined by 20 log10x. Even in doing so, the detected abundance amounts of detec-
tion maps were still very bright compared to those in Fig. 8.8, which are abundance
8.5 HYDICE Experiments 249
a
0.6
p13 p43
0.5 p23 p53
p33
0.4
Estimated abundances
0.3
0.2
0.1
0.1
0 20 40 60 80 100 120 140 160 180
Number of signal sources
b
0.4
0.35
0.3
p13
Estimated fraction
0.25 p43
p23 p53
0.2
p33
0.15
0.1
0.05
-0.05
0 5 10 15 20 25 30 35
Number of signal sources
Fig. 8.6 Abundance fractions of five subtarget pixels estimated by CRHSP-LSOSP from number
of signal sources growing from 1 to 37. (a) Estimated abundance fractions by CRHSP-LSOSP with
p ¼ 1–169; (b) zoomed-in plot of estimated abundance fractions of CRHSP-LSOSP with p ¼ 1–37
250 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
Fig. 8.7 Progressive RHSP-OSP-detected fractional maps using targets produced by RHSP-OSP
as p grows from 9, 18, 19, 29, 34, 35, 36, 37, and 169; (a) 9 signatures; (b) 18 signatures; (c)
19 signatures; (d) 29 signatures of RHSP-OSP; (e) 34 signatures; (f) 35 signatures; (g) 36 signa-
tures; (h) 37 signatures; (i) 169 signatures
8.5 HYDICE Experiments 251
fractions estimated by RHSP-LSOSP for true abundances within a range from 0.1
to 0.5, in which case the background was largely suppressed to yield better visual
inspection. Nevertheless, this by no means implies that RHSP-OSP-detected maps
were worse than RHSP-LSOSP-estimated maps. If we examine carefully the results
in Fig. 8.7, we can also find that the five rows of panels, including the five subpixel
panels in the third column, were actually detected effectively.
One note is worthwhile here. According to the ground truth, the panels in rows
4 and 5 were made of the same materials with slightly different colors of paint. As a
consequence, a technique to detect panels in row 4 would also be able to detect
panels in row 5 and vice versa. This phenomenon was confirmed by Figs. 8.7 and
8.8. Similarly, it was also applied to the panels in rows 2 and 3, which were also
made of the same materials with slightly different paint colors.
Interestingly, the good results in Figs. 8.7 and 8.8 were the same as those
produced by p ¼ 18 (Chang et al. 2010a, 2011b), nRHSP-OSP ¼ 19 and 29 (Chang
et al. 2014b), and then performance seemed to deteriorate after p ¼ 34 (Heinz and
Chang 2001), with the worst result given by p ¼ 169. This made sense because the
more signal sources generated by RHSP-OSP, the more annihilation due to the fact
that OSP uses P⊥ U in (8.1) to annihilate unwanted signal sources, which may include
signal sources similar to the signal sources one wishes to detect, for example, the
panels in rows 2 and 3 and those in rows 4 and 8. As a result, the detectability of the
desired signal sources was largely degraded and reduced by annihilation of similar
signal sources in U. But this did not occur for data unmixing via fully abundance-
252 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
Fig. 8.8 Progressive RHSP-LSOSP-detected abundance fractional maps using targets produced
by RHSP-LSOSP as p grows from 9, 18, 19, 29, 34, 35, 36, 37, and 169; (a) 9 signatures; (b)
18 signatures; (c) 19 signatures; (d) 29 signatures of RLSOSP; (e) 34 signatures; (f) 35 signatures;
(g) 36 signatures; (h) 37 signatures; (i) 169 signatures
constrained methods, which are designed to estimate rather than detect abundance
fractions. To verify this assertion, Fig. 8.9 shows the abundance fractions of the
same five subtarget panels in the third column produced in Fig. 8.7, where the FCLS
method in Heinz and Chang (2001) was used to perform linear spectral unmixing,
8.5 HYDICE Experiments 253
instead of using OSP to perform signal detection, as is done in Fig. 8.7 for panel
pixels. According to Fig. 8.9d, h, the FCLS-unmixed results were indeed very good
and very close after p + q ¼ 18, 19. Comparing the results to those in Fig. 8.8
produced by RHSP-LSOSP, it was demonstrated that RHSP-LSOSP-generated
target signal sources could also be used as unsupervised signal sources to perform
unsupervised LSMA for data unmixing as an alternative to unsupervised least
squares–based fully constrained LSMA (Chang et al. 2010a) and component
analysis–based fully constrained LSMA (Chang et al. 2011b).
As a final comment on the implementation of RHSP-OSP, it should be kept in
mind that the plots in Figs. 8.5 and 8.6 were obtained by CRHSP-OSP and CRHSP-
LSOSP for a single panel pixel, with p as a parameter varying from 1 to 169, while
Figs. 8.7 and 8.8 show progressive RHSP-OSP-detection maps and RHSP-LSOSP-
estimation maps, respectively, with p a fixed value for all data sample vectors,
N
fri gi¼1 .
254 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
Fig. 8.9 Progressive FCLS-unmixed abundance fractional maps using targets produced by RHSP-
OSP as p grows from 9, 19, 29, 34, 119, and 169; (a) 9 signatures; (b) 18 signatures; (c)
19 signatures; (d) 29 signatures; (e) 34 signatures; (f) 35 signatures; (g) 36 signatures; (h)
37 signatures; (i) 169 signatures
8.6 Computational Complexity Analysis 255
Figure 8.10a, b plots the computing times required by OSP and RHSP-OSP as
p grows from 1 to 35 and from 1 to 169, respectively, where the x-axis indicates that
p is the cumulative number of signatures used by OSP to account for all signatures
to be used to form the signature matrix. Figure 8.10a is simply a zoomed-in version
of the plots in Fig. 8.10b, particularly for p from1 to 35 for better visual assessment.
As seen in Fig. 8.10a, RHSP-OSP does not have significant computing time savings
compared to OSP when the cumulative number of signatures, p, is small. However,
as the number of signatures, p, grows large, the savings began to emerge, as shown
in Fig. 8.10b, where RHSP-OSP required nearly constant computing time to process
each signal source compared to OSP, whose computing time grew exponentially
with the value of p. This is not a surprise from a signal processing point of view,
such as Kalman filtering (Poor 1994), because only the new incoming signature,
which is the data sample currently being visited, is required for RHSP-OSP to be
256 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
a
x 10
6
4
Computing
1 RHSP-OSP
OSP
0
0 5 10 15 20 25 30 35
Number of signal sources
b
0.01
0.009
0.008
0.007
Computing time(s)
0.006
0.005
0.004
0.003
0.002
0.001 RHSP-OSP
OSP
0
0 20 40 60 80 100 120 140 160 180
Number of signal sources
Fig. 8.10 Comparative performance of computing times required by OSP and RHSP-OSP. (a, b)
Computing time
processed, regardless of how many signatures are already being processed, com-
pared to OSP, which requires reprocessing of all signatures when a new signature is
included in a new signature matrix in which case computing time is increased
exponentially as the number of signatures grows.
8.7 Conclusions 257
8.7 Conclusions
The OSP developed by Harsanyi and Chang (1994) has produced promising results
in hyperspectral data exploitation. Its potential remains underexplored. This chapter
revisits the Harsanyi–Chang OSP and further develops a recursive version of OSP,
RHSP-OSP, that can implement OSP without inverting matrices as OSP does. As a
result, its computational complexity is significantly reduced. Such recursiveness
allows users not only to find unknown target signal sources in an unsupervised
manner but also to determine the number of signal sources as signatures, p, that
must be found at the same time. With RHSP-OSP there is no need to know the value
of p beforehand because its value can be determined by its ongoing process. Since
RHSP-OSP also iterates the signal signatures to be found in the outer loop, while
iterating the data sample vectors in the inner loop, we can derive a causal version of
258 8 Recursive Hyperspectral Sample Processing of Orthogonal Subspace Projection
subspace approach to accomplish what RHSP-OSP does. In this case, GSOP can be
considered a simple special case of RHSP-OSP.
Interestingly, we can also follow a similar idea used to develop causal iterative
pixel purity index (C-IPPI) in Chap. 8 and progressive iterative pixel purity index
(P-IPPI) in Chap. 12 (Chang 2016) to implement RHSP-OSP in a two-loop process:
(1) finding new signal signatures one at a time by P-OSP and (2) updating P⊥ Up1 in
(8.1) recursively by RHSP-OSP. Depending on how these two loops are designed
and implemented, two different versions can be derived for RHSP-OSP. One is
called causal RHSP-OSP (CRHSP-OSP), which iterates signal signatures in an
inner loop while iterating data sample vectors in an outer loop. The other is called
progressive hyperspectral sample processing of OSP (PHSP-OSP), which iterates
data sample vectors in an inner loop while iterating signal sources in an outer loop.
Both CRHSP-OSP and PHSP-OSP implement their inner and outer loops in the
complete opposite order. As a result of CRHSP-OSP, each data sample vector r can
produce its own appropriate number of undesired signal signatures for OSP to find
abundance fractions of a desired signal signature present in that particular data
sample vector r. On the other hand, when PHSP-OSP is implemented, we see
abundance fractions of a particular desired signal signature present in each data
sample vector progressively as new signatures are added for annihilation. In
addition, PHSP-OSP also provides a signature-varying progressive profile of
detected abundance fractions for a particular desired signal signature when each
new undesired signal signature is added. In both cases CRHSP-OSP and PHSP-OSP
offer advantages of hardware implementation in chip design such as FPGA design
(Chang and Wang 2008).
There are several benefits. First, RHSP-OSP allows OSP to be implemented with
varying signatures via a recursive equation without repeatedly reinverting
undesired signature matrices. In this case, RHSP-OSP can be carried out in a
manner similar to that of a Kalman filter (Chap. 3) by only retaining the innovation
information to update results without reprocessing already known information,
including data sample vectors already visited. As a result, the computer processing
time required by RHSP-OSP is nearly constant regardless of how many targets need
to be processed. This significant advantage makes RHSP-OSP very attractive and
effective. Second, RHSP-OSP can also be used to find new unknown signal sources
recursively while simultaneously determining a desired number of signal sources
while data processing is happening. A similar idea is also used by RHSP-ATGP in
Chap. 7. As a result, with appropriate interpretations, RHSP-ATGP can be consid-
ered a special case of RHSP-OSP. Third, for practical applications RHSP-OSP can
also be expanded in two different ways, to a causal process and to a progressive
process, which give rise to causal RHSP-OSP and progressive RHSP-OSP, respec-
tively, both of which can be easily realized by hardware design. Finally, in analogy
with RHSP-ATGP, RHSP-OSP also provides a real-time stopping rule via a
Neyman–Pearson detector to determine whether each target that is newly generated
by RHSP-OSP is a target while it is being generated.
Chapter 9
Recursive Hyperspectral Sample Processing
of Linear Spectral Mixture Analysis
Abstract In linear spectral mixture analysis (LSMA) the signatures used to form a
linear mixing model (LMM) have a significant role and impact in unmixing
performance but are generally assumed to be fixed and known a priori. However,
in real-world applications, such prior knowledge is either very difficult if not
impossible to obtain or unreliable even if it is known. Consequently, a realistic
solution to this issue is to adapt signatures while LSMA is taking place. Another
issue is that according to the general consensus, the signatures used to perform
LSMA should be endmembers. As noted in Chaps. 4 and 9 in Chang (Real time
progressive hyperspectral image processing: endmember finding and anomaly
detection, Springer, New York, 2016), this is usually not true. In other words, the
signatures used to from LMM should be those that can best represent the data to be
processed but are not necessarily endmembers as pure signatures. Specifically,
background signatures which are generally mixed are considered to be the most
integral part of data and should be included for data unmixing. To address the
aforementioned issues, this chapter introduces a new concept called virtual signa-
tures (VSs), which is defined as spectrally distinct signatures that are different from
endmembers as pure signatures in the sense that VSs ought to be extracted directly
from the data themselves and cannot be from spectral libraries or databases or
outside the data space. Furthermore, VSs are also different from virtual
endmembers (VEs) introduced in nonlinear spectral mixture analysis according to
a bilinear model. With this definition, finding and adapting VSs is a very challeng-
ing issue because these unknown VSs must be found in an unsupervised manner
without prior knowledge. In addition, an associated issue is the need to determine
how to terminate the VS finding process once an unsupervised VS finding algorithm
is initiated. To accomplish this task, this chapter develops a theory of adaptive
recursive hyperspectral sample processing of LSMA (ARHSP-LSMA). It first
develops a recursive hyperspectral sample processing of LSMA (RHSP-LSMA),
which allows LSMA to be implemented progressively, by growing VSs one at a
time to form an adaptive linear mixing model (ALMM) that makes the LMM adapt
to signatures, as well as recursive by taking advantage of previously known
signatures to update LMM. In this way, LSMA does not necessarily have to be
reimplemented over and over again during data processing. This allows LSMA to
perform data unmixing at the same time that ALMM is being updated. The resulting
LSMA is called ARHSP-LSMA. Most importantly, one great benefit of ARHSP-
LSMA is that it can be used to fuse LSMA results obtained by multiple different
sets of signatures without going through the entire signature set to perform LSMA.
This is particularly useful in data communication and transmission when LSMA
can be processed at different locations by different users independently and sepa-
rately, and then these LSMA results can be further fused by ARHSP-LSMA without
performing LSMA using the complete set of signatures. Finally, it takes advantage
of the concept of target-specified virtual dimensionality developed in Chap. 4 to
determine when ARHSP-LSMA should be terminated automatically by a Neyman–
Pearson detector (NPD).
9.1 Introduction
Assume that any data sample vector r can be represented by a linear admixture of a
finite number of signatures, say, p signatures, m1, m2, . . ., mp by which r can be
unmixed into their corresponding abundance fractions of α1, α2, . . ., αp as an abun-
T
dance vector, αp ¼ α1 ; α2 ; . . . ; αp . In recent years these p signatures have been
assumed to be known and fixed a priori. Such linear spectral mixture analysis
(LSMA) is generally referred to as supervised LSMA (SLSMA). Unfortunately, in
practical applications, such prior signature knowledge is generally unobtainable.
Under this circumstance developing an unsupervised LSMA (ULSMA) is necessary
without appealing for prior knowledge of signatures that are used to form a linear
mixing model (LMM). However, two major challenging issues arise in ULSMA.
The first issue has to do with determining the number of signatures, p, followed by
the second issue, which is that the p signatures, m1, m2, . . ., mp used to form an
LMM must be found in a completely unsupervised manner. To determine an
appropriate value of p, the concept of virtual dimensionality (VD), recently devel-
oped in Chap. 17 (Chang 2003) and Chap. 5 (Chang 2013) and Chang and Du
(2004), can be used. As for finding m1, m2, . . ., mp a general approach is to assume
that these p signatures are endmembers, that is, pure signatures, so that many
endmember extraction algorithms (EEAs) developed in the literature can be used
to extract endmembers. Unfortunately, this assumption is generally not true. First,
the background is considered an integral part of data. Thus, an LMM should include
background signatures to faithfully represent data. Practically speaking, many
background signatures are indeed not pure. Second, even though endmembers
exist in the data, these endmembers may very well be contaminated by other
signatures or corrupted by noise in real-world environments. Therefore, the pres-
ence of endmembers is not guaranteed. In both cases, extracting something that is
not present in the data is meaningless. Interestingly, many EEAs that reportedly
extract endmembers actually find endmember candidates, not true endmembers,
because true endmembers may not be present in the data. A good example is the
N-finder algorithm (N-FINDR) (Winter 1999a, b), which was originally developed
to find endmembers, not extract them. It has been misinterpreted in the literature as
9.1 Introduction 263
an EEA. Third, the concept of VD, originally developed in Chang (2003) and
Chang and Du (2004), was to determine the number of spectrally distinct signa-
tures, not endmembers, present in the data. Accordingly, using VD to determine the
number of endmembers is misleading. This chapter clarifies this issue by introduc-
ing the new concept of virtual signature (VS), which was previously used in
unsupervised LSMA (Chang et al. 2010a, 2011b) to define a VS as a spectrally
distinct signature that is a real data sample vector and must be present in data. With
this clear definition a VS distinguishes itself from an endmember in two respects.
One is that a VS is not necessarily pure like an endmember must be. In other words,
a VS can be a mixed signature. Another difference is that the existence of VS in the
data is guaranteed. This is unlike endmembers, which may not exist in real data
owing to signature purity. More specifically, a real endmember extracted directly
from data does not imply that it is in fact a true endmember. Previously, to address
this issue, the term virtual endmember (VE) was used in Tompkins et al. (1997) and
Bowles and Gilles (2007) to distinguish it from true endmembers. With this
interpretation endmembers extracted by an EEA in this chapter should be treated
as VEs. However, it is very unfortunate that the same term, VE, was also adopted in
several reports in nonlinear spectral mixture analysis (Chen et al. 2011), where VEs
are defined as relevant endmembers resulting from multiple-scattering effects in
each pixel and can be described by their product terms according to a bilinear
model. Such VEs can be highly correlated, as opposed to true endmembers, which
are generally spectrally distinct. Thus, to avoid any controversy and confusion, VS
is specifically defined and used in this chapter to differentiate VSs from VEs and
endmembers.
One of major applications resulting from LSMA is linear spectral unmixing
(LSU), which is probably one of the most studied subjects in hyperspectral data
exploitation (Chang 2003, 2013). When LSMA is implemented, it generally
assumes that signatures used to form an LMM for mixture analysis must be
known and fixed during data processing. In real practical problems obtaining
such prior knowledge is usually very costly or may not even be reliable owing to
many unknown subtle substances present in the data. To resolve this dilemma, one
feasible approach is to implement LSMA in an unsupervised manner, such as
ULSMA developed by Chang et al. (2010, 2011b) and Chang (2013), where
signatures can be obtained directly from the data being processed by an
unsupervised target detection algorithm. This approach does not address two
major issues: determining the number of signatures, p, required to form an LMM
and finding signatures once the value of p is determined. In ULSMA, the first issue
was resolved by virtual dimensionality (VD), recently developed by Chang (2003)
and Chang and Du (2004), while the second issue was resolved by least squares–
based unsupervised target finding algorithms proposed in Chang et al. (2010a) and
component analysis in Chang et al. (2011b). However, such ULSMA also fixes the
number of signatures during data processing once these unsupervised signatures are
found. In other words, this ULSMA cannot adapt to signatures once they are found
and cannot vary but must be fixed. More specifically, if signatures are found to be
264 9 Recursive Hyperspectral Sample Processing of Linear Spectral Mixture Analysis
9.2 LSMA
LSMA is quite different from the orthogonal subspace projection (OSP) presented
in Sect. 4.4.2.1 in two crucial respects. First and foremost is the performance
criteria they use. OSP is designed based on signal-to-noise ratio (SNR), whereas
LSMA is derived from least-squares error (LSE). As a result, OSP detects signal
strength compared to LSMA, which estimates signal parameters. Second, LSMA is
a vector estimation technique that can estimate signal abundance fractions simul-
taneously. Thus, it can be used for data unmixing by imposing abundance con-
straints, abundance sum-to-one constraint (ASC) and abundance nonnegativity
constraint (ANC). However, since OSP performs detection for signals one at a
time, ASC cannot be directly applied to OSP unless ASC must be adaptively
applied to each signal it detects consecutively. Therefore, in essence, OSP and
LSMA can be considered two completely different analysis concepts.
Assume that L is the number of spectral bands and r is an L-dimensional image
pixel vector. Assume that there are p material substance signatures, m1, m2, . . ., mp,
which are generally referred to as digital numbers (DNs). A linear mixture model of
r models the spectral signature of r as a linear combination of m1, m2, . . ., mp with
appropriate abundance fractions specified by α1, α2, . . ., αp. More precisely, r is an
L 1 column vector and
M is an L p substance spectral signature matrix, denoted
by m1 m2 mp , where mj is an L 1 column vector represented by the
spectral signature of the jth substance tj resident in the pixel vector r. Let α ¼
T
α1 ; α2 ; . . . ; αp be a p 1 abundance column vector associated with r, where αj
denotes the abundance fraction of the jth substance signature mj present in the pixel
vector r. To restore the pixel vector r, we assume that the spectral signature of the
pixel vector r can be represented by an LMM by m1, m2, . . . , mp as follows:
r ¼ Mα þ n; ð9:1Þ
The simplest way to impose a constraint on (1) is ASC. In other words, a sum-to-one
constrained least-squares (SCLS) problem can be cast as follows
n o Xp
minα ðr MαÞT ðr MαÞ subject to j¼1
αj ¼ 1: ð9:3Þ
To solve (9.3), we use the Lagrange multiplier λ1 to constrain the ASC, 1Tα ¼ 1, by
1
J ðα; λ1 Þ ¼ ðMα rÞT ðMα rÞ þ λ1 1T α 1 ; ð9:4Þ
2
0 1T
where 1 is the unity vector given by 1 ¼ @1, 1, . . . , 1 A . Differentiating (9.4) with
|fflfflfflfflfflffl{zfflfflfflfflfflffl}
p
respect to α and λ1 yields
SCLS
∂J ðα;λÞ
∂α SCLS
^
¼ MT M α ðrÞ MT r þ λ*1 1 ¼ 0
^
α ð rÞ
1 1 ð9:5Þ
^
)α ðrÞ ¼ MT M MT r λ*1 MT M 1
SCLS
1
^ SCLS ðrÞ ¼ α
)α ^ LS ðrÞ λ*1 MT M 1
and
∂J ðα; λ1 Þ
^ LS ðrÞ 1 ¼ 0:
¼ 1T α ð9:6Þ
∂λ1 λ*
1
Note that (9.5) and (9.6) must be solved simultaneously in such a way that both
^ SCLS ðrÞ and λ1 appear in (9.5).
the optimal solutions, α
Using (9.6) further implies
1
^ SCLS ðrÞ λ*1 1T MT M 1 ¼ 1
1T α
h 1 i1 ð9:7Þ
) λ*1 ¼ 1T MT M 1 ^ SCLS ðrÞ :
1 1T α
9.2 LSMA 267
where
T 1 h T T 1 i1 T
P⊥
M, 1 ¼ I M M 1 1 M M 1 1 : ð9:9Þ
Since there is no physical constraint imposed on the solution α ^ LS ðrÞ, its generated
abundance fractions α ^ LS
1 ðrÞ, α ^ LS ^ LS
2 ðrÞ, . . . , α p ðrÞ may be negative. To avoid this
situation, an ANC must be imposed on (9.1). A general approach to solving such an
ANC-constrained problem is to introduce an LSE-based objective function imposed
by the ANC for j 2 f1; 2; . . . ; pg, defined as
which implies
T 1 *
M M λ ¼α ^ LS ðrÞ α^ NCLS ðrÞ
NCLS
) λ* ¼ MT M α ^ LS ðrÞ MT M α^ ðrÞ
h i NCLS ð9:12Þ
1
) λ* ¼ MT M MT M MT r MT M α ^ ðrÞ
) λ* ¼ MT r MT M^
α NCLS ðrÞ:
268 9 Recursive Hyperspectral Sample Processing of Linear Spectral Mixture Analysis
λi ¼ 0, i 2 P,
ð9:13Þ
λi < 0, i 2 R;
where P and R denote passive and active sets that contain indices representing
negative or positive abundances respectively. By virtue of (9.11)–(9.13), a numer-
ical algorithm, referred to as a nonnegativity constrained least-squares (NCLS)
algorithm, can be designed to start off with the initial estimate given by α ^ LS ðrÞ in
(9.2). If all abundance fractions are positive, the NCLS algorithm stops. Otherwise,
all indices corresponding to negative and zero abundance fractions will be moved to
passive set P and all positive abundance indices to R. According to Kuhn–Tucker
conditions (9.13), any λi with index i 2 P is set to zero, and other indices are
calculated based on (9.12). If all λi are negative, then the NCLS algorithm stops.
If not, the corresponding most negative index is moved from R to P. A new vector λ
is then recalculated based on the modified index sets, and a new Lagrange multi-
plier vector is further implemented to find a new set of abundance fractions. Then
any negative abundance indices are shuffled from P to R. By iteratively
implementing (9.11) and (9.12), using α ^ LS ðrÞ in (9.2) as an initial abundance
vector, the NCLS algorithm can be derived to find an optimal solution, α ^ NCLS ðrÞ
(Chang and Heinz 2000). A detailed step-by-step implementation of the NCLS
algorithm is given in what follows.
NCLS Algorithm
1. Initialization: Set the passive set Pð0Þ ¼ f1; 2; . . . ; pg, Rð0Þ ¼ ∅, and k ¼ 0.
2. Compute α ^ LS ðrÞ by (9.2), and let α ^ NCLS, ðkÞ ðrÞ ¼ αLS ðrÞ.
3. At the kth iteration, if all components in α ^ NCLS, ðkÞ ðrÞ are nonnegative, the
algorithm is terminated. Otherwise, continue.
4. Let k ¼ k þ 1.
5. Move all indices in P(k1) that correspond to negative components of
^ NCLS, ðk1Þ ðrÞ to R(k1), respectively. Create a new index S(k), and set it
α
equal to R(k).
6. Let α^ RðkÞ denote the vector consisting of all components of α ^ NCLS, ðk1Þ ðrÞ in R(k).
1
7. Form a matrix ΦðkÞ by deleting all rows and columns in the matrix MT M
that are specified by P(k).
ðkÞ 1
8. Calculate λðkÞ ¼ Φα ^ RðkÞ . If all components in λ(k) are negative, go to
α
step 13. Otherwise, continue.
n o
ðkÞ
9. Calculate λðmax
kÞ
¼ arg maxj λj , and move the index in R(k) that corresponds to
ðkÞ
λmax to P(k).
9.2 LSMA 269
ðkÞ 1
10. Form another matrix Ψ λ by deleting every column of MT M that is
specified by P(k).
^ NCLS, ðkÞ Ψ ðλkÞ λðkÞ .
^ S ðk Þ ¼ α
11. Set α
12. If any components of α ^ SðkÞ in S(k) are negative, then move these components
(k) (k)
from P to R . Go to step 6.
ðkÞ 1
13. Form another matrix Ψ λ by deleting every column of MT M that is
specified by P(k).
^ NCLS, ðkÞ ¼ α
14. Set α ^ LS Ψ ðλkÞ λðkÞ . Go to step 3.
Because the NCLS algorithm does not impose the ASC, its generated abundance
fractions may not necessarily sum up to one. In this case, it must solve the following
constrained optimization problem:
n o n Xp o
min ðr MαÞT ðr MαÞ subject to Δ ¼ α αj 0 for 8j, α j ¼ 1 :
α2Δ j¼1
ð9:14Þ
^ SCLS
The optimal solution to (9.14) first takes advantage of the SCLS solution, α
ðrÞ in (9.8), as an initial estimate to derive
1 1
^ FCLS ðrÞ ¼ P⊥
α ^ SCLS ðrÞ þ MT M 1 1T MT M 1
M, 1 α ð9:15Þ
and
T 1 T T 1 T
P⊥
M, 1 ¼ ILL M M 1 1 M M 1 1 : ð9:16Þ
It then uses SCLS together with ANC by introducing a new signature matrix
N and an observation vector s into the NCLS specified by
ηM ηr
N¼ and s¼ ; ð9:17Þ
1T 1
9.3 RHSP-LSMA
In (9.1) the signature matrix M is assumed to be known and fixed. Thus, if the
signatures m1, m2, . . ., mp in M are not sufficiently representative, then the LMM in
(9.1) needs to be updated by adding more signatures. The LMM adapting varying
signatures is referred to as an adaptive LMM (ALMM). In most cases, when
ALMM is implemented, the current signatures remain unchanged. Only future
signatures need to be updated. In this case, we should be able to take advantage
of what has already been done for the current signatures and only process new
signatures without reprocessing LSMA, which uses signatures already visited.
This type of scenario is very common in communications and signal processing
when a new incoming input is fed into a system. In LSMA, we consider LMM a
linear system with inputs determined by signatures m1, m2, . . ., mp. Thus, if a
new input, such as a new signature mp+1, comes in to update ALMM, the
system should not be processed by entire inputs again but rather updated by
only the new incoming input, mp+1. A good representative example is a Kalman
filter (Chap. 3), which operates in exactly this way. Applying an idea similar to
Kalman filtering, the following subsection develops the so-called RHSP-LSMA,
which can update LSMA recursively in signatures in the same way that a
Kalman filter does.
" #!1
1 MpT Mp MpT mpþ1
T
Mpþ1 Mpþ1 ¼
T T
mpþ1 Mp mpþ1 mpþ1
2 1 T 3
6 M T
p M p þ βM # T
p pþ1 pþ1 Mp
m m #
βM#p mpþ1 7
¼6
4 T
7;
5
βmpþ1T
M#p β
ð9:19Þ
1
1 1
where M#p¼ Mp Mp
T
Mp and β ¼ mpþ1 I Mp Mp Mp
T T T T
Mp mpþ1
n h i o1
¼ mpþ1T
P⊥Mp mpþ1 . Note that P⊥
Mp ¼ I PMp can be easily calculated via PMp
1
¼ Mp MpT Mp MpT ¼ Mp M#p using (9.19). The least-squares estimate of αp+1,
^ LS
α pþ1 ðrÞ is then given by
272 9 Recursive Hyperspectral Sample Processing of Linear Spectral Mixture Analysis
02 311 2 3
1 MpT Mp MpT mpþ1 MpT
^ LS
α pþ1 ðrÞ ¼ T
Mpþ1 Mpþ1 T
Mpþ1 r ¼ @4 5A 4 5r
T T T
mpþ1 Mp mpþ1 mpþ1 mpþ1
2 1 T 3
2 3
6 M p
T
M p þ βM #
p m pþ1 m T
pþ1 M #
p βM #
p m pþ1 7 Mp
T
¼6
4 T
74
5
5r
T
βmpþ1 Mp
T #
β m pþ1
2 1 T 3
6 M p
T
M p M T
p r þ βM #
p m pþ1 m T
pþ1 M #
p M T
p r βM #
p m pþ1 M T
p 7r
¼6
4 T
7
5
βmpþ1 T
Mp Mp r þ βmpþ1
# T T
r
2 T 3
6 ^
α LS
p ð r Þ þ βM #
p m pþ1 m T
pþ1 M #
p M p
T
r βM #
p m pþ1 M T
p r 7
¼4 5
βmpþ1 I PMp r
T
2 LS 3 2 LS 3
^ p ðrÞ þ βM#p mpþ1 mpþ1
α T
PM p I r ^ p ðrÞ βM#p mpþ1 mpþ1
α T
P⊥
Mp r
¼4 5¼4 5;
βmpþ1T
P⊥Mp r αLS
pþ1 ð r Þ
ð9:20Þ
⊥
where PM T
p
¼ PMp ¼ Mp M#p and α ^ LS
pþ1 ðrÞ ¼ βmpþ1 PMp r in (9.20) is a scalar
T
obtained by (9.2) that is exactly the least-squares OSP (LSOSP) estimate of the
T LSOSP
^ pþ1 ðrÞ, via OSP
( p + 1)st abundance fraction of αpþ1 ¼ α1 ; α2 ; ; αp ; αpþ1 , α
in (4.24), where U and d are specified by U ¼ m1 mp1 mp and mp+1, respec-
tively. According to (9.20), there are three pieces of information to implement
RHSP-LSMA.
1. Processed information:
⊥
^ LS
α p ðrÞ, PMp :
2. New information:
mp þ1 :
LS
If we further define ^ LS
α ð r Þ^ α ð r Þ ¼ α^ LS ðrÞ, . . . , α
^ LS ðrÞ,
pþ1 pþ1 pþ1 1 pþ1 p pþ1
T
^ LS ðrÞÞT and α
α ^ LS ðrÞ ¼ α ^ LS ðrÞ, . . . , α
^ LS ðrÞ , then α ^ LS
pþ1 ðrÞ ¼
pþ1 pþ1 p pþ1 1 pþ1 p pþ1
0 LS 1
^ ðrÞ
α
LS
^ ðrÞ ¼ @ LS A, where
p pþ1
α
pþ1 pþ1 ^ ðrÞ
α
pþ1 pþ1
^ LS
α ^ LS ðrÞ βM#p mpþ1 mpþ1
ðrÞ ¼ α T
P⊥
Mp r; ð9:21Þ
p pþ1 p p
^ LS ðrÞ ¼ βmpþ1
α T
P⊥
Mp r: ð9:22Þ
pþ1 pþ1
^ LS
Apparently, (9.21) is a recursive equation used to update α ^ LS
pþ1 ðrÞ, that is, α p
α LS ðrÞ ) α
ðrÞ^ ^ LS ðrÞ ) α ^ LS ðrÞ ¼ α
^ LS
pþ1 ðrÞ, which is exactly how a
p p ð9:21Þ p pþ1 ð9:22Þ pþ1 pþ1
Kalman filter is implemented (Chap. 3). In addition,
2 1 T 3
1 h
T
i6 Mp Mp þ βM#p mpþ1 mpþ1T
M#p βM#p mpþ1 7" 1p #
T
1pþ1 T
Mpþ1 Mpþ1 1pþ1 ¼ 1pT 1 6
4 T
7
5
1
βmpþ1
T
M#p β
1 T
¼ 1pT MpT Mp βM#p mpþ1 mpþ1
T
M#p 1p βmpþ1
T
M#p 1p β1pT M#p mpþ1 þ β
1
T
¼ 1pT MpT Mp 1p β 1pT M#p mpþ1 M#p mpþ1 1p 2 1pT M#p mpþ1 þ 1
1 h i2
¼ 1pT MpT Mp 1p β 1pT M#p mpþ1 1 :
ð9:23Þ
1
By virtue of (9.23), P⊥
M, 1 in (9.9) can be calculated recursively by 1
T
MpT Mp 1 in
(9.23), which is a scalar, and α ^ NCLS ðrÞ in (9.12), and α
^ SCLS ðrÞ in (9.8), α ^ FCLS ðrÞ in
^ LS
(9.15) can also be calculated recursively. Note that sign α pþ1 ðrÞ can be calcu-
1
lated directly from the obtained α ^ LS
pþ1 ðrÞ ¼ Mpþ1 Mpþ1
T T
Mpþ1 r once M#pþ1 ¼
1
T T
Mpþ1 Mpþ1 Mpþ1 is calculated.
274 9 Recursive Hyperspectral Sample Processing of Linear Spectral Mixture Analysis
1 1 n h i o1
where Γ ¼ MqT I Mp MpT Mp MpT Mq ¼ MqT P⊥
Mp Mq , which
can be more specifically expressed as Γ( p,q) to indicate its dependency on matrices
Mp and Mq;
" #!1 " #
1 MpT Mp MpT Mq MpT
^ LS
α pþq ðrÞ ¼ Mpþq
T
Mpþq T
Mpþq r¼ r
MqT Mp MqT Mq MqT
2 1 T 3
" T#
MpT Mp þ M#p Mq ΓMqT M#p M#p Mq Γ
6 7 Mp
¼4 T 5 r
ΓT MqT M#p Γ MqT
2 1 T 3
MpT Mp MpT r þ M#p Mq ΓMqT M#p MpT r M#p Mq ΓMqT r
6 7
¼4 T 5
ΓT MqT M#p MpT r þ ΓMqT r
2 T 3
^ LS
α ðrÞ þ M#p Mq ΓMqT M#p MpT r M#p Mq ΓMqT r
¼4 p
5
ΓT MqT I PMp r
" LS # " LS #
^ p ðrÞ þ M#p Mq ΓMqT PMp I r
α ^ p ðrÞ M#p Mq ΓMqT P⊥
α Mp r
¼ ¼ ;
ΓT MqT P⊥Mp r αLS
q ðrÞ
ð9:26Þ
9.4 Adaptive RHSP-LSMA 275
1 T
where P⊥
Mp ¼ M M
p p
#
¼ M p M T
p M p Mp
T
¼ P ⊥
Mp owing to its symmetry, M#p
n h i o1
T ⊥ ⊥
Mp ¼ I, and αLS
q ð r Þ ¼ Γ M q PM p r ¼ M q PM p M q
T T
MqT P⊥
Mp r in (9.26) is an
estimate of αq(r) in a least-squares sense. Some comments on (9.24) are
worthwhile.
T T
1. Since P⊥ Mp is also idempotent, that is, PMp
⊥
¼ P⊥Mp and PMp
⊥
P⊥ ⊥
Mp ¼ PMp ,
h iT h
i 1
Γ¼ P⊥
Mp Mq P⊥
Mp Mq . This gives rise to αLS
q ðrÞ ¼
T
1
P⊥
Mp Mq P⊥Mp Mq MqT P⊥ ⊥
Mp r , where PMp r maps a data sample vector
⊥
r onto the linear subspace Mp orthogonal to hMpi spanned by Mp via P⊥ Mp .
2. It is also interesting to note that αLS
q ðrÞ has exactly the same form as the LSOSP
1 T ⊥
obtained in (4.21) and (9.2) by αjLSOSP ðrÞ ¼ mj P⊥ U mj mj PU r, with mj and
U being the signature matrix consisting of all signatures but with mj replaced by
Mq and Mp, respectively. As also noted in the first component of (9.25),
M#p Mq ΓMqT P⊥ ⊥
Mp r, the vector of PMp r is simply the residual of the data sample
⊥
vector r leaking from the space Mp into the space Mp .
3. If we let Mq be a singleton set consisting of only one signature mp+1, then (9.26)
is reduced to (9.20), where the matrix Γ becomes the scalar β.
As a final note, the results presented in this section can be extended to multiple
sets of signatures greater than two by processing, using RHSP-LSMA, two sets of
disjoint signature matrices at a time. Conversely, it can be also used to break up the
signature matrix Mp into two disjoint sub-matrices, Mr and Ms with p ¼ r þ s.
The RHSP-LSMA developed in Sect. 9.3 assumes that the signature knowledge of
m1, m2, . . ., mp is provided a priori. In fact, this is generally not true in practical
applications, and such knowledge must be obtained directly from the data to be
processed. In doing so, two key issues must be addressed: determining the number
of signatures, p, that need to be generated and finding p signatures once the value of
p is determined. The ULSMA was the first method developed to deal with both
issues (Chang et al. 2010a, 2011b; also Chap. 17 in Chang 2013). It first estimated
appropriate values for p by virtual dimensionality (VD), developed in Chang (2003)
and Chang and Du (2004); then it used least-squares methods in Chang et al. (2010)
or component analysis in Chang et al. (2011b) to find the desired signatures, m1,
m2, . . ., mp. In this type of ULSMA, determining the value of p and finding
p desired signatures m1, m2, . . ., mp are decoupled. However, in reality, the
276 9 Recursive Hyperspectral Sample Processing of Linear Spectral Mixture Analysis
signatures used to form an LMM are actually closely related to the value of p, which
is in turn determined by various applications. In other words, if signatures are
selected adequately, the needed value of p will be smaller; otherwise, it may be
larger than its real value if the signatures are poorly selected. Thus, to cope with this
dilemma, this section presents a new alternative approach, called ARHSP-LSMA,
to resolve these two issues via ALMM defined by adaptively adjusting p signatures
used in LMM in (9.1) as
r ¼ Mp αp þ n; ð9:27Þ
where p is considered to be a variable and Mp ¼ m1 m2 mp1 mp , with mp being
generated by RHSP-LSMA as a new pth signature.
Implementing ARHSP-LSMA requires several pieces of information, which
must be obtained and generated directly from the data to be processed: (1) how to
find appropriate signatures to be used to form an ALMM (9.27); (2) how to
determine the number of signatures needed for an ALMM; (3) how to implement
LSMA without repeatedly processing entire signatures for an ALMM. Each of
these is described in steps 2, 4, and 3, respectively, and summarized as follows.
ARHSP-LSMA
1. Initial condition:
Find the initial signature, t1LSMA ¼ argfmaxr rT rg, over all data sample vectors r.
2. Design a particular algorithm to find the desired signatures to form an ALMM.
3. Implement RHSP-LSMA with the found signatures one by one recursively while
signatures are being generated.
4. Apply a specific stopping rule to determine whether ARHSP-LSMA is to be
terminated, in which case the number of signatures, p, is automatically set at this
stage.
In ARHSP-LSMA, finding the signatures in step 2 and applying a stopping rule
in step 4 are very closely related to applications. For example, if LSMA is used for
subtarget detection or mixed target classification, finding unsupervised targets is the
first priority, and the criterion to determine how many such targets should be
coupled with the algorithm used to generate these targets. On the other hand, if
LSMA is used for LSU to unmix data, an appropriate criterion for optimality may
be an unmixed error, where the algorithm for finding signatures should be a spectral
unmixing method. If LSMA is used to find endmembers, an appropriate criterion
for optimality will be maximum simplex volume, and the endmember finding
algorithm should be one used to generate endmembers. In what follows, two
approaches to generating signatures for ALMM.
9.4 Adaptive RHSP-LSMA 277
To see how to find new signatures, we consider the second term of (9.21),
^ LS
new signature, mp+1. To make the best possible estimate for α ðrÞ, we need to
p pþ1
minimize this error term as much as we can. This is equivalent to minimizing
n h i o1
β ¼ mpþ1
T
P⊥
Mp m pþ1
T
or maximizing mpþ1 P⊥
Mp in (9.28), both of which turn
out to be the same thing. If we interpret U ¼ Mp and d ¼ mp+1 in (9.1), then β
is exactly the constant used to correct the estimation error in the LSOSP
estimate of α^ pþ1
LSOSP
ðrÞ and α^ LS ðrÞ ¼ βmpþ1
T
P⊥
Mp r , as derived in (9.22).
pþ1 pþ1
n h i o
Now let ρ mpþ1 ; Mp be a measure defined by mpþ1 T
P⊥Mp mpþ1 . Since P⊥
Mp
T
is idempotent and P⊥ Mp ¼ P⊥
Mp ,
n T h ⊥ i o
T T T
ρ mpþ1 ; Mp ¼ mpþ1 PMp mpþ1 ¼ P⊥Mp m pþ1 P ⊥
Mp m pþ1
T 2
⊥
¼ P⊥
Mp m pþ1 P ⊥
Mp m pþ1 ¼ P Mp m pþ1 ¼ β1 :
ð9:29Þ
The significance of (9.29) is that ρ mpþ1 ; Mp is a measure of the residual resulting
from the leakage of signatures in Mp into its orthogonal complement subspace
⊥
Mp , which determines how much energy can be generated from the new
signature mp+1. In other words, maximizing (9.29) is equivalent to minimizing β,
which in turn means minimizing the estimation error. This quantity provides an
effective criterion for generating new signatures.
3. Stopping rule:
Use (9.29) and (9.30) to calculate
ρ tOSP
p ; M p1 ð9:31Þ
N
as a criterion to replace ρ tOSP
p ; Mp1 in (9.31), where fri gi¼1 are all data sample
vectors. Unlike (9.31), which is monotonically decreasing and requires a prescribed
error threshold to terminate the algorithm, (9.32) is not monotonically decreasing as
Mp is increasingly augmented owing to a model fitting error. In other words, when a
set of {tp} is underrepresented by LMM, the unmixed error in (9.32) will be
increased until it reaches a certain value of p, which is optimal; after that, (9.32)
begins to increase again due to an overfitting error resulting from {tp} which is
overrepresented by LMM. Thus, applying this fact we can use this model fitting
error resulting from LMM to find the optimal value of p, p*, as follows:
p* ¼ arg minp2 LSU Mp : ð9:33Þ
Note that ideas similar to (9.33) are also explored by HySime and GENE
(Ambikapathi et al. 2013).
9.5 Determination of Number of Signatures for ARHSP-LSMA 279
which is a data sample vector that yields the maximum unmixed error among all
data sample vectors. The reason for maximizing (9.34) is that tUE
p is causing the
maximum unmixed error and it should be included in the signature matrix to
reduce the unmixedh error. i
3. Augment Mp ¼ Mp1 tLSU
p , and check whether p ¼ L, where L is the total
number of spectral bands. If it is not, go to step 2. Otherwise, continue.
4. Stopping rule:
Find the minimal unmixed error over p via (9.33). However, it is worth noting
that p* is largely dependent on the set of found {tLSU
p } that are used to form the
LMM. Different algorithms find different sets of {tLSUp }, which of course results
*
in different values of p .
^ p in (9.32).
In the LSU-based algorithm, there is no specific algorithm to find α
Thus, when (9.32) is performed, three least-squares algorithms can be used: an
unconstrained-abundance least-squares method, such as α ^ LS
p specified by (9.2);
^ pNCLS specified by
partially abundance-constrained least-squares methods, such as α
(9.11); and fully abundance-constrained least-squares method, such as α ^ pFCLS
specified by (9.15). Thus, if the FCLS method is used to find tLSU
p via (9.34) in
step 2, the LSU-based algorithm becomes the unsupervised fully constrained least-
squares (UFCLS) method in Sect. 4.4.2.1.3.
where ηp in (9.35) is the maximum residual of the pth target data sample, tRp - LSMA ,
found by RHSP-LSMA and leaked from < Up1 > into < Up1 >⊥ , which is the
complement space orthogonal to the space < Up1 >. It should be noted that since
n oL
the found tRHSP-LSMA
p obtained by the RHSP-LSMA algorithm may be highly
p¼1
correlated, ηp in (9.35) is used instead because ηp represents the maximum residuals of
tpRHSP - LSMA leaked into < Up1 >⊥ .
It is this sequence, {ηp} given by (9.35), that will be used as the signal source in a
binary composite hypothesis testing problem to determine whether or not the pth
potential target candidate, tpRHSP - LSMA is a true target by a detector formulated as
follows:
H0 : ηp p ηp H 0 ¼ p0 ηp
versus for p ¼ 1, 2, . . . , L ð9:36Þ
H1 : ηp p ηp H 1 ¼ p1 ηp
where the alternative hypothesis H1 and the null hypothesis H0 represent two cases
of tRHSP - LSMA being a target signal source under H and not a target signal source
p 1
under H0, respectively, in the sense that H0 represents the maximum residual
resulting from the background signal sources, while H1 represents the maximum
residual leaked from the target signal sources. Note that when p ¼ 0, η0 is undefined
in (9.35). To make (9.36) work, we need to find probability distributions under both
hypotheses. The assumption made on (9.35) is that if a signal source is not an
endmember under H0, it should be considered part of the background that can be
characterized by a Gaussian distribution. On the other hand, if a signal source is
indeed a desired target signal source, it should be uniformly distributed over a range
from 0 to ηp1. As shown in Sect. 4.5.1, the probability distribution under each
hypothesis can be obtained by
η p pν p η p
p H 0 ηp ¼ ; ð9:37Þ
ηp pνp ηp þ Fνp ηp
9.5 Determination of Number of Signatures for ARHSP-LSMA 281
where PF is a predetermined false alarm probability, δNP - ηp ¼ 1, only
NP ARHSP LSMA
if δNP
ARHSP-LSMA p η ¼ 1 and δ ARHSP-LSMA p η ¼ 0 if δARHSP-LSMA ηp < 1:
NP
In analogy with Sect. 4.5.3, we can also replace the signal energy specified by ηp
pffiffiffiffiffi
in (9.35) with the signal strength ηp specified by
pffiffiffiffiffi ⊥ RHSP-LSMA
ηp ¼ PMp1 tp : ð9:41Þ
a b
p11, p12, p13
Fig. 9.1 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels
Table 9.1 tabulates the values of nRHSP-LSMA via VDNP ARHSP - LSMA (PF) determined
by (9.40) and (9.41), with various false alarm probabilities, PF, using three
unsupervised target finding algorithms, the OSP-based ATGP, described in Sect.
4.4.2.3, and two LSE-based methods, UNCLS and UFCLS, described in Sects.
4.4.2.2.2 and 4.4.2.2.3. It should be noted that since ATGP is the same as ULSOSP
in Sect. 4.4.2.2.1 scaled by a constant, their TSVD results are the same. Also
included in Table 9.1 are the VD results estimated by the Harsanyi–Farrand–
Chang (HFC)/noise-whitened HFC (NWHFC) method tabulated in Table 4.7 in
Sect. 4.4 for comparison.
9.5 Determination of Number of Signatures for ARHSP-LSMA 283
3
10
105
104
102
3
10
101 102
0 5 10 15 20 25 30 35 40 45 0 5 10 15 20 25 30 35 40 45
Iteration Iteration
106
Maximum Residual Energy
Maximum Residual Length
103
105
102 104
103
101
102
100 101
0 20 40 60 80 100 120 0 20 40 60 80 100 120
Iteration Iteration
106
Maximum Residual Energy
Maximum Residual Length
103
105
102 104
103
101
102
100 101
0 20 40 60 80 100 120 0 20 40 60 80 100 120
Iteration Iteration
n oL
Fig. 9.2 Plots of values of (9.40) and (9.41) versus targets tRHSP -LSMA generated by ATGP,
p
p¼1
UNCLS, and UFCLS
284 9 Recursive Hyperspectral Sample Processing of Linear Spectral Mixture Analysis
As shown in Table 9.1, the value of nRHSP-LSMA is always greater than the value
of VD estimated by eigenanalysis-based methods, such as the HFC and NWHFC
methods. Interestingly, the value of nRHSP-LSMA is increased proportionally to the
number of abundance constraints imposed on the used target finding algorithms.
For example, ATGP produced the least value of nRHSP-LSMA, with UFCLS produc-
ing the largest value.
To extract panel pixels in five different rows from the HYDICE scene in Fig. 9.1,
the value of VD, nVD, must be at least 18, as demonstrated in Chang et al. (2010a,
2011b). However, according to Table 9.1, the values of nRHSP-LSMA via
VDNPARHSP - LSMA (PF) must be at least 39. In this case, we chose these two values to
form an ALMM to perform ARHSP-LSMA, where Fig. 9.3a, b shows 18 and
39 RHSP-LSMA extracted target pixels, respectively.
As shown in Fig. 9.3a, RHSP-LSMA could extract one panel pixel for each of
the five different rows. This implies that a value of 18 is the minimum number of
target pixels that must be generated by ARHSP-LSMA. Table 9.2 tabulates panel
pixels among target pixels 1–9 and 10–18 extracted by RHSP-LSMA in Fig. 9.3a,
respectively.
Table 9.3 tabulates the abundance fractions of 20 panel pixels, including 19 R
panel pixels and a Y panel pixel, p212, estimated by FCLS using target pixels
extracted by RHSP-LSMA in Fig. 9.3a, b as signatures to form ALMM. The first
and second columns of Table 9.3 tabulate the FCLS-estimated abundance fractions
of the 1st to 9th RHSP-LSMA-generated target pixels, including 3 R panel pixels,
p11, p312, p521, and the 10th to 18th RHSP-LSMA-generated target pixels, including
1 R panel pixel, p521, and 1 Y panel pixel, p212, given in Table 9.2, respectively. As
we can see from Table 9.3, no R panel pixels in rows 2 and 4 were estimated in the
second column since no panel pixels were found by RHSP-LSMA in its first nine
target pixels. Similarly, no panel pixels in rows 1, 3, and 5 were found by RHSP-
LSMA in its 10th–18th target pixels (Fig. 9.3a). Thus, no abundance fractions of the
panel pixels in rows 1, 3, and 5 were estimated in the third column. However, we
can also fuse the results in the second and third columns via (9.26) to obtain the
results in the fourth column, which turn out to be the same as the results in the fifth
column obtained by applying FCLS directly to the signatures made up of the 1st to
18th target pixels. Following the same treatment, we can also fuse the FCLS-
estimated abundance fraction results of 20 panel pixels using the first 18 RHSP-
LSMA-generated target pixels with FCLS-estimated abundance fraction results of
20 panel pixels using the 19th to 39th RHSP-LSMA-generated target pixels in
Fig. 9.3b to find the FCLS-estimated abundance fractions of the 20 panel pixels in
the sixth column, which are also equal to the FCLS-estimated abundance fraction
9.6 HYDICE Experiments 285
results of the 20 panel pixels in the seventh column using the 39 RHSP-LSMA-
generated target pixels in Fig. 9.3b.
As a final concluding remark, despite the fact that the popular Cuprite scene has
been widely studied for endmember extraction, it is impossible to use this scene in
286
our experiments to evaluate data unmixing due to insufficient ground truth about
background within the scene which actually plays a key role in LSMA. Nonethe-
less, the experiments conducted for the HYDICE scene should be sufficient to
justify our findings.
9.7 Conclusions
LSMA is generally performed with signatures, which are used to form an LMM to
be provided a priori. In reality, such signature knowledge must be obtained directly
from the data to be processed since the assumed signature knowledge may not be
reliable. However, to extend LSMA in an unsupervised manner, two main issues
need to be addressed: how many signatures are needed and how these signatures are
to be found. This chapter developed a theory of adaptive recursive hyperspectral
sample processing of LSMA (ARHSP-LSMA) that deals with these two issues
altogether as a whole. Many benefits can be gained from ARHSP-LSMA that
cannot be found in LSMA. (1) The LMM used by LSMA is extendable to an
adaptive LMM (ALMM). (2) To find signatures to update ALMM, ARHSP-LSMA
is developed to allow LSMA to perform recursively signature by signature.
(3) ARHSP-LSMA can fuse LSMA results obtained using two different sets of
signatures without reprocessing all signatures. (4) With the development of
ARHSP-LSMA, the innovations information provided by the correlation between
processed information and new signature information in the recursive equation
explores a means of finding new signatures that can adapt to different applications.
(5) The signatures found by RHSP-LSMA can be used for TSVD (Chap. 4) to
develop RHSP-LSMA-specified VD, nRHSP-LSMA, which further determines how
many signatures are needed to perform ARHSP-LSMA. (6) By virtue of ARHSP-
LSMA, there is no need for data storage or complex calculations. With such
advantages, ARHSP-LSMA can be very useful in data transmission and commu-
nication. (7) Most importantly, ARHSP-LSMA can take advantage of ALMM to
fuse LSMA results obtained by multiple difference sets of signatures without
reprocessing the complete set of signatures. This is particularly useful in satellite
data communication when data processing can be carried out at different locations
by different users. This idea is similar to that of parallel processing, where each
processor corresponds to data processing using one set of signatures. (8) Finally,
owing to its recursive structures, ARHSP-LSMA can be realized in hardware
design, such as in Field Programmable Gate Array implementation.
Chapter 10
Recursive Hyperspectral Sample Processing
of Maximum Likelihood Estimation
10.1 Introduction
results, Chang (1998) further showed that MLE and OSP were actually two
different approaches designed with different rationales. More specifically, if MLE
performs as an estimator and not a classifier, as is common in traditional remote
sensing, it is indeed an a priori estimation approach with a Gaussian assumption
(Poor 1994), while OSP is an a posteriori signal detection technique that uses
signal-to-noise ratio (SNR) as an optimal criterion without making an additive
Gaussian noise assumption (Tu et al. 1997; Chang 2005, 2013). In particular, the
signatures used to form an LMM can be found directly from the data to be
processed, referred to as virtual signatures (VSs) in Chang et al. (2010a, 2011b).
The relationship between MLE and LSMA in mixed pixel classification was further
explored in Chang et al. (1998) and Chang (2007b). With this in mind we can
expect that there should also be a very close relationship between recursive
hyperspectral sample processing of linear spectral mixture analysis (RHSP-
LSMA) and recursive hyperspectral sample processing of maximal likelihood
estimation (RHSP-MLE) to be derived in this chapter. Accordingly, for least
squares (LS)-based LSMA the three criteria LSE (Chang et al. 1998b), maximal
OSP residual (Chang et al. 2011c), and MLE error (Settle 1996) can be used to find
VSs for LSMA to unmix data.
Similarly, as discussed in Chap. 9, two major issues are associated with finding
VSs. One is the number of VSs that must be determined. The other issue is how to
find VSs to be used by LSMA for data unmixing. Despite the fact that the first issue
can be addressed by virtual dimensionality (VD) defined in Chap. 4, in many reports
the value of VD really has nothing to do with LSMA because it was not determined
by real signatures found in the data to be processed, as pointed out by Chang
et al. (2014). To resolve this dilemma, we can use target-specified VD developed in
Chap. 4 to produce LS-specified VSs for LSMA. With regard to the second issue,
one interesting approach is called minimum estimated abundance covariance
(MEAC), recently proposed by Du (2012). It assumes that the LMM can be
described by a Gaussian distribution and then the resulting Gaussian MLE error
matrix used as a criterion to produce real data samples as VSs. To facilitate the
process, it further develops a fast procedure to find desired VSs via a recursive
equation. Unfortunately, MEAC did not really address the first issue since it did not
use its generated VSs but rather used VD to determine the number of VSs needed
for data unmixing by LSMA. As mentioned earlier, VD is an eigen-based but not an
LS-based criterion. This chapter addresses this issue and further develops an RHSP-
MLE that extends MEAC to deal with the first issue. In addition to RHSP-MLE,
two other recursive approaches, OSP-based recursive hyperspectral sample
processing of OSP (RHSP-OSP), derived in Chap. 8, and LSMA-based recursive
hyperspectral sample processing of LSMA (RHSP-LSMA), derived in Chap. 9, are
also included for comparative study and analysis. All three of these recursive
approaches take advantage of the Neyman–Pearson detection theory Poor (1994)
and Chang et al. (2011c) to determine the number of VSs. It turns out that the
number of VSs is generally much higher than the value determined by VD. This
makes perfect sense because VSs are not necessarily pure and many of them may
also be background mixed signatures. As shown in Chang et al. (2010a, 2011b),
10.2 Criteria for Finding Virtual Signatures for LSMA 291
Chang (2013, 2016), and Gao et al. (2015), for LSMA to work effectively, it is
essential to include background signatures for data unmixing. The experimental
results conducted in this chapter also confirm this fact.
Assume that L is the total number of spectral bands and r is an L-dimensional image
pixel vector. Assume that m1, m2, . . ., mp are p material substance signatures. An
LMM models the spectral signature of r as a linear combination of m1, m2, . . ., mp
with appropriate abundance fractions specified by α1, α2, . . ., αp. More precisely,
r is an L 1 column vector and Mp an L p substance spectral signature matrix,
denoted by Mp ¼ m1 m2 mp1 mp , where mj is an L 1 column vector
represented by the spectral signature of the jth substance tj resident in the pixel
T
vector r. Let αp ¼ α1 ; α2 ; . . . ; αp be a p 1 abundance column vector associated
with r, where αj denotes the abundance fraction of the jth substance signature mj
present in the pixel vector r. To restore the pixel vector r, we assume that the
spectral signature of the pixel vector r is linearly mixed by m1, m2, . . ., mp as
follows:
r ¼ Mp αp þ n; ð10:1Þ
A key issue that arises in (10.2) is how to find a desired mp+1. In what follows,
we develop three LS-based criteria to find VSs to be used for LSMA.
^ LS
where α pþ1 ðrÞ ¼ α^ LS ^ LS
1 ðrÞ, α ^ LS
2 ðrÞ, . . . , α pþ1 ðrÞ ^ LS
and α j ðrÞ is the abundance
fraction of the jth substance signature mj estimated from the data sample vector r.
292 10 Recursive Hyperspectral Sample Processing of Maximum Likelihood Estimation
^ LS
In (10.3), the subscript of α ^ LS
pþ1 ðrÞ, “p + 1” is specifically used to emphasize α pþ1
^ LS
ðrÞ obtained by p + 1 signatures and boldfaced α is used to indicate that α pþ1 ðrÞ is
^ LS
an abundance vector, not a scalar α pþ1 ðrÞ obtained by Chang et al. (1998b) and
Chang (2003, 2005, 2013) as follows:
1
⊥
^ LS
α pþ1 ð r Þ ¼ m T
P m
pþ1 Mp pþ1
T
mpþ1 P⊥
Mp r; ð10:4Þ
where
1
P⊥
Mp ¼ I M p M #
p ¼ I Mp M T
p M p MpT ð10:5Þ
1
and M#p is the pseudo-inverse of Mp given by MpT Mp MpT . In particular,
1
T
mpþ1 P⊥
Mp mpþ1 ð10:6Þ
is included to account for the LSE resulting from estimating the abundance fraction
αp+1 of mp+1. As a consequence, minimizing (10.6) provides an optimal criterion to
find a desired VS:
1
⊥
tLS
pþ1 ¼ arg minmpþ1 mpþ1 PMp mpþ1
T
n o ð10:7Þ
¼ arg maxmpþ1 mpþ1
T
P⊥
Mp mpþ1 :
Using (10.7), an algorithm for finding new VSs can be described as follows.
h i
2. Apply P⊥
Mp via (10.5), with Mp ¼ t1 tp
OSP OSP
, to all pixels
ATGP
mp+1 in the image to find the ( p + 1)th target tpþ1 by
n h
i o
tOSP ¼ arg max m
P⊥ mpþ1
2 : ð10:10Þ
pþ1 pþ1 Mp
T
P mpþ1
2 ¼ P⊥ mpþ1 P⊥
Mp Mp Mp mpþ1
T ð10:11Þ
¼ mpþ1
T
P⊥
Mp P⊥ ⊥
Mp mpþ1 ¼ mpþ1 PMp mpþ1 :
T
In light of (10.11) it easy to see that (10.10) is indeed the same as (10.7) and also
similar to (9.27) in Sect. 9.4.1. In other words, using an LS-based criterion to find
new VSs by minimizing the LSE specified by (10.7) is identical to using an
SNRHSP-based criterion to find new VSs via OSP specified by (10.10). This is
not a coincidence. It was shown in Chap. 12 of Chang (2013) that an OSP can also
be derived as a least-squares OSP (LSOSP) estimate for αp, denoted by α ^ pþ1
LSOSP
ðrÞ,
1
by including a constant mpþ1 T
P⊥Mp mpþ1 ^ OSP
in α pþ1 ðrÞ in (10.9) as follows:
1
^ pþ1
α LSOSP
ðrÞ ¼ mpþ1
T
P⊥
Mp m pþ1 ^ OSP
α pþ1 ðrÞ: ð10:12Þ
α pþ1
Comparing (10.12) to (10.4) it turns out that^ LSOSP
ðrÞin (10.12) is exactly the same
α LS
as^ pþ1 ðrÞin (10.4). However, it should be noted that there are fundamental differences
T
between α^ OSP
pþ1 ð r Þand ^
α LS
pþ1 ð r Þ. First, ^
α LS
pþ1 ðr Þ ¼ ^
α LS
1 ð r Þ , . . . , ^
α LS
p ð rÞ , α LS
pþ1 ðr Þ is an
T
estimator that estimates the abundance fraction vector αpþ1 ¼ α1 ; . . . ; αp ; αpþ1 ,
whereas α^ OSP
pþ1 ðrÞis merely a detector that detects the amount of abundance fraction of a
^ OSP
particular VS, mp+1, αp+1 present in r. In order for α pþ1 ðrÞ to perform the estimation of
abundance fraction for αp+1, α ^ pþ1
LSOSP
ðrÞ is developed in Settle (1996), Chang
1
T
et al. (1998b), and Chang (2005, 2007c, 2013) by including mpþ1 P⊥
Mp mpþ1 in
^ OSP
α ^ LS
pþ1 ðrÞ , which turns out to identical to α ^ OSP
pþ1 ðrÞ. Second, α pþ1 ðrÞ performs the
detection of the specific substance signature mp+1 as a signal compared to α ^ LS
pþ1 ðrÞ,
which estimates the abundance fractions of all signatures, with m1 , m2 , . . . , mpþ1 as a
parameter vector. Nevertheless, it was shown in Chang (2007c, 2013) that for all
1 j p þ 1, α^ LS
j ðrÞ ¼ α^ jLSOSP ðrÞ. Third, α
^ OSP
pþ1 ðrÞhas an advantage over α^ LS
pþ1 ðrÞin
10.2 Criteria for Finding Virtual Signatures for LSMA 295
^ OSP
that α pþ1 ðrÞ provides a constructive means of producing new signal sources in an
unsupervised manner. One such approach is the ATGP described earlier to perform
automatic target recognition. Finally, it is worth noting that the interpretations of (10.7)
and (10.10) are completely different. While (10.7) is used to minimize the abundance
estimation error, (10.10) is used to maximize the maximal OP residual in the hMpi
hyperplane.
An earlier approach to solving (10.1) was to assume that the n in (10.1) is a white
Gaussian noise with its covariance matrix given by σ 2 ILL (Settle 1996). Then its
solution could be obtained by finding the MLE of α, α^ MLE
pþ1 ðrÞ by
1
^ MLE
α pþ1 ðrÞ ¼ Mpþ1 Mpþ1
T T
Mpþ1 r ¼ M#pþ1 r; ð10:13Þ
which is identical to (10.3). Its prediction error matrix is also given by the following
MLE error matrix:
1
Error Matrix α^ MLE
pþ1 ð r Þ ¼ σ 2
M T
M
pþ1 pþ1 : ð10:14Þ
is then used to measure its error performance. Like LS and OSP, (10.15) can be used
as an optimal criterion to find VSs to derive an MLE-based algorithm, summarized
as follows.
If the n in (10.1) is not a white noise but a color Gaussian noise with its
covariance matrix given by Σn, then (10.10) becomes
1
1
^ MLE
α pþ1 ðrÞ ¼ Mpþ1 Σn Mpþ1
T T
Mpþ1 Σ1
n r: ð10:17Þ
Now we can use the square-root representation of Σ1n in Poor (1994) to whiten the
T
color Gaussian noise. In this case, Mpþ1 Σ1
n M pþ1 in (10.17) can be reexpressed as
T
Σ1 1=2
Σ1=2 eT M e
n Mpþ1 ¼ Σn Mpþ1 ¼ M pþ1 pþ1 ; ð10:18Þ
T
Mpþ1 Mpþ1 n
1
1
^ LS
α pþ1 ðrÞ ¼ Mpþ1 Σn Mpþ1
T
Mpþ1T
Σ1 nr
h T i1 1=2 T
¼ Σ1=2
n Mpþ1 Σ1=2n Mpþ1 Σn Mpþ1 Σ1=2 n r ð10:19Þ
1
eT Me eT e e# r ¼α
¼ M pþ1 pþ1 M pþ1 r ¼ M pþ1 e ^ LS
pþ1 ðe
r Þ;
r ¼ Σ1=2
which is reduced to (10.13) with e n r .
Several comments are in order here.
1. The same idea that used (10.18) and (10.19) as criteria to find a new VS was also
used in Du (2012) to derive an abundance covariance-based method, MEAC. As
a consequence, MEAC can be considered a special case of MLE in (10.17). That
is, if we let Rpþ1 ¼ K1=2 Mpþ1 used in Du (2012), MLE specified by (10.17)
becomes exactly the same as MEAC.
2. The sequential forward feature search (SFFS) used in MEAC is essentially the
same as a sequential process using (10.16), where each signature produced by
SFFS is the same as each new VS generated by MLE via (10.16).
3. Using the noise covariance matrix to perform noise whitening is completely
different from using the data covariance matrix to perform data whitening
because the data covariance matrix K is formed by entire data samples and the
noise covariance matrix Σn must be estimated by a reliable method such as the
one developed by Roger and Arnold (1996), which is used in all the experiments
performed in this chapter. The difference between using noise-whitened data by
Σ1=2
n and data whitened by K1/2 will be clearly demonstrated in the experi-
ments conducted in Sects. 10.6 and 10.7.
4. It should be noted that the three criteria OSP, LS, and MLE/MEAC presented in
this section are not interchangeable, even though they produce the same results.
For example, OSP is a detection technique that is derived based on a signal
detection model, with SNR used as an optimal criterion. Unlike OSP which is a
detection technique, both LS and MLE/MEAC are estimation techniques. Thus,
10.3 Recursive Hyperspectral Sample Processing of MLE 297
all these techniques are indeed quite different in their design rationales. More
specifically, LS is a least-squares error-based estimation technique that makes no
assumptions about the probability distributions of data. By contrast,
MLE/MEAC assumes the model error or noise term n in (10.1) and (10.2) to
be a Gaussian distribution. Thus, it is an a priori approach. On the other hand, LS
makes no such statistical assumption about the LMM, so it is an a posteriori
approach. Therefore, OSP, LS, and MLE/MEAC are not the same techniques.
Nevertheless, they can be interpreted by one another in an appropriate manner.
In Sect. 10.2, three criteria, LSE in (10.7), maximal OSP residual in (10.10), and
MLE error in (10.16), are derived to find VSs. However,
to find new VSs,
mp+1,
using these equations, the signature matrix Mpþ1 ¼ m1 m2 mp , mpþ1 must be
recalculated. These three equations cannot
take advantage
of the results previously
produced by Mp to calculate Mpþ1 ¼ Mp mpþ1 using two separate components,
Mp and mp+1. This section addresses this issue and develops recursive algorithms to
implement these three criteria to find each new VS one after another recursively.
(10.6) is
P⊥ Mp m pþ1
" #1
h i1 MpT Mp MpT mpþ1
T
Mpþ1 Mpþ1 ¼ T T
mpþ1 Mp mpþ1 mpþ1
2 1 T 3
MpT Mp þ β
mpþ1T
M#p β
pþ1 p pþ1 p
ð10:20Þ
1
where M#p ¼ MpT Mp MpT and
1 1
β
¼ T
mpþ1 I Mp Mp Mp
T T
Mp mpþ1
pþ1 p
ð10:21Þ
n h i o1
¼ T
mpþ1 P⊥Mp mpþ1 ;
T
mpþ1 P⊥
Mp mpþ1 ¼ ⊥
β
mpþ1
mpþ1 PMp1 mpþ1
T T et p tp et p tp T mpþ1
pþ1 p
h i2
¼ mpþ1
T
P⊥ β
m T et p tp ;
Mp1 m pþ1
pþ1 p pþ1
ð10:22Þ
RHSP-LS/OSP Algorithm
1. Initial condition:
RHSP-LS=OSP
(a) Find an initial target pixel vector t1 ¼ arg maxm1 m1T m1 . Set
h i
RHSP-LS=OSP
M1 ¼ t1 .
(b) Calculate m2T P⊥
M1 m2 as follows:
m2T P⊥RHSP-LS=OSP m2
t1
" 1 #
T T
RHSP-LS=OSP RHSP-LS=OSP RHSP-LS=OSP RHSP-LS=OSP
¼ m2T I t1 t1 t1 t1 m2
" 1 #
T T
RHSP-LS=OSP RHSP-LS=OSP RHSP-LS=OSP RHSP-LS=OSP
¼ m2T I t1 t1 t1 t1 m2
ð10:23Þ
and use (10.11) to find
10.3 Recursive Hyperspectral Sample Processing of MLE 299
RHSP-LS=OSP T ⊥
t2 ¼ arg maxm2 m2 PtRHSP-LS=OSP m2 : ð10:24Þ
1
h i
RHSP-LS=OSP RHSP-LS=OSP
(c) Form M2 ¼ t1 t2 .
(d) Let p ¼ 2.
RHSP-LS=OSP
2. At the ( p + 1)th iteration, find tpþ1 by maximizing (10.22) over all data
sample vectors, mp+1:
n h io
RHSP-LS=OSP
tpþ1 ¼ arg maxmpþ1 mpþ1
T
P⊥
Mp mpþ1 ð10:25Þ
h i
RHSP-LS=OSP RHSP-LS=OSP
where Mp ¼ t1 tp .
3. Stopping rule:
If a stopping rule to be discussed in the following section is satisfied, RHSP-LS/
OSP is terminated. Otherwise, let p p þ 1 and go to step 2.
10.3.2 RHSP-MLE
1 n h i o1
LS
¼ trace MpT Mp þ mpþ1
T
P⊥
Mp mpþ1 ^ <Mp > mpþ1
2
1 þ
α
1
2
LS
¼ trace MpT Mp þ
P⊥
Mp mpþ1
1 þ
α
^ <Mp > mpþ1
2
0
12
1 2
α^ LS ð Þ
þ @
<Mp >
A ,
m
¼ trace ^ pþ1
þ α OP pþ1
MpT Mp mpþ1
PMp mpþ1
ð10:26Þ
T
^ LSM ¼ mpþ1
with α h pi
T ^ pþ1
M#p M#p mpþ1 , where α OP
mpþ1 ¼
P⊥
Mp mpþ1 is the OP
of mp+1 to account for the maximal OP residual from the hyperplane of ‹Mp› and
300 10 Recursive Hyperspectral Sample Processing of Maximum Likelihood Estimation
OSP
OP OSP
αˆ p + 1 (t p +1 ) t p +1
m p+1
OP
αˆ p + 1 (m p + 1 ) MLE
t p +1
OP MLE f
αˆ p + 1 (t p + 1 )
q LS MLE
αˆ p + 1 ( t p + 1 )
LS OSP
αˆ p + 1 (t p + 1 ) LS MEAC
αˆ p + 1 (t p + 1 )
^ LSM mpþ1 is the abundance vector of mp+1 unmixed by t1, t2, . . ., tp in the
α h pi
hyperplane ‹Mp› via (10.3). In addition, tRHSP-MLE can be found by pþ1
8 8 0
12 9 9
>
< >
<
>>
2
α^ hMp i ðmpþ1 Þ
==
LS
RHSP-MLE
tpþ1 ¼ arg minmpþ1
PMp mpþ1
þ @
>
: >
:
P⊥Mp mpþ1
;>
> ;
8 8 0
12 99
>
< >
<
> =>=
2
α^ LS
hMp i ð pþ1 Þ
m
¼ arg minmpþ1
α^ pþ1 mpþ1
OP
þ @ A :
>
: >
: jjα^ pþ1 ðmpþ1 Þjj
OP
;>
> ;
ð10:27Þ
12
2
α^ LS
hMp i ð m Þ
PMp mpþ1
and @
pþ1
A at the same
P⊥Mp mpþ1
2
time, while (10.7) only has to minimize
P mpþ1
. Mp
jjα^ LS ðmpþ1 Þjj
Now, let tan φ ¼ tan ðπ=2 θÞ ¼ α^ pþ1 . Figure 10.1 illustrates the rela-
jj pþ1OP
ðmpþ1 Þjj
^
α OP
mpþ1
and minimizes
α ^ pþ1 mpþ1
, while tOSP
LS
pþ1 only maximizes
pþ1OP
α^ pþ1 mpþ1
. In this case, if there is more than one target sample that yields
OP
LS
α
^ mpþ1
α ^ mpþ1
, whereas
pþ1 pþ1 pþ1
ATGP
tpþ1 does not specify which one should be selected. However, it should be noted
OP ATGP
that
α
^ t
pþ1
α
pþ1 ^ OP mpþ1 for all mp+1. In this case, t ATGP always yields
pþ1 pþ1
an OP at least greater than or equal to the OP produced by tMLE
pþ1 .
10.4 Stopping Rule for RHSP-MLE 301
RHSP-MLE
Interestingly, according to the preceding derivations, finding tpþ1 by
RHSP-MLE is identical to the MEAC method using fast processing, with Rp ¼
Σ1=2
n Mp , in Du (2012). Furthermore, using RHSP-MLE to find the set of
RHSP - MLE
{tp } via (10.27) is also identical to MEAC using the sequential forward
feature search (SFFS).
;
p ¼ arg maxr PMp1 r
tVS ð10:28Þ
2
ηp ¼
P⊥ VS
2
Mp1 tp ¼ maxr
P⊥
Mp1 r ; ð10:29Þ
where ηp in (10.29) is the maximal residual of the pth target data sample, tVS
p is
⊥
found by the VS leaked from Mp1 into Mp1 , which is the complement space
orthogonal to the space Mp1 and turns out to be exactly the same criterion used
P⊥
M0 r ¼ r r in (10.29)
n oL
when p ¼ 1. Since the found tVSp obtained by a VS may be highly correlated,
p¼1
ηp in (10.29) is used instead because ηp represents the maximum residuals of tVS p
⊥
L
leaked into Mp1 . It is this sequence, ηp p¼1 , that will be used as a signal
source in a binary composite hypothesis testing problem to determine whether or
not the pth potential target candidate tVS
p is a true target using a detector formulated
as follows:
302 10 Recursive Hyperspectral Sample Processing of Maximum Likelihood Estimation
⊥
h iT h i
P tVS
2 ¼ P⊥ tVS P⊥ VS
Mp1 p Mp1 p Mp1 tp
T T ð10:30Þ
¼ tVS
p P⊥
Mp1 P⊥ VS
Mp1 tp :
T
Since P⊥
Mp1 is symmetric and idempotent, that is, P⊥
Mp1 P⊥ ⊥
Mp1 ¼ PMp1 ,
(10.30) is reduced to
⊥
T
P tVS
2 ¼ tVS P⊥ VS
; ð10:31Þ
Mp1 p p Mp1 tp
which is exactly the reciprocal of (10.6). This shows that the stopping rule using
(10.29) for terminating VSs is identical to signal sources used under each hypoth-
esis by the Neyman–Pearson detection theory proposed in Chang (2003) and Chang
and Du (2004) to determine VD.
It is this sequence, {ηp} given by (10.29) that will be used as a signal source in a
binary composite hypothesis testing problem to determine whether or not the pth
VS tVS
p is a true target using a detector formulated as follows:
H 0 : ηp p ηp
H 0 ¼ p0 ηp
versus
for p ¼ 1, 2, . . . , L ; ð10:32Þ
H 1 : ηp p ηp
H 1 ¼ p1 ηp
where the alternative hypothesis H1 and the null hypothesis H0 represent two cases
of tVS
p being a target signal source under H1 and not a target signal source under H0
respectively in the sense that H0 represents the maximum residual resulting from
the background signal sources, while H1 represents the maximum residual leaked
from the target signal sources. To make (10.32) work, we need to find the proba-
bility distributions under both hypotheses. The assumption made in (10.32) is that if
a signal source is not an endmember under H0, it should be considered part of the
background, which can be characterized by a Gaussian distribution. On the other
hand, if a signal source is indeed a desired target signal source, it should be
uniformly distributed over the interval of (0, ηp1). For details on deriving proba-
bility distributions, we refer the reader to Kuybeda et al. (2007). By virtue of the
Neyman–Pearson detection theory, we can derive the following Neyman–Pearson
detector (NPD), which maximizes the detection power PD subject to the false alarm
probability PF, which determines the threshold value τp:
8
< 1, if Λ ηp > τp ,
δNP η ¼ 1 with probability κ, if Λ ηp ¼ τp, ð10:33Þ
VS p
:
0, if Λ ηp < τp ;
where the likelihood ratio test Λ(ηp) is given by Λ ηp ¼ p1 ηp =p0 ηp , with p0(ηp)
and p1(ηp) given by (10.33), and κ is the probability of saying H1. Thus, a case of
10.5 Discussions 303
T
ηp ¼ tVS
p P⊥
Mp1 tp > τ p indicates that δVS ηp ¼ 1 in (10.33) fails the test, in
VS NP
10.5 Discussions
concept of VS arises from this need. It is derived from VD, which was originally
developed to find the number of spectrally distinct signatures in data that can be
used by LSMA to unmix data via these signatures defined as VSs. With this
interpretation the signatures m1, m2, . . ., mp used to form the LMM in (10.1) are
actually VSs.
Since VSs do not necessarily have to be endmembers, those algorithms that are
designed for the purpose of finding endmembers may not be applicable to finding
VSs owing to constraints imposed on the algorithms, such as simplex-based algo-
rithms, N-FINDR, and simplex growing algorithm (SGA) (Chang et al. 2006). To
resolve this dilemma, this chapter takes a rather different approach by looking at
various criteria used for LSMA and other designs and develops approaches to
finding VSs where three criteria, LS, OSP, and MLE, are developed in Sect. 10.2
and two recursive algorithms (RHSP-LS/OSP for LS/OSP and RHSP-MLE for
MLE) are designed in Sect. 10.3.
Recently, endmember variability has received considerable attention from
researchers (Dennison and Roberts 2003; Somers et al. 2011) who take into account
the variability of an endmember present in the data. Instead of working on a single
endmember a group of signatures, referred to as an endmember class, one type of
endmember is represent so as to address signature corruption caused by physical
effects, such as noise or interference encountered in real environments. In this case,
several issues need to be addressed. One is how to find an endmember class that can
appropriately account for a given true endmember. Another issue is dealing with
individual samples in an endmember class. A third issue concerns finding a
signature that represents its endmember class. Interestingly, the concept of VS
can be used to address these issues because a VS need not be a true endmember
as a pure signature. Instead, it must be a real data sample vector directly extracted
from the data. More details can be found in Gao et al. (2014).
The synthetic image simulated in Fig. 1.15 is shown in Fig. 10.2 with five panels in
each row simulated by the same mineral signature and five panels in each column
having the same size.
Among 25 panels are five 4 4 pure-pixel panels for each row in the first column
and five 2 2 pure-pixel panels for each row in the second column, five 2 2 mixed
pixel panels for each row in the third column, and five 1 1 subpixel panels for
each row in both the fourth and fifth columns, where the mixed and subpanel pixels
were simulated according to the legends in Fig. 10.2. Thus, a total of 100 pure
pixels (80 in the first column and 20 in second column), referred to as endmember
pixels, were simulated in the data by the five endmembers A, B, C, K, and M. An
area marked “BKG” in the upper right corner of Fig. 1.14a was selected to find its
sample mean, that is, the average of all pixel vectors within the BKG area, denoted
by b and plotted in Fig. 1.14b, to be used to simulate the background (BKG) for an
10.6 Synthetic Image Experiments 305
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
image scene with a size of 200 200 pixels in Fig. 10.2. The reason for this
background selection is empirical since the selected BKG area seemed more
homogeneous than other regions. Nevertheless, other areas can also be selected
for the same purpose. This b-simulated image background was further corrupted by
an additive noise to achieve a certain SNR, defined as 50 % signature (i.e., reflec-
tance/radiance) divided by the standard deviation of the noise in Harsanyi and
Chang (1994). Once target pixels and background are simulated, two types of target
insertion can be designed to simulate experiments for various applications.
The first type of target insertion is target implantation (TI), which can be
simulated by inserting clean target panels into a noisy image BKG by replacing
their corresponding BKG pixels, where the SNR is empirically set to 20:1. That is,
TI implants clean target panel pixels into noise-corrupted image BKG with an
SNR of 20:1, in which case there are 100 pure panel pixels in the first and second
columns.
A second type of target insertion is target embeddedness (TE), which is simu-
lated by embedding clean target panels into a noisy image BKG by superimposing
target panel pixels over the BKG pixels where the SNR is empirically set to 20:1.
That is, TE embeds clean target panel pixels into noise-corrupted image BKG with
SNR ¼ 20:1 in which case all 100 pure panel pixels in the first and second columns
are no longer pure. In other words, a salient difference between TI and TE is worth
mentioning. Since TE inserts targets by adding target pixels to and superimposing
over background pixels instead of replacing background pixels like TI does for
target insertion. As a consequence, the abundance fraction of the pixel into which a
target pixel is embedded is not summed to one.
RHSP-LS/OSP and RHSP-MLE were then implemented in conjunction with
(10.29) to estimate nVS for TI and TE scenarios. For the purpose of comparative
analysis the unsupervised fully constrained least-squares (UFCLS) method devel-
oped by Heinz and Chang (2001) was also implemented. However, in this case, due
306 10 Recursive Hyperspectral Sample Processing of Maximum Likelihood Estimation
Table 10.1 nVS estimated by ATGP, MLE, and UFCLS for TI data
NPD NPD NPD NPD NPD
PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105 MLD
ATGP F F F F F F
ATGP (K) 9 8 8 7 7 8
ATGP (Σn) 8 8 8 7 7 8
RHSP-MLE N/A N/A N/A N/A N/A N/A
RHSP-MLE (K) N/A N/A N/A N/A N/A N/A
(MEAC)
RHSP-MLE (Σn) N/A N/A N/A N/A N/A N/A
UFCLS F F F F F F
UFCLS(K) 8 7 7 7 7 7
UFCLS (Σn) 6 6 6 6 6 6
Table 10.2 nVS estimated by ATGP, MLE, and UFCLS for TE data
NPD NPD NPD NPD NPD
PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105 MLD
ATGP F F F F F F
ATGP(K) 8 7 6 6 6 6
ATGP (Σn) 7 6 6 6 6 6
RHSP-MLE N/A N/A N/A N/A N/A N/A
RHSP-MLE (K) N/A N/A N/A N/A N/A N/A
(MEAC)
RHSP-MLE (Σn) N/A N/A N/A N/A N/A N/A
UFCLS F F F F F F
UFCLS(K) 7 6 6 6 6 6
UFCLS (Σn) 6 6 6 6 6 6
to its use of full abundance constraints, UFCLS does not have an analytical form to
be implemented recursively.
Tables 10.1 and 10.2 tabulate the results with various false alarm probabilities
along with maximum likelihood detector (MLD), where an “F’ indicates that NPD
failed the test and all 189 generated target pixels are treated as VSs. Since RHSP-
MLE makes use of a linear model in (10.1) with assuming white Gaussian noise
(Σ ¼ I) or color Gaussian noise (Σ ¼ Σn) (i.e., MEAC method), experiments were
also conducted on data noise-whitened by Σ1=2 n , with Σn estimated using the
method in Du (2012), and data whitened by K1/2 (i.e., the square root of the sample
covariance matrix K, K1/2 was used as the whitened matrix (Poor 1994)). Simi-
larly, both RHSP-LS/OSP and UFCLS were also implemented on the original data
with white Gaussian noise, Σ ¼ I, where no preprocessing is required, noise-
whitened data by Σ1=2n , and whitened data by K1/2.
Interestingly, for both TI and TE, RHSP-MLE did not work for all scenarios, nor
did RHSP-LS/OSP and UFCLS on the original data with Σ ¼ I. This implied that
more than 189 background pixels were also treated as VSs due to the additive white
10.6 Synthetic Image Experiments 307
Fig. 10.3 Targets found by MEAC for TI scenario. (a) Noise-whitened data. (b) Whitened data
a
17 18
16
10
17 18 5
412 14
7 412 14
7
2 3
15
1 13 86
9 2 3
15
11
2468
10 12
157 157
Fig. 10.4 Targets found by MEAC for TE scenario. (a) Noise-whitened data. (b) Whitened data
Figures 10.5, 10.6, and 10.7 also show the VSs found by RHSP-LS/OSP, RHSP-
MLE, and UFCLS for TI and TE from the original data space, noise-whitened data,
and whitened data, respectively. Since the VSs were generated sequentially and
those generated by a smaller value of nVD were also part of the VSs generated by a
large value of nVS, the number of VSs was set to the largest value, for example, 9 for
TI and 8 for TE.
Apparently, RHSP-MLE did not work well for both scenarios. No matter which
data set was used, it missed at least one panel signature (Fig. 10.5). As also
demonstrated in Fig. 10.5, RHSP-MLE did not pick up panel pixels corresponding
to endmembers in five rows. This is because MLE is not a criterion particularly
designed to find endmembers. By contrast, RHSP-LS/OSP and UFCLS worked
very effectively for both scenarios for all types of data sets in Fig. 10.4, where five
panel pixels corresponding to five different mineral signatures were already found
by the first five RHSP-LS/OSP-generated pixel vectors. This is because OP has
been widely used for finding endmembers such as PPI (Boardman 1994). Compared
to RHSP-LS/OSP, UFCLS always extracted a panel pixel in the fifth row as its first
a
9 9
9
2 3 3 73
7
4
8 6
6
24 248 5
6
1 5 18 1
7
5
3 64
8
7 4
9
258 63 2 6 4
1 1 7 138
7
2 3 3
8
3 4 4
5 5 5
4 2 2
7 7 7
1 1 1
6 6
6
8 8
2 8 3 3
3 4 4
5 5 5
4 2 2
7 8 6 8 6
1 1 1
6
7 7
a
7 8
3 8 4 4
5 5 5
7 9 6 6
2 2 2
4 9 3 3
7
9
1 1 1
6 8
3 4 4
9 9
4 5 5
8
6 6 6
7
2 2 2
5 7 3 3
9
1 1 1 7
8 8
VS and a background pixel as its second VS on all data sets. This indicates that a
background signature is a crucial signature for both TI and TE scenarios and must
be included in the LMM to properly represent data so that UFCLS can perform
LSMA effectively. However, such a background signature is certainly not a pure
signature but rather a mixed signature.
The image scene shown in Fig. 10.8 (also shown Fig. 1.10a) was used for exper-
iments. It was acquired by the airborne Hyperspectral Digital Imagery Collection
Experiment (HYDICE). It has a size of 64 64 pixel vectors with 15 panels in the
scene and the ground truth map in Fig. 10.8b (Fig. 1.10b).
As noted in Fig. 10.8b, the panel pixel, p212, marked in yellow is of particular
interest. According to the ground truth, this panel pixel is not a pure panel pixel but
rather a boundary panel pixel marked in yellow. Thus, when an endmember finding
algorithm, such as N-FINDR developed by Winter (1999), is implemented, p221 is
always extracted. However, if an unsupervised target detection algorithm is used,
panel pixel p212, instead of panel pixel p221, is always the one that gets extracted to
10.7 Real Image Experiments 311
a b
p11, p12, p13
Fig. 10.8 (a) A HYDICE panel scene containing 15 panels and (b) ground truth map of spatial
locations of the 15 panels
Table 10.3 nVS estimated by (10.29) with various false alarm probabilities for HYDICE data
where RHSP-LS/OSP, RHSP-MLE, and UFCLS are used to find VSs
NPD NPD NPD NPD NPD
PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105 MLD
RHSP-LS/OSP 118 118 116 115 114 118
RHSP-LS/OSP (K) 54 53 49 49 46 53
RHSP-LS/OSP (Σn) 33 32 32 30 29 32
RHSP-MLE 128 128 127 126 123 128
RHSP-MLE (K) 79 70 68 59 59 73
(MEAC)
RHSP-MLE (Σn) 44 44 43 43 43 44
UFCLS F F F F F F
UFCLS (K) 133 133 130 129 128 133
UFCLS (Σn) 81 79 79 75 69 79
represent the panel signature in row 2. Also, because of such ambiguity, the panel
signature representing panel pixels in the second row is either p221 or p212, which
endmember finding algorithms always have a difficult time finding. This implies
that the ground truth of R panel pixels in the second row provided in Fig. 10.8b may
not be as accurate as thought.
Similar to synthetic image experiments, RHSP-MLE was also implemented on
the original data, data noise-whitened by Σ1=2
n , and data whitened by K1/2, where
the noise covariance matrix Σn was estimated by the method in Du (2012). For
comparison, RHSP-LS/OSP and UFCLS were also implemented for three cases.
Table 10.3 tabulates nVD estimated for various scenarios using signal energy
specified by (10.24).
Obviously, the number of VSs estimated by VD in Table 20.3 for this HYDICE
scene is higher than the number of endmembers obtained in Chang (2003, 2013)
and Chang and Du (2004). This is mainly because the generated VSs include
endmembers, and some of them are not necessarily pure signatures as endmembers.
312 10 Recursive Hyperspectral Sample Processing of Maximum Likelihood Estimation
Fig. 10.9 VSs found by MLE, RHSP-LS/OSP, and UFCLS using (10.24). (a) 118 VSs produced
by RHSP-LS/OSP. (b) 128 VSs produced by RHSP-MLE. (c) 81 target pixels produced by UFCLS
Figure 10.8 shows the VSs found by RHSP-LS/OSP, RHSP-MLE, and UFCLS
via (10.24) from the original data space, noise-whitened data, and whitened data,
respectively. Since these VSs were generated sequentially and those generated by a
smaller value of nVS were also part of the VSs generated by a large value of nVS, the
number of VSs was set to the largest value, for example, 128. As shown in Fig. 10.9,
all methods were able to find pixels corresponding to five panel signatures, and a
large number of VSs are actually background pixels representing mixed signatures.
10.8 Unmixed Error Analysis 313
a
x 105
5.5 ATGP
UFCLS
5 MLE
4.5
N=9
FCLS-unmixed errors
3.5 N=13
2.5 N=71
N=92
2 Minimum error N=168
1.5
N=25 Minimum error N=168
1
2.5
FCLS-unmixed errors
N=13
N=15
2
N=23
1.5 N=25
N=62 N=71
Minimum error N=168
1
N=61 Minimum error N=169
N=18
N=35
0.5 N=43 N=49
Mimum error N=167
Fig. 10.10 FCLS-unmixing errors using ATGP-, MLE (MEAC)-, and UFCLS-generated VSs. (a)
FCLS-unmixed error for entire image. (b) FCLS-unmixed error for 19 pixels. (c) FCLS-unmixed
error for entire image (K). (d) FCLS-unmixed error for 19 pixels (K). (e) FCLS-unmixed error for
entire image (Σn). (f) FCLS-unmixed error for 19 pixels (Σn)
10.8 Unmixed Error Analysis 315
c
x 105
16 ATGP
UFCLS
N=15 MLE
14
12
FCLS-unmixed errors
10 N=25
8
N=35
N=45
6 N=10
4 N=97
N=23 Minimum error N=169
N=73
2
2 N=12
N=13
N=20
1.5
N=85
Minimum error N=168
N=28
1 N=57
N=25 Minimum error N=166
N=33
0.5 N=37 Minimum error N=169
20 40 60 80 100 120 140 160
Number of VSs
Fig. 10.10 (continued)
316 10 Recursive Hyperspectral Sample Processing of Maximum Likelihood Estimation
e
x 105
ATGP
6
UFCLS
MLE
5
FCLS-unmixed errors
N=14
4
N=18
N=10 N=74
2 Minimum error N=169
N=72
1 N=29
N=50 Minimum error N=168
20 40 60 80 100 120 140 160
Number of VSs
f
x 104
ATGP
UFCLS
MLE
2 N=10
N=15
FCLS-unmixed errors
N=20
1.5
N=24 Minimum error N=169
N=13
1 Minimum error N=166
N=20 N=78
N=31
0.5
Table 10.4 First four sudden drops at unmixed curves generated by ATGP, UFCLS, and MLE
First drop Second drop Third drop Fourth drop
VD (entire scene/19R) (entire scene/19R) (entire scene/19R) (entire scene/19R)
ATGP 9/18 NA/62 NA/71 NA/NA
ATGP (K) 10/12 13/14 73/25 NA/57
ATGP (Σn) 18/10 72/15 NA/20 NA/NA
MLE 13/13 71/15 92/25 NA/61
MLE (K) 15/12 25/85 35/NA 45/NA
MLE (Σn) 14/10 74/13 NA/NA NA/NA
UFCLS 11/23 25/35 NA/43 NA/49
UFCLS (K) NA/13 NA/20 NA/28 NA/33
UFCLS 10/10 29/20 50/24 NA/31
(Σn)
10.9 Conclusions
In recent years, the simplex volume has been widely used to find endmembers that
are supposed to be pure signatures (Schowengerdt 1997). Unfortunately, these
endmembers may not be effective for LSMA when they are used to perform data
unmixing. This is mainly because the signatures used to form an LMM for LSMA
must be representative, and they are not necessarily pure signatures. This is
particularly true of some background signatures that may be mixed, as demon-
strated in Chang et al. (2010a, 2011b). This issue was already addressed in Chaps. 4
and 9 (Chang 2016). This chapter dealt with the same issue from the viewpoint of
MLE from several new aspects. First and foremost, the concept of VS for LSMA
was introduced. Second, three criteria, LSE, OSP, and MLE, were designed to find
VSs where each of them leads to a particular approach, called LS, OSP, and MLE
methods, respectively. Third, to produce VSs in an unsupervised manner, this
chapter also derived recursive algorithms to implement LS, OSP, and MLE,
where RHSP-MLE turned out to be MEAC (Du 2012). In addition, RHSP-LS and
RHSP-OSP were also shown to be the same algorithm, which happens to be the
same recursive algorithm, called recursive ATGP (R-ATGP), developed in Gao and
Chang (2014) and Chang et al. (2015g), and RHSP-ATGP in Chap. 7. Finally, to
further determine the number of VSs, a Neyman–Pearson detection theory-based
approach using real target pixels derived from Chang et al. (2014), called target-
specified virtual dimensionality (TSVD) described in Chap. 4, is proposed to
replace VD to reflect real applications.
Chapter 11
Recursive Hyperspectral Sample Processing
of Orthogonal Projection-Based Simplex
Growing Algorithm
11.1 Introduction
processing time required for calculating SV. However, all the aforementioned
algorithms involve computing SV via a matrix determinant calculation, referred
to as determinant-based SV (DSV) calculation, which usual requires excessively
high computing time. In particular, both N-FINDR and SGA run into the same
challenging numerical issue. That is, the matrices used to calculate the DSV is
generally not of full rank because the number of endmembers, p, to be used to
form a simplex is relatively small compared to the data band dimensionality, L,
the total number of spectral bands. Under such a circumstance, two commonly
used approaches are suggested. One is to reduce the band dimensionality, L, to
p 1. A second approach is to use singular value decomposition (SVD) without
band dimensionality reduction (DR). In the former case, different DR transforms
result in different sets of endmembers (Li et al. 2015a). On the other hand, in the
latter case, SVD may cause numerical instability (Li et al. 2015a). Most impor-
tantly, the DSVs found by both approaches are not necessarily true SV, as shown
in Li et al. (2015a). In particular, it was shown in Chang (2013) and Chang et al.
(2016b) that DSV-based SGA and OP-based vertex component analysis (VCA)
are essentially the same as the automatic target generation process (ATGP)
developed in Sects. 4.4.2.3, provided that the initial conditions were appropri-
ately selected [also see Chaps. 4 and 13 in Chang (2016)]. However, a challeng-
ing issue in computational complexity remains: finding the matrix inverse in
OP-based endmember finding algorithms (EFAs) and calculating DSV in
SV-based EFAs. Interestingly, the issue of how to calculate simplex volume
has received little attention.
In this chapter we present an OP-based SGA (OPSGA) recently developed by Li
and Chang (2016b) that transforms the DSV computational issue into one of finding
OP. Its key idea lies in the fact that an SV can be calculated from a geometric point
of view, referred to as geometric SV (GSV) calculation, by multiplying the SV of its
base by its height, where the base of a simplex is formed by previously found
endmembers and its height is actually the OP of the next new endmember perpen-
dicular to the base. If we grow simplexes one vertex at a time from a single vertex,
then the vertices of previously constructed simplexes will form part of the vertices
of subsequently constructed simplexes. By taking advantage of such a growing
process, we can construct a ( j + 1)-vertex simplex from the previously generated j-
vertex simplex by adding a newly generated endmember, mj+1, as the ( j + 1)th
vertex. SGA turns out to be a perfect candidate to be used for this purpose because it
grows SV one after another in sequence, and vertices previously found by SGA are
indeed part of vertices of subsequent simplexes found by SGA. The next issue is
finding a desired mj+1. In doing so we first consider the SGA-generated j-vertex,
SSGA
j as a base from which SGA grows by adding its next ( j + 1)th endmember mj+1
to SSGA
j to form a new ( j + 1)-vertex simplex, SSGA SGA
jþ1 . In this case, the volume of Sjþ1
can be simply calculated by multiplying the volume of SSGA j by its height, which can
be calculated by finding the mj+1 perpendicular to the base that is exactly SSGA j .
322 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
Interestingly, this height can be further solved by finding the OP of mj+1 in the
complement subspace orthogonal to the base SSGA j via orthogonal subspace projec-
tion (OSP) discussed in Sect. 4.4.2.1 of Chap. 4. In other words, we can use vertices
previously
found by SGA, m1, m2, . . ., mj to form an undesired signature matrix
Uj ¼ m1 m2 mj and find its orthogonal complement subspace via an operator
P⊥Uj ¼ I Uj Uj specified by (4.19), with U being the pseudo-inverse of Uj given
# #
1
by U#j ¼ UjT Uj UjT . Then the height of SSGA
j must be on the hyperplane
specified by P⊥
Uj . More specifically, we can find
n o
mjþ1 ¼ arg maxr P⊥
Uj r ð11:1Þ
According
to Sect. 2.2
of Chap. 2, the volume of a p-vertex simplex,
Sp ¼ S m1 ; m2 ; . . . ; mp , formed by any p data sample vectors m1, m2, . . ., mp is
defined as V(m1, m2, . . ., mp) and can be calculated by the following determinant:
" #
1 . . . 1
1
Det
m1 m2 . . . mp
DSV m1 ; m2 ; . . . ; mp ¼ : ð11:2Þ
ðp 1Þ!
The p vertices of S(m1, m2, . . ., mp), m1, m2, . . ., mp in (11.2) are L-dimensional
vectors. The SGA developed by Chang et al. (2006) uses (11.2) to calculate the
DSV without DR. More specifically, SGA is designed to find a set of j data sample
vectors, denoted by {m1 , m2 , . . ., mj } one after another in such a way that for each
1 j p, mj yields the maximum DSV of (11.2), that is,
n o
m*j ¼ arg maxmj DSV m*1 ; . . . ; m*j1 ; mj : ð11:3Þ
In order for SGA to find"endmembers, we need to #calculate the DSV via (11.2)
1 1
1 1
through the determinant ð1Þ for each j to produce the jth
e e ð2Þ
e ðj1Þ
r
endmember. Although there are fast computational methods proposed in Xiong
et al. (2010) to relieve the computational complexity, it does not change the fact of
determinant computation. In a recent study by Chen et al. (2014), it has shown that
finding the height of a simplex is equivalent to finding the OP onto the hyperplane
linearly spanned by its base. Since the volume of a simplex can be calculated by
multiplying its base by its height, this implies that the simplex OP is actually the OP
from a vertex perpendicular to the simplex formed by all the remaining vertices lies.
Figure 11.1a illustrates this concept with p ¼ 3, where a two-dimensional
(2D) three-vertex simplex with m1 and m2 forms its base and a third vertex
represented by three different data sample vector, m3, m e 3 , and m^ 3 , with the same
magnitude of OP specified by m3A, m e 3 B, and m^ 3 C as its heights when these three
324 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
a b
~ m̂ 4 ~
m
m 3 m3 m̂ 3 m4 4
m3
m1 B m2 C m1 B A
A C
m2
Fig. 11.1 Finding volumes of 2D and 3D simplexes. (a) 2D three-vertex simplexes. (b) 3D four-
vertex simplexes
ATGP
t4 ATGP
SGA
t4
t 4 VCA
SGA
t4
t4
VCA
t4
m3
A
B
C
m1 m2
PU r
U m1 , m 2 , m 3
Fig. 11.2 Finding endmembers by ATGP, VCA, and SGA via OP and simplex
vertices are orthogonally projected onto the line connecting m1 and m2. This
indicates that the volumes of the three three-vertex simplexes, m1m2m3,
m1 m2 me 3 , m1 m2 m
e 3 , are identical because they have the same base formed by the
line segment connecting m1 and m2 and the same height specified by the identical
magnitude of the three different OPs.
Similarly, Fig. 11.2b shows the case of p ¼ 4, where a three-dimensional
(3D) four-vertex simplex with a triangle formed by m1, m2, and m3 as its base
and a fourth vertex represented by three different data sample vectors, m4, m e 4, and
m^ 4, with the same magnitude of OP specified by m4A, m e 4 B, and m
^ 4 C, as its heights
when these three vertices are orthogonally projected onto a hyperplane where the
triangle formed by m1, m2, and m3 lies. Figure 11.2 further shows how this concept
plays out by the three different EFAs discussed in Chap. 3 (Chang 2016): ATGP,
VCA, and SGA.
11.3 Orthogonal Projection-Based Simplex Growing Algorithm 325
Assume that m1, m2, and m3 are three previously found endmembers and form a
hyperplane U ¼ hm1 ; m2 ; m3 i and a fourth endmember to be found by ATGP,
⊥
⊥
VCA, and SGA is denoted by tATGP , tVCA and tSGA with t4ATGP , tVCA , and
SGA
⊥ 4 4 4 4
t4 representing their corresponding OP vectors in the hyperplane specified by
⊥
⊥
PU . The resulting OPs are denoted by their vector lengths, t4ATGP , tVCA
⊥ ,
SGA
⊥ 4
and t 4
.
From Fig. 11.2 we see that, instead of using (11.2) directly to calculate the DSV
formed by the three-vertex (m1, m2, m3) simplex and the fourth vertex, tSGA 4 as
DSV(m1, m2, m3, tSGA
4 ) an equivalent method of finding the DSV is to calculate the
GSV as the product of its base, which is the GSV(m1, m2, m3) of a simplex formed
by the three previously found endmembers m1, m2, m3, that is, the area of
⊥
the triangle formed by m1, m2, m3 with its height specified by the OP, tSGA 4
that is,
⊥
GSV m1 ; m2 ; m3 ; tSGA
4 ¼ ð1=3!Þ tSGA
4
GSVðm1 ; m2 ; m3 Þ: ð11:4Þ
Since tSGA
4 must be the one to produce the maximal GSV, this implies that
GSV m1 ; m2 ; m3 ; tSGA
4 GSV m1 ; m2 ; m3 ; t4ATGP ; ð11:7Þ
which implies
SGA
⊥ ATGP
⊥
t t ð11:8Þ
4 4
because of (11.4) and (11.5). On the other hand, ATGP finds tATGP to produce the
4 ATGP
⊥ 2
maximal OP among all data sample vectors, that is, t4 ¼
⊥
T ⊥
⊥ ATGP
T ⊥ ATGP
max PU r PU r ¼ PU t4 P U t4 , i.e.,
ATGP
⊥ 2 ⊥ SGA
T ⊥ SGA
SGA
⊥ 2
t P t PU t4 ¼ t4 ; ð11:9Þ
4 U 4
which yields
ATGP
⊥ SGA
⊥
t t : ð11:10Þ
4 4
326 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
⊥
⊥
Combining (11.8) and (11.10) yields t4ATGP ¼ tSGA
4
. As a result, their
simplex volumes must be the same given by
With regard to VCA, if the maximal OP is used as a criterion for optimality, then
ATGP
⊥ VCA
⊥
t t : ð11:12Þ
4 4
GSV m1 ; m2 ; m3 ; tSGA
4 GSV m1 ; m2 ; m3 ; tVCA
4 : ð11:14Þ
In the cases of both (11.13) and (11.14), the best performance of the VCA can be
achieved by ATGP in terms of maximal OP and by SGA in terms of the
maximal GSV.
The foregoing argument concludes that, even though the three target pixels
tATGP
4 , tVCA
4 and tSGA
4 may be different in the original data space, tATGP
4 , tVCA
4 and
SGA ⊥
t produce an identical OP in the orthogonal complement subspace PU , that is,
4 ATGP
⊥ VCA
⊥ SGA
⊥
t ¼ t ¼ t . Accordingly, their GSVs are actually
4 4 4
the same as
GSV m1 ; m2 ; m3 ; t4ATGP
¼ GSV m1 ; m2 ; m3 ; tVCA
4 ð11:15Þ
¼ GSV m1 ; m2 ; m3 ; tSGA
4 :
m 2 – m1 , m 3 – m 1
m3
m3 = P[ m – m ] m 3
2 1
m 3 – m1 m
3
m3 – m 2
A
m1 m 2 – m1 m2
m3
t 3 = P[m m ]t 3
1 2
m1 , m 2
O
⊥
be expressed as a sum of two terms, m3 ¼ m⊥ ⊥
3 þ ðm3 Þ⊥, where m3 ¼ P½m2 m1 m3 is
the maximal OP on hm2 m1 , m3 m1 i⊥ perpendicular to the hyperplane
hm2 m1 , m3 m1 i , and ðm3 Þ⊥ ¼ m3 m⊥ 3 ¼ P½m2 m1 m3 is the projection
onto the hyperplane hm2 m1 , m3 m1 i (Fig. 11.3). This
Sðm1 , m2 m1 , m3 m1 Þ is simply the triangle Δ(m1, m2, m3) obtained by adding
to the two-vertex simplex Sðm n 1 , m2 m1 Þ a third endmember m3, which yields
⊥ o
the maximal OP, m3 ¼ arg maxr P r , where the segment Am3 is
½m2 m1
perpendicular to Sðm1 , m2 m1 Þ and its volume is calculated as
GSVðm1 , m2 m1 , m3 m1 Þ ¼ m⊥ 3 ð m 2 m1 Þ.
⊥
However, as also shown in Fig. 11.3, t⊥ 3 ¼ P½m1 m2 t3 is assumed to be the one
yielding the maximal OP perpendicular to the hyperplane
n highlighted inored linearly
⊥
spanned by m1 and m2 with no ASC, where t3 ¼ arg maxr P r is derived
½m1 m2
from the well-known ATGP and is different from m3, which is found by the
segment m2 m1 rather than the space linearly spanned by m1 and m2. Then the
simplex specified by the triangle formed by m1, m2, and t⊥ 3 is highlighted in red, and
its volume can be calculated by GSVðm1 ; m2 ; t3 Þ ¼ t⊥ 3 GSV ðm1 ; m2 Þ according to
(11.5). Obviously the simplex formed by the
triangle Δ(m 1, m2, m3) is different
from that formed by the triangle Δ m1 ; m2 ; t⊥ 3 . This major difference arises from
⊥
the fact that P½m1 m2 operates on a 2D space linearly spanned by two independent
vectors, m1 and m2, while P⊥ ½m2 m1 operates on a 1D space spanned by m2 m1 due
to the fact that the effect of the vector m1 is removed from the simplex volume
calculation simply by placing m1 at the origin. Nevertheless, if all data sample
vectors
in
the data space are linearly translated by m1, then
Δ m1 ; m2 ; t⊥
3 ¼ Δðm1 ; m2 ; m3 Þ. In other words, m1 is considered a biased vector.
Once it is removed, DSGA and ATGP will produce the same set of endmembers
and the same GSV.
328 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
ð hðSjþ1 Þ !j
h
GSV Sjþ1 V Sjþ1 ¼ V B Sjþ1
dh
0 h Sjþ1
ð11:16Þ
h Sjþ1
where h(Sj+1) is the height of the simplex Sj+1, which is perpendicular to its base
B(Sj+1). For example, h(S2) is the height of a two-vertex simplex S2, which is simply
a distance from one point, h(S3) is the height of a triangle S3, which is the third
vertex perpendicular to the line segment connecting two vertices specified by S2,
h(S4) is the height of a pyramid S4, which is the fourth vertex perpendicular to
the plane containing a triangle S3 formed by the first three vertices, and so on. The
method using (11.16) to calculate the GSV is called OPSGA.
As an alternative, the volume of a simplex Sn can also be calculated as
" #
1 1 1
v1 v2 vn
DSVðSn Þ ¼ ; ð11:17Þ
ðn 1Þ!
which is identical to (11.2). The v1, v2, . . ., vn in (11.17) are n + 1 vertices of Sn,
which are L-dimensional vectors. The method using (11.9) to calculate the DSV is
called DSGA. Details on DSV calculation by (11.16) and (11.17) can be found in Li
et al. (2015).
Theoretically, both (11.16) and (11.17) should yield the same SV. In fact, when
it comes to practical implementation, they do not produce the same answers owing
to the use of SVD in (11.17) compared to (11.16), which does not have such an
issue. Consequently, OPSGA is generally preferred to DSGA for two reasons. First,
DSGA may not yield correct results when it is used to find the determinant of a
11.5 Recursive OP-Simplex Growing Algorithm 329
nonsquare matrix. The other reason is that DGSA generally entails very high
computational costs, while OPSGA does not.
Finally, it should be noted that since a simplex is fully constrained, it requires
extra care when it comes to computing its volume. Assume that a simplex Sp ¼ S
1
DSV Sp ¼ DetðME Þ; ð11:18Þ
ðp 1Þ!
where
" # " #
1 1 1 1 0 0
DetðME Þ ¼ Det ¼ Det
m1 m2 mp m1 m2 m1 mp m1 ð11:19Þ
¼ Det½ m2 m1 m3 m1 mp m1 :
In other words, the correct volume of the simplex Sp should be computed based
on the simplex with m1 centered at the origin. In doing so, we simply subtract m1
e j ¼ mj m1 for all 2 j
from all the p 1 endmembers, m2, . . ., mj by letting m
p, and the correct volume of Sp can be calculated as follows:
1
GSV Sp ¼ Det½ m
e2 e3
m e p :
m ð11:20Þ
ðp 1Þ!
According
to
(11.4), two operations are involved in GSV computation,
V m1 ; . . . ; mj1 and the maximal OP found by SGA.
which is equivalent to finding a target data sample vector tj producing the maximal
D E⊥ D E
OP in the space, Ue j1 , perpendicular to the linear subspace U e j1 spanned by
e 2, . . . , m
the previously found j 1 endmembers, m e j1 . Interestingly, solving
(11.22) is equivalent to solving
⊥
e j ¼ arg maxe t j
m e ¼ arg max
⊥e
r e
T
r ; ð11:23Þ
tj er ¼rm12 e
U j1
1
with P⊥ ¼ I U e # and U
e j1 U e# ¼ U eT U e j1 e T being the pseudo-
U
e
U j1
j1 j1 j1 j1
2 T T
⊥
e
inverse of U j1 , and Pe e ⊥
r ¼ Pe e r Pe e⊥
r ¼e ⊥
r Pe e
T ⊥
r and Pe ¼
U j1 U j1 U j1 U j1 U j1
2
P⊥e ¼ Pe
⊥
being idempotent. In other words, (11.4) can be solved via
U j1 U j1
(11.23) and (11.24) by
e j ¼ arg maxr GSV m
m e 2; . . . ; m
e j1 ; e
r
n ⊥
o
¼ arg maxer ð1=ð j 1ÞÞe r GSV m e j1
e 2; . . . ; m
( ð11:25Þ
T )
¼ arg max
e ⊥ e r e
T
r ¼ arg maxer Pe e ⊥
r ⊥
Pe e r :
er2 U j1 U j1 U j1
Coincidentally, the solution to (11.25) can also be found by ATGP. For example,
Fig. 11.1 can be redepicted in Fig. 11.4 to illustrate its idea.
11.5 Recursive OP-Simplex Growing Algorithm 331
a b
ATGP
SGA ATGP VCA SGA
t4 VCA t4
t3 t3 t3 t4
m3-m1
0 m2 - m1 C 0 B C A
B A
m 2- m1
Fig. 11.4 Finding volumes of 2D and 3D simplexes. (a) 2D three-vertex simplexes. (b) 3D four-
vertex simplexes
While we can find an SV via (11.16), we can further derive a recursive equation
that reduces one dimension at a time for (11.21) recursively. More specifically, we
can rewrite (11.21) via (11.22) as
⊥
ð11:26Þ
where GSV(m1, m2) is a 1D simplex with two vertices whose volume is the length
of the line segment connecting m1 and m2, m e 2 ¼ m2 m1 . It should be noted that
RHSP-OPSGA does not directly use (11.1) to calculate an SV. Instead it makes use
of (11.21) and (11.27) without DR. In addition, it is also worth noting that the
volume GSV(m1, . . ., mj) is actually calculated by
1
GSV m1 ; . . . ; mj ¼ e 2m
Det m e 3 m e j :
e j1 m ð11:28Þ
ðj 1Þ!
j3
Y
⊥
me ji ; ð11:29Þ
i¼0
1
e# ¼ U
where U eT U e j1 e T and
U
j1 j1 j1
1 1 1
β¼ e jT
m e e T e
I U j1 U j1 U j1 e T
ej
U j1 m ej P
¼ m T ⊥
e ; ð11:31Þ
e mj U j1
1
e# ¼ U
U e TUej UeT
j j j
2 1 T 32 3
UeT U e j1 þ βU e# m e jm
e jT U e# βU e# m ej eT
U
6 j1 j1 p1 j1 74 j1
5
¼4 T 5
βm T e
e j U j1 #
β e
m j
T
ð11:32Þ
2 T 3
e # þ βU
U e# m e e
m T e#
U Ue T βU e# m e e
m T
6 j1 j1 j j j1 j1 j1 j j 7
¼4 T 5;
βm e jT Ue# Ue T þ βm e T
j1 j1 j
11.5 Recursive OP-Simplex Growing Algorithm 333
h i 1
e jU
U e# ¼ U e j1 m
ej U e TU ej eT
U
j j j
2 T 3
h e # þ βU
i6 U e# m e me T e#
U Ue T
β e
U #
e
m e
m T
j1 j1 j j j1 j1 j1 j j 7
¼ Ue j1 m
ej 6 7
4 T 5
T e# e
βmj U j1 U j1 þ βmj
T T
T
e j1 U
e # þ βU e# m
e j1 U e# e T βU e# m
e j1 U
¼U j1 j1 e j m
e jT U j1 U j1 j1 e j m
e jT
T
βm e jT U
e jm e# e T þ βm
U e jT
e jm
j1 j1
T
¼U e # þ βe
e j1 U u j1 e
u T
βe
u j1 e
m T
β e
u j1 e
m T
þ βm e jT ;
e jm
j1 j1 j j
ð11:33Þ
e# m
e j1 U
ej1 ¼ U
where u j1 e j ;
1
P⊥ ¼ I U e# ¼ I U
e jU ej Ue TUej eT
U
eUj
j j j
¼ P⊥
e βe ej1
u j1 u T
þ βe e jT þ βm
u j1 m e ju
ej1
T
βm e jT
e jm ð11:34Þ
U j1
T
¼ P⊥
e β u ej u
ej1 m ej ;
ej1 m
U j1
1
where P⊥ ¼ I e j1 U
U e# ¼ I U e j1 U eT U e j1 e T . It should also be
U
e
U j1
j1 j1 j1
T
Te
noted that all mj u j1 ¼ u ej1 m
T e j, m ej u
T ej1 m ej ¼ u ej m
ej1 m e j , β in
e e e
(11.30)–(11.34) are scalars. If U p1 is invertible, then U p1 U p1 ¼ I. It should be
#
1
e e
ej1 ¼ U j1 U j1 m
noted that u #
e j, m
ej u
T ej1 ¼ uej1 m
T e j, and β ¼ m ej P
T ⊥ e are
e mj
U j1
e j1 and m
used to account for the correlation between U e j.
11.5.3 RHSP-OPSGA
RHSP-OPSGA
1. Initial conditions:
(a) Find two endmembers, m1 and m2, that yield a line segment with the
maximal length. In other words, find m1 and m2 with the maximal Euclidean
distance, that is, ðm1 ; m2 Þ ¼ arg maxðr;sÞ dðr; sÞ , where d ð; Þ is the
Euclidean distance measure.
1
(b) Find P⊥ ¼ I U e# ¼ I U
e 2U e2 U e TUe2 e T , with U
U e 2 ¼ ½m2 m1 and
e U2
2 2 2
V ðm1 ; m2 Þ ¼ dðm1 ; m2 Þ.
(c) Set j ¼ 3.
2. At the jth recursion:
⊥
(a) Find m j ¼ arg maxer Pe r
e e via (11.22).
U j1
(b) Calculate
⊥ ⊥
e j ¼
m mj
e
Pe
U j1 : ð11:35Þ
e⊥
(c) Compute, using m j obtained in step 2(b),
⊥
T
P⊥ ⊥
ej ¼ PU
ej1 β u
ej u
ej1 m ej ;
ej1 m ð11:36Þ
U
e# m
e j1 U
ej1 ¼ U
where u j1 e j .
4. In the meantime we can also find the SV via (11.27) in step 2(c).
5. Most importantly, the SV calculation equation specified by (11.27) is exactly the
same as (11.21) but can be recursively implemented by RHSP-OPSGA to find
SVs of growing simplexes.
All the algorithms derived from DSGA can be considered variants of simplex
growing volume analysis (GSVA)-based algorithms. These include SGA,
RT-SGA, GSGA, OPSGA, and a recently developed distance-based SGA (Dist-
SGA) (Wang et al. 2013). However, depending on how initial conditions are
implemented, these GSVA-based algorithms also produce different results. For
example, an initial condition can be initialized by one endmember m1 with zero
simplex volume or 1D two-vertex simplex with maximal volume, which is the
maximal Euclidean distance of a pair (m1, m2) among all data sample vectors. This
section takes initial conditions into account to derive various GSVA algorithms.
Since initial conditions are applied to all GSVA variants and SGA-like algorithms,
the terminology of GSVA is used for a generic and general purpose, while the
algorithm’s implementation is described by OPSGA for illustration. Nevertheless,
the same algorithmic description can also be applied to other SVA-based algo-
rithms, such as DSGA, GSGA, and Dist-SGA.
11.6.1 1-GSVA
The first version of GSVA starts the initial condition with a zero-dimensional
one-vertex simplex of zero SV where a one-vertex simplex is simply a single data
sample vector and can be considered a degenerated simplex. Using such a
degenerated simplex we can derive a special version of 1-GSVA.
1-GSVA
1. Initial condition:
(a) Find m1 with maximal vector length, that is, m1 ¼ arg maxr rT r .
T
1 T
(b) Find P⊥ U1 ¼ I U1 U1 ¼ I U1 U1 U1
#
U1 , with U1 ¼ ½m1 and
pffiffiffiffiffiffiffiffiffiffiffiffiffi
GSVðm1 Þ ¼ m1 ¼ m1 m1 , where U1 ¼ ½m1 ¼ U
T e 1.
(c) Set j ¼ 2.
2. At the jth recursion:
⊥
(a) Find m
e j ¼ arg maxer P e r via (11.22), where e
r ¼ r m1 ,
U j1
e
P⊥ ¼IU e# .
e j1 U
e
U j1
j1
336 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
⊥ ⊥
(b) Calculate m
e j ¼ P
U e
ej1 .
m j
(c) Compute
⊥
e⊥
using m j obtained in step 2(b).
(d) Use the following recursive equation to calculate Pe j ⊥ via P⊥
e :
Uj U j1
T
P⊥ ⊥
ej ¼ PU
ej1 β u
ej u
ej1 m ej ;
ej1 m ð11:36Þ
U
e# m
e j1 U
ej1 ¼ U
where u j1 e j . Continue.
A second version of the GSVA is called the adaptive GSVA (AGSVA), which can
adapt its initial condition to different needs by starting with any dimensional
simplex as its initial condition. In particular, let m1, . . ., mj be the initial
j endmembers with j 3 that are provided a priori by any means, either through
visual inspection or from prior knowledge. To specify how many endmembers are
used as initial endmembers, AGSVA using ninitial endmembers as its initial
endmembers is denoted by ninitial-GSVA. With this definition, 1-GSVA and
GSVA can be considered special cases of ninitial-ROPSGA with ninitial ¼ 1 and
2, respectively.
11.7 Determining the Number of Endmembers for RHSP-OPSGA 337
ninitial-RHSP-GSVA
1. Initial condition:
(a) Let ninitial ¼ j and find j endmembers, m1, . . ., mj that yield a j-vertex simplex
with the maximal volume, GSV(m1, . . ., mj).
1
(b) Find P⊥ ¼ I U e# ¼ I U
e jU ej Ue TUej eT ,
U with ej ¼
U
e j j j
Uj
m2 m1 mj m1 and the simplex volume V(m1, . . ., mj)
2. At the ( j + 1)th recursion:
⊥
(a) Find m
e jþ1 ¼ arg maxer P e r via (11.22).
Uj
e
⊥ ⊥
(b) Calculate m e jþ1 ¼ P
U ej
e
m jþ1
.
⊥
e jþ1 obtained in step 2(b),
(c) Compute, using m
⊥
e jþ1 GSV m1 ; . . . ; mj :
GSV m1 ; . . . ; mjþ1 ¼ ð1=jÞm ð11:27Þ
T
P⊥
e ¼ P⊥
e β u
e jþ1 u
ej m e jþ1 ;
ej m ð11:36Þ
U jþ1 Uj
ej ¼ U
where u e #e
e jU
j r.
In OPSGA and RHSP-OPSGA, the value of p, that is, the total number of
endmembers, is known and fixed. When its value is unknown, it must be determined
in an unsupervised fashion. The concept of VD, recently developed in Chang
(2003) and Chang and Du (2004), has been widely used for this purpose. In this
section, we take a rather different approach using real target data sample vectors
338 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
e pT P⊥ m
ηp ¼ m ep ð11:37Þ
e U p1
T T
e RHSP -OPSGA ⊥ ⊥ e RHSP-OPSGA
¼ m p P e P e m p :
E p1 E p1
ð11:40Þ
T
⊥ ⊥ ⊥ ⊥
Because Pe is symmetric and idempotent, that is, Pe Pe ¼ Pe ,
E p1 E p1 E p1 E p1
(11.38) is reduced to
⊥ RHSP-OPSGA 2 RHSP-OPSGA T ⊥ RHSP-OPSGA
P ep
m ¼ m
ep Pe ep
m : ð11:41Þ
eE p1 E p1
⊥
Now if we replace Pe in (11.41) with P⊥ e , with Ep1 ¼ Up1, (11.41)
E p1 U p1
becomes (11.37) and (11.39). This shows that the stopping rule using (11.41) to
terminate RHSP-OPSGA is identical to the signal sources used under each
hypothesis by Neyman–Pearson detection theory to determine the VD.
It is this sequence, {ηp} given by (11.39), that will be used as the signal sources
in a binary composite hypothesis testing problem to determine whether or not the
pth potential endmember candidate m e RHSP -OPSGA is a true endmember by a detector,
p
formulated as follows:
H 0 : ηp p ηp H0 ¼ p0 ηp
versus for p ¼ 1, 2, . . . , L ; ð11:42Þ
H 1 : ηp p ηp H1 ¼ p1 ηp
where the alternative hypothesis H1 and the null hypothesis H0 represent two cases
e pRHSP-OPSGA being an endmember signal source under H1 and not an endmember
of m
signal source under H0 respectively in the sense that H0 represents the maximum
residual resulting from the background signal sources, while H1 represents the
maximum residual leaked from the endmember signal sources. Note that when
p ¼ 0, η0 is undefined in (11.42). To make (11.42) work, we need to find probability
distributions under both hypotheses. The assumption made on (11.42) is that if a
signal source is not an endmember under H0, it should be considered part of the
background, which can be characterized by a Gaussian distribution. On the other
340 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
h
i
Since pηp ηp ¼ p H 0 ; ηp þ p H 1 ; ηp ¼ 1=ηp1 ηp pνp ηp þ Fνp ηp , we
can obtain an a posteriori probability distribution of p(H0|ηp) given by
ηp pνp ηp
p H 0 ηp ¼
ð11:46Þ
η p pν p η p þ Fν p η p
Fνl ηp
p H 1 ηp ¼
: ð11:47Þ
ηp pνp ηp þ Fνp ηp
>
> 1, if Λ ηp > τp,
<
δNP
RHSP-OPSGA ηp ¼ 1 with probability κ, if Λ ηp ¼ τp , ð11:48Þ
>
>
: 0, if Λ ηp < τp ;
where the likelihood ratio test Λ(ηp) is given by Λ ηp ¼ p1 ηp =p0 ηp , with p0(ηp)
and p1(ηp) given by (11.44) and (11.45). Thus, according to (11.29), a case of
ηp > τp indicates that δNPRHSP - OPSGA (ηp) in (11.48) fails the test, in which case
11.8 Real Image Experiments 341
e RHSP
m -OPSGA is assumed to be an endmember. Note that the test for (11.48) must be
p
performed for each of the L potential endmember candidates. Therefore, for a
different value of p the threshold τp varies. Using (11.48) the value of VD, nRHBP-
OPSGA, can be determined by calculating
n h io
RHSP-OPSGA
VDNP
RHSP-OPSGA ð P F Þ ¼ arg max p δ NP
RHSP-OPSGA ηp ¼ 1 ; ð11:49Þ
where PF is a predetermined false alarm probability, δNP - η ¼ 1 only if
NP
RHSP
p
OPSGA
δRHSP-OPSGA ηp ¼ 1, and δRHSP-OPSGA ηp ¼ 0 if δRHSP-OPSGA ηp < 1. Note that
NP NP
since p starts with 2, the actual value of VD estimated for RHSP-OPSGA in (11.49)
should be VDNP RHSP-OPSGA ðPF Þ þ 1.
Since the sequence of subspaces {Ẽp} is nested in terms of E e1
Ee p , the
⊥
sequence of maxer Pe e r produced by (11.38) is monotonically decreasing.
E p1
This further implies that the sequence of {ηp} produced by (11.39) is also mono-
tonically decreasing. Using this fact, the sequence of detectors {δNP (ηp)}
- OPSGA
RHSP
starts with a failure of the test specified by (11.48), that is, δRHSP-OPSGA ηp ¼ 1, and
NP
1=2
pffiffiffiffiffi T ⊥
ρp ¼ ηp ¼ e e
m p Pe m p : ð11:50Þ
U p1
Although many real hyperspectral image scenes can be used for experiments, their
conclusions are very similar. In this case, we used two real image scenes that have
been studied extensively in the literature for comparative analysis.
342 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
The image scene shown in Fig. 11.5 (also in Fig. 1.10a) was used for the experi-
ments. It was acquired by the airborne HYperspectral Digital Imagery Collection
Experiment (HYDICE). It has a size of 64 64 pixel vectors with 15 panels in the
scene and the ground truth map in Fig. 11.5b (Fig. 1.10b).
It is worth noting that panel pixel p212, marked in yellow in Fig. 11.5b, is of
particular interest. Based on the ground truth, this panel pixel is not a pure panel
pixel and is marked by yellow as a boundary panel pixel. However, with our
extensive and comprehensive experiments, this yellow panel pixel is always
extracted as the one with the most spectrally distinct signature compared to R
panel pixels in row 2. This indicates that the signature of spectral purity is not
equivalent to the signature of spectral distinction. As a matter of fact, in many cases
panel pixel p212 instead of panel pixel p221 is the one extracted by EFAs to represent
the panel signature in row 2. Also, because of such ambiguity, the panel signature
representing panel pixels in the second row is either p221 or p212, which is always
the last one to be found by EFAs. This implies that the ground truth of R panel
pixels in the second row in Fig. 11.5b may not be as pure as was thought.
Even though the RHSP-OPSGA can be used to find endmember candidates up to
the number of total bands, L, there is no necessity for doing so. Using (11.37) and
(11.50) we can estimate the number of endmembers that must be generated by the
RHSP-OPSGA. Table 11.1 tabulates nRHBP-OPSGA estimated by the MOCA as well
a b
p11, p12, p13
Fig. 11.5 (a) HYDICE panel scene containing 15 panels and (b) ground truth map of spatial
locations of the 15 panels
as NPD for various false alarm probabilities, where MOCA is actually a maximum
likelihood detector.
According to Chang (2003) and Chang and Du (2004), the VD estimated for this
HYDICE scene by the Harsanyi–Farrand–Chang (HFC) method (Harsanyi et al.
1994) would be 9. Recently, an approach using real target signal sources proposed
for VD estimation in Chang et al. (2014) and described in Chap. 4 was used to
estimate nRHBP-OPSGA to be in a range between 39 and 45. However, when
unsupervised fully constrained least squares (FCLS) was used in Heinz and
Chang (2001) to find targets, the number of targets was 34. Thus, the values of
VD, nVD, was set from 9 to 45. Figure 11.6 shows endmember pixels found by the
four EFAs RHSP-OPSGA, OPSGA, DSGA, and a new distance-based SGA (Dist-
SGA) developed in Wang et al. (2013a). However, looking more closely at the
development of Dist-SGA, we see that the used distance is actually the orthogonal
distance, which is exactly an OP. Accordingly, it is not surprising to see from these
figures that the RHSP-OPSGA, OPSGA, and Dist-SGA produced identical
endmember pixels, while the DSGA produced completely different results. The
reason the DSGA produced different endmembers from the other algorithms is
mainly that the SV calculated by the DSGA using a determinant via SVD to find
endmember pixels is different from that calculated by the OPSGA, RHSP-OPSGA,
and Dist-SGA using the OP. Interestingly, none of the four algorithms could find
pure panel pixels corresponding to the five panel signatures when nVD ¼ 9, but they
could do so, as shown in Fig. 11.6b, when nVD ¼ 18, which is twice value of
nVD ¼ 9, a fact that was also noted in Chang et al. (2010a, 2011b). Since the four
algorithms are sequential, the endmember pixels generated by a smaller value of
nRHBP-OPSGA are always part of endmember pixels generated by a larger value of
nRHBP-OPSGA, and the endmember pixels found in Fig. 11.6 are separated into
four different ranges: (a) nVD ¼ 9, (b) nVD ¼ 18, (c) nRHBP-OPSGA ¼ 34, and
(d) nRHBP-OPSGA ¼ 45.
As expected, the Dist-SGA, OPSGA, and RHSP-OPSGA found identical sets of
endmember pixels owing to the fact that the maximal OP of a vertex perpendicular
to a simplex base is indeed its height.
Fig. 11.6 Endmember pixels found by DSGA, Dist-SGA, OPSGA, and RHSP-OPSGA on
HYDICE. (a) 9 endmember pixels. (b) 10–18 endmember pixels. (c) 19–34 endmember pixels.
(d) 35–45 endmember pixels
data, are also available for study. There are five pure pixels in Fig. 11.7a, b, which
can be identified as corresponding to five different minerals: alunite (A),
buddingtonite (B), calcite (C), kaolinite (K), and muscovite (M), labeled by A,
B, C, K, and M in Fig. 11.7b, along with their spectral signatures plotted in
Fig. 11.7c, d.
Although the data set contains more than five minerals in the data set, the ground
truth available for this region only provides the locations of the pure pixels: alunite
11.8 Real Image Experiments 345
(A), buddingtonite (B), calcite (C), kaolinite (K), and muscovite (M), shown in
Fig. 11.7b with their reflectance and radiance spectra plotted in Fig. 11.7a, b.
Since there is no available prior knowledge about the spatial locations of
endmembers, we must rely on an unsupervised means of identifying whether an
extracted target pixel is an endmember. To address this issue, the following
Endmember IDentification algorithm (EIDA), developed in Chang et al. (2014),
was used for this purpose.
EIDA
J p
1. Assume that tj j¼1 are J extracted target pixels and fmi gi¼1 are known p ground
truth endmembers.
J p
2. Cluster all extracted pixels, tj j¼1, into p endmember classes Cj j¼1 according
to the following rule:
tj 2 Cj* , j* ¼ arg min1ip SAM tj ; mi ; ð11:52Þ
Fig. 11.7 (a) Cuprite AVIRIS image scene. (b) Spatial positions of five pure pixels corresponding
to the following minerals: alunite (A), buddingtonite (B), calcite (C), kaolinite (K), and muscovite
(M); (c) five mineral reflectance spectra; (d) five mineral radiance spectra
346 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
Table 11.2 nRHBP-OPSGA estimated for Curpite reflectance data by (11.37) and (11.50)
ηp in (11.37) MOCA PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
DSGA F F F 119 119 117
RHSP-OPSGA 75 82 74 69 64 62
ρp in (11.50) MOCA PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
DSGA 119 119 119 117 103 95
RHSP-OPSGA 75 82 74 69 64 62
Table 11.3 nRHBP-OPSGA estimated or Cuprite radiance data by (11.37) and (11.50)
ηp in (11.37) MOCA PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
DSGA F F 109 106 104 95
RHSP-OPSGA 72 79 72 69 65 62
ρp in (11.50) MOCA PF ¼ 101 PF ¼ 102 PF ¼ 103 PF ¼ 104 PF ¼ 105
DSGA 103 106 103 95 95 85
RHSP-OPSGA 67 71 67 62 58 56
nVD ¼ 139 for Cuprite radiance data. It should be noted that for Cuprite data the
DSGA could not produce 189 target signal sources owing to its use of SVD to
calculate SV.
In addition, according to Chang (2003) and Chang and Du (2004), the value of
the VD estimated by the HFC method was nVD ¼ 22 for reflectance data with
2nVD ¼ 44 and nVD ¼ 15 for radiance data with 2nVD ¼ 30. Using these values,
along with the values of nRHBP-OPSGA in Tables 11.2 and 11.3, the experiments were
conducted for three ranges: (a) nVD ¼ 22 for both reflectance and radiance data,
(b) nRHBP-OPSGA ¼ 46 for both reflectance and radiance data, and (c) nRHBP-
OPSGA ¼ 75 for both reflectance and radiance data. The reason that we could do
this is because the DSGA, Dist-SGA, OPSGA, and RHSP-OPSGA are all sequen-
tial algorithms and the endmember pixels generated by a smaller value of nRHBP-
OPSGA are always part of the endmember pixels generated by a larger value of
nRHBP-OPSGA.
Figures 11.8 and 11.9 show the endmembers found by the DSGA, OPSGA,
RHSP-OPSGA, and Dist-SGA for Cuprite reflectance data and Cuprite radiance
data, respectively, where the pixels marked by yellow open circles were found by
the algorithms using numerals used to indicate their found orders.
To identify which pixels among the found endmembers in Figs. 11.8 and 11.9 are
desired endmember pixels, the EIDA, described earlier, was used to find the spatial
locations of these desired endmember pixels shown in Figs. 11.10 and 11.11, where
the pixels marked by lowercase a, b, c, k, and m with red triangles are the desired
endmember pixels found by the EIDA that correspond to the five ground truth
mineral endmembers marked by uppercase A, B, C, K, and M with yellow crosses.
Table 11.4 further tabulates the five desired endmember pixels found by the
EIDA among target pixels found by the DSGA, {tDSGA j } and target pixels found by
OP - SGA
the OPSGA, {tj } where the subscript j indicates the order in which a
particular target pixel was found.
348 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
Fig. 11.8 Endmember pixels found by DSGA without DR, Dist-SGA, OPSGA, RHSP-OPSGA
on Cuprite reflectance data. (a) 22 endmember pixels. (b) 23–46 endmember pixels. (c) 47–75
endmember pixels
As we can see from Table 11.4 for the reflectance data, the DSGA required at
least 64 target pixels to find the last mineral signature K to complete all five mineral
signatures, while 53 target pixels were required by the RHSP-OPSGA to find the
last mineral signature B. For radiance data it took the DSGA at least 103 target
pixels to find the last mineral signature K to complete all five mineral signatures,
while 68 target pixels were required by the RHSP-OPSGA to complete all five
mineral signatures with the mineral signature C being the last one to be found.
Nonetheless, both the DSGA and RHSP-OPSGA found the same first mineral
signature, which was M because the spectrum of M is probably the most distinct
among the five mineral signatures according to Fig. 11.7a, b. Once the spatial
locations of the five desired endmember pixels were found in Figs. 11.8 and 11.9,
we can further perform a comparative spectral analysis between EIDA-identified
endmember pixels in Table 11.4 and the ground truth pixels in Fig. 11.7b. For each
of the spectral signatures of the five mineral signatures, A, B, C, K, and M, for
reflectance data and radiance data, Figs. 11.12a–e and 11.13a–e plot three spectra,
ground truth pixel spectrum, and the spectra of tDSGA j and tjR - OPSGA identified in
Table 11.4.
11.8 Real Image Experiments 349
Fig. 11.9 Endmember pixels found by DSGA without DR, Dist-SGA, OPSGA, RHSP-OPSGA
on Cuprite radiance data. (a) 22 endmember pixels. (b) 23–46 endmember pixels. (c) 47–75
endmember pixels
Fig. 11.10 Endmembers found by EIDA using DSGA and RHSP-OPSGA for Cuprite reflectance
data. (a) DSGA. (b) RHSP-OPSGA
350 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
a b
a
A A
c
b B b B
m M m M
K k K
k
c
a
C C
Fig. 11.11 Endmembers found by EIDA using DSGA and RHSP-OPSGA for Cuprite radiance
data. (a) DSGA. (b) RHSP-OPSGA
While the plots of Figs. 11.12 and 11.13 offer the advantage of visual assessment
about how close a found endmember pixel is to a ground truth pixel, it does not
provide quantitative measurements on their spectral similarity. Tables 11.5 and
11.6 calculate the spectral similarity values of the plots in Figs. 11.12 and 11.13,
respectively, where SAM and spectral information divergence (SID) were used as
spectral measures. As we can see in Tables 11.5 and 11.6, the spectral similarity
values between pixels found by the DSGA and RHSP-OPSGA compared to ground
truth pixels were indeed very close even though the pixels found by the DSGA and
RHSP-OPSGA were identified by EIDA in different locations.
One of the great advantages that the RHSP-OPSGA provides is computational cost
savings. Figures 11.14 and 11.15 show the cumulative computing time in seconds
required by the four algorithms DSGA, Dist-SGA, OPSGA, and RHSP-OPSGA to
run HYDICE and Cuprite data sets respectively as the number of endmembers, p,
11.9 Computer Processing Time Analysis 351
a b
6000 6000
5500 5500
5000 5000
4500 4500
Reflectance
4000
Reflectance
4000
3500 3500
3000 3000
2500 2500
Reflectance
3500
2500
3000
2000
2500
1500 2000
Groundtruth 1500 Groundtruth
1000 DSGA DSGA
RHSP-OPSGA 1000 RHSP-OPSGA
500 500
0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200
Band Band
e
7000
6000
5000
Reflectance
4000
3000
Groundtruth
2000 DSGA
RHSP-OPSGA
1000
0 20 40 60 80 100 120 140 160 180 200
Band
Fig. 11.12 Comparative plots between spectral signatures found by DSGA and R-OPSGA
and ground truth pixels for Cuprite reflectance data. (a) “a” signature against “A.” (b) “b”
signature against “B.” (c) “c” signature against “C.” (d) “k” signature against “K.” (e) “m”
signature against “M”
increases to 169, where the computer environment was executed on the following
system: Windows operating system, 7 64 bits, with an Intel Core i5-2500 CPU @
3.30 GHz; Memory: 8.00 GB; programming language: MATLAB V8.0. Note that
the computer processing time to run the Cuprite reflectance and radiance sets was
nearly the same because the same image scene was used. Thus, there is no visible
difference. In this case, only reflectance data were plotted. As shown in Figs. 11.14a
and 11.15a, the DSGA required extremely high computing time compared to the
352 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
a b
12000 9000
Groundtruth Groundtruth
DSGA 8000 DSGA
10000 RHSP-OPSGA RHSP-OPSGA
7000
8000 6000
Radiance
Radiance
5000
6000
4000
4000 3000
2000
2000
1000
0 0
0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200
Band Band
c d
8000 10000
Groundtruth Groundtruth
7000 9000
DSGA DSGA
RHSP-OPSGA 8000 RHSP-OPSGA
6000
7000
5000
6000
Radiance
Radiance
4000 5000
3000 4000
3000
2000
2000
1000
1000
0 0
0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200
Band Band
e
12000
Groundtruth
DSGA
10000 RHSP-OPSGA
8000
Radiance
6000
4000
2000
0
0 20 40 60 80 100 120 140 160 180 200
Band
Fig. 11.13 Comparative plots between spectral signatures found by DSGA and R-OPSGA
and ground truth pixels for Cuprite radiance data. (a) “a” signature against “A.” (b) “b”
signature against “B.” (c) “c” signature against “C.” (d) “k” signature against “K.” (e) “m”
signature against “M”
other three, so there is no visible difference among the other three due to the large
magnitude caused by the DSGA. To provide a better visual assessment, Figs. 11.14b
and 11.15b replot the computing times for the RHSP-OPSGA, OPSGA, and Dist-
SGA on the same scale, where the RHSP-OPSGA was slightly better than the
OPSGA, but both the OPSGA and RHSP-OPSGA were indeed better than the Dist-
SGA in terms of computing time.
11.9 Computer Processing Time Analysis 353
Table 11.5 SAM/SID of closet endmembers to ground truth endmembers by DSGA and RHSP-
OPSGA on Cuprite reflectance data
Table 11.6 SAM/SID of closet endmembers to ground truth endmembers by DSGA and RHSP-
OPSGA on Cuprite radiance data
a b
1000 4
DSGA
Accumulative computing time (sec)
Accumulative computing time (sec)
500 2
400 1.5
300
1
200
100 0.5
0 0
0 20 40 60 80 100 120 140 160 180 0 20 40 60 80 100 120 140 160 180
Number of endmembers (p) Number of endmembers (p)
Fig. 11.14 Cumulative computing time in seconds of DSGA without DR, DSGA, Dist-SGA,
OPSGA, RHSP-OPSGA as p increases on HYDICE data
Tables 11.7 and 11.8 also tabulate computer processing times required for
the RHSP-OPSGA, OPSGA, Dist-SGA, and DSGA to generate Figs. 11.6
and 11.11.
As we can see from Tables 11.6 and 11.7, the RHSP-OPSGA was the best,
requiring the least computing time, while the DSGA was the worst, requiring a
much higher computing time. In addition, the more endmember pixels that had to be
found, the more computational savings can be achieved. It is worth noting that the
Dist-SGA was shown to be the best in Wang et al. (2013) among all variants of
SGA. Figures 11.14 and 11.15 and Tables 11.6 and 11.7 further show that our
developed OPSGA and RHSP-OPSGA outperformed the Dist-SGA.
354 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
a b
12000 140
2000 20
0 0
0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200
Number of endmembers (p) Number of endmembers (p)
Fig. 11.15 Cumulative computing time in seconds of DSGA without DR, DSGA, Dist-SGA,
OPSGA, RHSP-OPSGA as p increases on Cuprite data
To conclude this section, some comments are noteworthy. It is known that VCA
is also an OP-based algorithm and has become popular for finding endmembers in
recent years. However, it suffers from several implementation issues. One is its use
of random initial conditions that result in finding inconsistent sets of endmembers.
Another is that it generally requires preprocessing prior to finding endmembers
such as DR transform to reduce data volumes. As shown in Chang (2013), different
DR transforms generally produce different sets of endmembers. Most importantly,
VCA is not a fully constrained algorithm that generally finds suboptimal solutions.
It is shown that VCA cannot compete against ATGP in the sense of finding the
maximal OP or against the SGA in the sense of finding the maximal SV. Compared
to VCA, Wang et al.’s Dist-SGA and the proposed OPSGA, along with RHSP-
OPSGA, have no such issues. The only major advantage that VCA can offer is its
lower computational complexity, which requires only inner products. But this
advantage can also be gained by the RHSP-OPSGA, which requires only inner
and outer products of vectors, none of which are matrix multiplications.
11.9 Computer Processing Time Analysis 355
a b
10 140
ATGP ATGP
9
Accumulative computing time (sec)
3 40
2
20
1
0 0
0 20 40 60 80 100 120 140 160 180 0 20 40 60 80 100 120 140 160 180 200
Number of endmembers (p) Number of endmembers (p)
Fig. 11.16 Computing times for ATGP, VCA with/without DR, Dist-SGA, OPSGA, and RHSP-
OPSGA. (a) HYDICE data. (b) Cuprite data
Figure 11.16a, b shows that the computing time of VCA varies with data sets and is
not stable compared to other algorithms because VCA with no DR required the least
computing time for both data sets, while VCA had the worst computing time for
HYDICE but the second best computing time for Cuprite data. Overall, the Dist-
SGA required the most computing time among all six test algorithms (except VCA
with DR in the HYDICE data set), while the OPSGA and RHSP-OPSGA are ranked
as the third best algorithm next to VCA with no DR and the ATGP. However, since
VCA is not designed to find the maximal SV, according to Chang (2013) and Chen
et al. (2014), despite its savings in computing time, VCA did not produce good
endmember results, except when VCA used ATGP-generated target pixels as initial
conditions. It is also interesting to see that VCA with DR required the most time for
HYDICE (Fig. 11.16a) but had the second best time for Cuprite data (Fig. 11.16b).
This is because DR is performed prior to VCA. When the data size is very large, as
with Cuprite, VCA with DR actually takes advantage of a DRHSP-reduced data
cube to save computing time owing to the fact that DR is only a one-time operation.
As a result, VCA with/without DR achieved the best computing times (Fig. 11.16b)
regardless of its performance.
Another comment is in order. Since VCA makes use of random initial conditions
for each of the endmembers it generates, it is interesting to see how random initial
conditions affect its final endmember results. In Chang (2013) and Chen et al.
(2014), three initial conditions were investigated for VCA with no DR: (1) initial
conditions generated by Gaussian random variables originally used by VCA
Nascimento and Dias (2005); (2) the unity vector with all components specified
0 1T
by 1s, that is, @1, 1, . . . , 1 A ; and (3) ATGP-generated target pixels. It was shown
|fflfflfflfflfflffl{zfflfflfflfflfflffl}
L
in Chang (2013) and Chen et al. (2014) that the best performance produced by VCA
was the one that used ATGP-generated target pixels as the initial conditions.
356 11 Recursive Hyperspectral Sample Processing of Orthogonal Projection-Based. . .
Interestingly, in this case, VCA was simply reduced to the ATGP, as shown in
Chang (2013) and Chen et al. (2014). Moreover, it was also shown in Chang (2013)
and Chen et al. (2014) that if VCA is implemented with DR or using random initial
conditions, VCA generally did not perform as well as the ATGP.
Finally, we would point out that as long as the maximal OP and maximal SV are
used as criteria for finding endmembers, VCA may not be effective, even though
VCA with/without DR may require the least computing time (Fig. 11.16). Unfor-
tunately, this must be traded off against its suboptimal performance as a compro-
mise, for which we consider it not worthwhile. All things considered, the RHSP-
OPSGA is generally preferred among all currently available SV-based EFAs in
terms of performance and computing efficiency.
11.10 Conclusions
Owing to its simple architecture and competing performance, the SGA has emerged
as a good alternative to N-FINDR when it comes to reducing computational
complexity and savings. However, calculating an SV remains a challenging issue
in terms of computational cost. This chapter develops an OP-based SGA (OPSGA)
that transform the calculation of an SV into finding an OP. In particular, a recursive
version of OPSGA, RHSP-OPSGA, is derived to allow OPSGA to perform the SGA
via an OP in a recursive manner. The RHSP-OPSGA offers several significant
benefits. First, it provides an effective means of determining how many
endmembers the RHSP-OPSGA needs to generate in an unsupervised fashion
versus the SGA, which must know this number in advance prior to implementation.
Another benefit is that it reduces the computational cost of calculating an SV to
only finding inner and outer products of vectors, not matrices, specifically, no
matrix inverses are required during recursion. Finally, the recursive equation
developed to implement the RHSP-OPSGA is a Kalman filter-like filter that can
be easily implemented in hardware design.
Chapter 12
Recursive Hyperspectral Sample Processing
of Geometric Simplex Growing Algorithm
Abstract Simplex volumes (SVs) have been used in the literature as a criterion for
finding endmembers. A main issue that arises in finding SVs is inverting a
nonsquare matrix, which involves excessive computing time in calculating the
matrix determinant. This type of SV calculation is referred to as a determinant-
based SV (DSV) calculation (Chap. 2). Therefore, several preprocessing steps are
suggested for DSV calculation in Chap. 2 to ease computational complexity, such
as reducing data dimensionality, easing computational complexity by manipulating
determinant calculation, narrowing search regions via Pixel Purity Index (PPI), or
developing algorithms such as the simplex growing algorithm (SGA) to grow
simplexes one after another sequentially instead of finding all endmembers together
simultaneously. However, all these DSV calculation techniques remain stuck with
certain inherent drawbacks encountered in finding SVs using a matrix determinant
calculation, in addition to another issue: the calculated SV may not be a true SV, as
pointed out in Chap. 2. To resolve this dilemma, Chap. 11 developed an orthogonal
projection (OP)-based growing simplex volume analysis (GSVA) approach, called
orthogonal projection SGA (OPSGA), which calculates a geometric SV (GSV)
from a simplex geometry point of view instead of resorting to the matrix determi-
nant. It transforms finding a new endmember that yields the maximal DSV by
growing previously simplexes into finding an endmember with the maximal OP
onto a hyperplane that is linearly spanned by simplexes specified by previously
found endmembers. Accordingly, OPSGA can be considered a technique for
calculating GSV by OP (GSV-OP). As an alternative to OPSGA, this chapter
presents another GSVA approach to finding GSV, referred to as the geometric
SGA (GSGA), developed by Chang, Li, and Song (IEEE Journal of Selected Topics
in Applied Earth Observations and Remote Sensing 99:1–13, 2016c), which con-
verts GSV calculation into a product of the height and base of a simplex. In parallel
to OPSGA, GSGA can also be considered a technique for calculating the GSV by
the simplex height (GSV-SH). In other words, finding the maximal SV is equivalent
to finding the maximal height of a simplex. As a result, when simplexes grow one
vertex at a time, GSV can be calculated by multiplying only successive heights
where the heights can be found by the Gram–Schmidt orthogonalization process
(GSOP). This is because the bases of previously grown simplexes are already
known and need not be recalculated. Furthermore, in analogy with recursive
hyperspectral sample processing of OPSGA (RHSP-OPSGA), a recursive
hyperspectral sample processing of GSGA (RHSP-GSGA) can also be derived
12.1 Introduction
As discussed in Chang (2016), the two most widely used algorithms for finding
endmembers are the OP-based Pixel Purity Index (PPI) developed by Boardman
(1994) and the simplex volume (SV)-based N-finder algorithm (N-FINDR) devel-
oped by Winter (1999a, b). In recent years N-FINDR seems to have emerged as a
preferred endmember finding algorithm (EFA) over other EFAs using different
criteria such as OP because its use of SVs as a criterion for optimality turns out also
to satisfy the convexity required for the purity of endmembers by imposing two
physical abundance constraints, abundance sum-to-one constraint (ASC) and abun-
dance nonnegativity constraint (ANC), used by linear spectral mixture analysis
(LSMA) for linear spectral unmixing (LSU).
Although SGA significantly reduces computer processing time for N-FINDR,
finding SVs generally requires calculating the determinant of a nonsquare matrix
defined in (2.1) in Chap. 2 by
" #
1 1
1 1
Det
m1 m2 mp1 mp
DSV m1 ; m2 ; . . . ; mp ¼ ; ð12:1Þ
p!
12.1 Introduction 359
p
where mj j¼1 are p endmembers. This is because the dimensionality of the
simplexes specified by p endmembers is p 1, which is relatively smaller com-
pared to the total number of bands, L. In this case, SGA repeatedly uses (12.1) to
calculate DSV via matrix determinants. Under such a circumstance two approaches
are usually adopted. One is to perform data dimensionality reduction (DR) to make
a nonsquare matrix a square matrix. The other is to perform singular value decom-
position SVD to find a square submatrix. Unfortunately, neither approach gives true
a SV, as discussed in Chap. 2. This issue was overlooked since all simplex-based
EFAs are designed to find endmembers, not to find true SVs. As a result, finding
true SVs is immaterial to EFAs. In fact, it was shown in Li et al. (2015a) and
Chap. 2 that the determinant-based SV (DSV) calculated by N-FINDR and SGA are
not true SVs. Similarly, when vertex component analysis (VCA), using the maxi-
mal OP, is implemented to find endmembers via convex hulls; whether or not SVs it
finds yield maximal SVs is not important. As a matter of fact, it is also shown in
Chap. 4 in Chang (2016), Chang et al. (2016b), and Chen et al. (2014a, b, c) that the
OP found using VCA is indeed not necessarily a maximal SV. Furthermore, the
convex hulls found by VCA do not yield maximal convex cone volume either, as
shown in Chap. 11 in Chang (2016) and in Chen et al. (2014a, b, c). The main issue
arises in VCA from the fact that the DSV or convex cone volume calculated by
vertices or endmembers is not a true SV because the vertices found by VCA do not
satisfy convexity constraints, and the dimensionality of the simplex and convex
cone formed by these vertices is always one dimension less than the total number of
vertices. To resolve this issue, the matrix determinant used to calculate SVs must
include a unity vector with ones in all components as an additional row in (2.1) in
Chap. 2 or (12.1) to impose ASC. Second, the computing time involved in finding a
DSV grows exponentially with increasing numbers of endmembers because the
computing time required in inverting larger matrices to find matrix determinants via
SVD is increased substantially. Third, DSV calculation may run into an inevitable
numerical issue as the number of endmembers, p, grows.
More specifically, assume that a given p-vertex simplex is formed by a set of
p
p endmembers mj j¼1 already found by a sequential EFA, which finds
endmembers one at a time. To find the ( p + 1)th endmember, mj+1, as a new
endmember, the DSGA grows ( p + 1)-vertex simplexes from the p-vertex simplex
p
specified by mj j¼1 and finds a particular ( p + 1)-vertex simplex with the mj+1 as a
new vertex that yields the maximal SV among all possible ( p + 1)-vertex simplexes
by calculating matrix determinants.
The OPSGA developed in Chap. 11 considers a hyperplane linearly spanned by
p p
the p previously found mj j¼1 with the p-vertex simplex specified by mj j¼1 ,
denoted by S(m1, m2, . . ., mp) embedded in this hyperplane. It then finds the mj+1 as
a new endmember that yields the maximal OP via P⊥ U in (4.19), with the undesired
p
signature matrix U formed by mj j¼1 . It turns out that finding such a maximal OP
can be carried out by the automatic target generation process (ATGP) developed in
Sect. 4.4.2.3. In other words, OPSGA can be interpreted as a new application
360 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
of GSGA over DSGA and OPSGA is that, compared to DSGA and OPSGA, which
are endmember-dependent methods, the simplexes found by GSGA have nothing to
do with the spatial locations of endmembers because they are determined by
simplex edges, not simplex vertices. In this case, GSGA is actually a simplex
edge-based technique. In particular, it is also translation-invariant and rotation-
invariant.
Analogously to RHSP-OPSGA, GSGA can also be extended to recursive GSGA
(RSGA), where two Kalman filter (KF)-like versions can be derived (Chang and Li
2016a, b). One is called KF-based orthogonal subspace projection RHSP-GSGA
(KF OSP-RHSP-GSGA), which takes advantage of OSP to derive a recursive
equation. The other is called KF-based orthogonal vector projection RHSP-
GSGA (KF OVP-RHSP-GSGA), which makes use of GSOP to derive a recursive
equation. The recursive structures of both algorithms provide further advantages in
the application of GSGA in hardware implementation and design.
The method for finding the maximal height via (12.2) for each 1 i j to calculate
maximal SV is called GSGA, as opposed to the original SGA developed by Chang
et al. (2006a, b, c), referred to as a determinant-based SGA (DSGA), which uses the
matrix determinant specified by (12.1) to find the maximal SV. Interestingly, (12.2)
also shows that GSGA is essentially the same as DSGA. Nevertheless, it is worth
noting that SV in (12.2) can only be calculated when ME is a square matrix since a
determinant is defined as a value associated with a square matrix. More details on
(12.2) can be found in Stein (1966). The GSGA presented in this section was
developed by Chang et al. (2016c) to address all the issues mentioned in the
introduction as an alternative approach to DSGA by taking advantage of simplex
geometric structures to find SVs without performing either DR or SVD.
The idea of GSGA comes very naturally from the geometric structures of a
simplex where a simplex can be considered a convex set formed by its edges instead
of vertices and its volume can be calculated by finding GSV formed by its simplex
edges. In this case, GSGA starts with its initial condition, which is one single edge,
kme 2 k ¼ km2 m1 k , connected by two endmembers m1 and m2 with maximal
length defined by l2 norm kxk2 ¼ xT x. This is because GSVðm1 ; m2 Þ ¼ km2 m1
k is the maximal GSV of a degenerated one-dimensional (1D) two-vertex simplex, S
(m1, m2). In this case, we can select either m1 or m2 as its initial endmember, and
there is no random issue encountered in GSGA. Also, it should be noted that GSGA
362 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
does not start with a single vertex since a one-vertex simplex has zero dimensions
with zero volume. This is quite different from all EFAs, which start with random
initial conditions by either randomly selecting the first endmember for initializa-
tion, such as SGA or VCA, or p endmembers, such as N-FINDR. To find a third
vertex, GSGA finds the data sample vector, denoted by m3, that yields the maximal
height h2 perpendicular to the edge vector m e 2 ¼ m2 m1 . We define m e 3 ¼ m3
m1 as a second simplex edge vector and further decompose m e 3 into m
e3 ¼ m e⊥
3
k
þme 3, where me⊥3 is the vector orthogonal to the segment m e 2 ¼ km2 m1 k, with the
⊥ jj
maximal OP given by h2 ¼ km e 3 k , and me 3 is the vector along the segment
kme 2 k ¼ km2 m1 k. These three vertices are then used to form a triangle Δ(m1,
m2, m3) as a two-dimensional
(2D) three-vertex simplex, S(m1, m2, m3) with three
3
simplex edges, 3 ¼ , specified by km2 m1 k, km3 m1 k, and km3 m2 k. Its
2
corresponding volume is calculated by GSV ¼ ðm1 ; m2 ; m3 Þ ¼ GSVðm1 ; m2 Þ h2,
with m e3 ¼ me⊥3 þm e k3 and h2 ¼ km e⊥3 k. The same procedure is repeated over and
over again to find the fourth data sample vector m4 to define a third simplex edge
vector, me 4 ¼ m4 m1 , where m e 4 can be decomposed into two vectors, me4 ¼ m e⊥4
þm e k4, where me⊥4 yields the maximal OP, defined as h3 ¼ km e⊥ 4 k, perpendicular to the
space hm e 2; m
e 3 i linearly spanned by m e 2 and m e 3 , which is a plane containing the
triangle formed by Δ(m1, m2, m3), and m e k4 is the vector lying in the space hm e ;me i.
2 3
4
We then form a 3D four-vertex simplex, S(m1, m2, m3, m4) with 6 ¼ simplex
2
edges specified by km2 m1 k, km3 m1 k, km3 m2 k, km4 m1 k, km4 m2 k,
and km4 m3 k, and the same procedure is repeated over and over again. Fig-
ure 12.1 provides an example to illustrate how m3 and m4 are found by GSGA.
Suppose that for j 2, m1, m2, . . ., mj are previously found endmembers and mj
+1 is the next endmember to be found. Assume that S(m1, m2, . . ., mj) is the ( j 1)-
dimensional j-vertex simplex formed by m1, m2, . . ., mj with the volume given by V
(m1, m2, . . ., mj). Let tj+1 be a new data sample
to be added as the ( j + 1)th vertex to
form a ( j + 1)-vertex simplex, S m1 ; m2 ; . . . ; mj ; tjþ1 . Then its SV,
V m1 ; m2 ; . . . ; mj ; tjþ1 , can be calculated proportional to the value of multiplying
V m1 ; m2 ; . . . ; mj ; tjþ1 as its base, with et jþ1 ¼ tjþ1 m1 ¼ et ⊥ ek
jþ1 þ t jþ1 and hj ¼ k
et ⊥
jþ1 k as its height, which can be obtained by finding the distance from tj+1
perpendicular to the base, denoted by hj, that is,
GSV m1 ; m2 ; . . . ; mj ; tjþ1 / GSV m1 ; m2 ; . . . ; mj hj ; ð12:3Þ
where the notation “/” means “proportional” scaled by a constant. Then the desired
( j + 1)th endmember, mj+1, is the one that maximizes V m1 ; m2 ; . . . ; mj ; tjþ1 over
tjþ1 , that is,
12.2 Geometric Simplex Growing Algorithm (GSGA) 363
m4
~ m3 ~
m m
3 h4 4
h3
m1 m2
~ ||
m
m~ || 4
O 3
4-vertex simplex with 6 edges
mjþ1 ¼ arg maxtjþ1 GSV m1 ; m2 ; . . . ; mj ; tjþ1 : ð12:4Þ
Accordingly,
the volume of a j-dimensional ( j + 1)-vertex simplex,
GSV m1 ; m2 ; . . . ; mj ; mjþ1 , can be simply calculated by multiplying GSV(m1,
m2, . . ., mj) of the simplex S(m1, m2, . . ., mj) as its base by its maximal height
hj ¼ km e⊥jþ1 k, where m e jþ1 ¼ mjþ1 m1 ¼ m e⊥ e kjþ1 . As a result, calculating
jþ1 þ m
GSV can be simply performed by successively computing (12.3) by finding mj+1
via (12.4) without calculating the matrix determinant by (12.1). Most importantly,
there are two significant differences between DSGA using (12.1) and GSGA using
(12.3). One is that DSGA calculates DSV using vertices as endmembers, with ASC
!T
1 , 1 , . . . , 1
imposed by a unity vector 1 ¼ |fflfflfflfflfflffl{zfflfflfflfflfflffl} , whereas GSGA calculates GSV
L
using p 1 simplex edge vectors without referring to vertices since GSV is
calculated by multiplying the simplex edge length km e 2 k ¼ km2 m1 k by the
product of successive heights, F, with no knowledge of the spatial positions of
endmembers. Nonetheless, the rank of the matrix in (12.1) has the same dimen-
sionality, p 1, as (12.2), which is specified by p 1 vectors, m e 2 ¼ m2 m1 ,
e 3 ¼ m3 m1 , m
m e 4 ¼ m4 m1 , . . ., m
e p ¼ mp m1 , each of which yields one
dimension. The other difference is that DSGA is determined by its initial condition
compared to GSGA, which is independent of the initial condition because the initial
p
endmember m1 has been subtracted out from all other endmembers mj j¼2 . In
other words, DSV calculated by DSGA is determined by endmembers, but GSV
calculated by GSGA is determined by relative distances among endmembers as
simplex edge lengths. Accordingly, endmembers found by DSGA are completely
determined by initial conditions and also generally require DR, which does not
364 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
which yields
" # " # " #
1 1 0 0 1
¼
r ðLþ1Þ1
m1 m2 m1 mp m1 ðLþ1Þp
e
α ðpþ1Þ1
T
where αe ðp1Þ1 ¼ α2 ; α3 ; ; αp is a ðp 1Þ-dimensional abundance vector that
removes α1 from the original p-dimensional abundance vector αp1 and
e e j ¼ mj m1 . Using ASC the abundance fraction α1 can be
r L1 ¼ rL1 m1 , m
obtained by
X
p
α1 ¼ 1 αj : ð12:8Þ
j¼2
Using (12.7) the jth abundance αj can be solved by Cramer’s rule as follows:
12.2 Geometric Simplex Growing Algorithm (GSGA) 365
1 1 1 1 1
Det
m1 mj1 r mjþ1 mp
αj ¼
for 1 j p ð12:8Þ
1 1 1 1 1
Det
m1 mj1 mj mjþ1 mp
and
along with α1 given by (12.8) and (12.2), which can then be used to calculate
(12.10) and (12.11).
Furthermore, a comparison of (12.6) and (12.7) clearly shows that the ASC
imposed on (12.6) has been removed in (12.7). This is because the SV calculated by
(12.6) is based on p signatures, m1, m2, . . ., mp which must satisfy the ASC,
while the SV calculated by (12.7) is based on ( p 1) simplex edge vectors,
me 2 ¼ m2 m1 , . . . , m
e p ¼ mp m1 , with dimensions less than p by one since
(12.7) does not need to impose ASC. This further indicates that, unlike (12.6),
which is determined by signatures themselves, GSV calculated by (12.7) according
to their edges has nothing to do with spatial locations of signatures. This crucial
difference allows OP to be used to calculate the GSV.
Most recently, Wang et al. (2013a, b, c) developed an approach, called distance-
based SGA (Dist-SGA), which used a new distance measure to find tj+1. Interest-
ingly, this approach is the same as finding a distance perpendicular to the simplex
base, which can be considered as an OP onto the hyperplane containing previously
grown simplexes. As a result, Dist-SGA can be considered a technique for calcu-
lating GSV by perpendicular distance (GSV-PD) and is essentially the same as
OPSGA developed in Chap. 11, which is a GSV-OP technique; a comparison
366 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
between OPSGA and Dist-SGA can be found in Chap. 11. In addition to Dist-SGA,
Geng et al. (2014) recently also proposed a similar approach for band selection but
not for finding endmembers.
The solution to (12.12) can be found by the well-known GSOP. Two versions of
GSOP can be derived.
1. Initial condition:
Find two data sample vectors, m1 and m2, with the maximal segment length as
two endmembers. Let
e 2 ¼ m2 m1 , u1 ¼ jjm e2 me 2 .
m1 jj, and u2 ¼ 1=2 ¼
m
m 1
e 2T m
ðm e2Þ e 2
m
2. For the given jth target mj for j > 2 find
e j ¼ mj m1
m ð12:13Þ
j1
X
e j; m
m ei
ej
mj ¼ m e
m
i¼2
hm e ii i
e i; m
X j1
X
j1
ej
¼m me j ; ui ui ¼ m
ej me jT ui ui ð12:14Þ
i¼2 i¼2
X
j1
ej
¼m e j;
ui uiT m
i¼2
where
mj mj
uj ¼
1=2 ¼ : ð12:15Þ
mj
mjT mj
e jGSGA þ m1 ;
mjGSGA ¼ m ð12:18Þ
where e
r ¼ r m1 .
4. Check whether j ¼ p. If so, the algorithm is terminated. Otherwise, let p p þ 1,
and go to step 2.
e jGSGA is obtained by finding maximal OPs of all data
According to (12.16), m
sample vectors er ¼ r m1 along the unit vector u j. In this case, er can be
⊥ k ⊥ ⊥ k #
decomposed into er ¼erj þ e r j ¼ PU e
r j , where e r and er j ¼ PUj1 e
r ¼ Uj1 Uj1 r
j1
#
using the orthogonal subspace projector defined by P⊥ Uj1
¼ I Uj1 Uj1 , using
1
# T T
Uj1 ¼ u2 u3 uj1 and the pseudo-inverse of Uj1 , Uj1 ¼ Uj1 Uj1 Uj .
k
From a signal processing point of view, e r⊥j and e
r j can be considered as the high-pass
and low-pass filter parts of the signal source e r . Thus, we can further decompose
k k
lower-pass filter e r⊥
r j1 into its high-pass-filter part, ej2, and low-pass-filter part, e
r j2,
and process continues until it reaches first basis vector u 2 as follows:
k
e ⊥ ¼e
r j1 r⊥
j2 þ e r⊥
r j2 ¼ ¼ e r⊥
j2 þ þ e3 þ u2 : ð12:19Þ
368 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
X
j1 j1
X
he
ri; me ii
rj ¼ e
r ei ¼ e
m rj er jT ui ui ð12:20Þ
i¼2
e i; m
hm e ii i¼2
with e
r j ¼ r m1 . As a result, er can also be expressed as
e
r¼ r1T u1 , r2T u2 , . . . , rjT ujin terms of basis vectors u 3, . . ., u
1, u j. This leads to
another version of GSOP for finding hjGSGA .
1. Initial condition:
Find two data sample vectors, m1 and m2, with the maximal segment length as
two endmembers. Let
e 2 ¼ m2 m1 , u1 ¼ jjm e2 me 2 .
m1 jj, and u2 ¼ 1=2 ¼
m
m 1
e 2T m
ðm e2Þ e 2
m
2. For the given jth target mj for j > 2, find m e j ¼ mj m1 via (12.13) and the
orthonormal vector u ej
j to the space linearly spanned by u2 , u3 , . . . , uj1 , mj ¼ m
Xj1 m e j; m
ei
i¼2 e i by (12.14), and uj ¼ Tmj 1=2 ¼ mj by (12.15).
m
hme i; m
e ii ðmj mj Þ jjmj jj
3. Furthermore, find the jth mutual orthogonal component rj given by (12.20).
4. Using (12.17) to find the maximal height of GSGA, hjGSGA can be found and
computed by
n o n o
mjGSGA ¼ arg maxrj krj k ¼ arg maxrj rjT uj ; ð12:21Þ
Even though GSOP1 and GSOP2, developed in Sect. 12.3.1, produce identical
results, there is a key difference between them. The equation used in GSOP1,
(12.16), is used to find a new endmember and projects all data sample vectors
e
r ¼ r m1 directly onto the jth orthonormal basis vector u j, whereas (12.21), used
in GSOP2, is used to find a new endmember and orthonormalizes all data sample
vectors e
r ¼ r m1 by u2 , u3 , . . . , uj1 via (12.20) to obtain rj prior to projecting the
jth orthonormal basis vector u j. Nonetheless, either one can be used in the imple-
mentation of GSGA, described as follows.
GSGA
1. Initial conditions:
Find two data sample vectors with the maximal segment with two endpoints
specified by m1 and m2, denoted by mGSGA
1 and mGSGA2 respectively.
Define m e2GSGA
¼ m2
GSGA
m1 GSGA
.
2. For each 3 j p we implement GSOP to find hjGSGA by (12.17) via finding
n o
e jGSGA ¼ arg maxer e
mjGSGA in (12.14) through m r T uj in (12.15). Or we can find
n o
mjGSGA ¼ arg maxrj krj k using (12.21).
e jGSGA þ m1 via (12.18) or
3. Find either mjGSGA ¼ m
Apparently, the main steps of GSOP are to find (12.16) and (12.17). In this section
j in (12.15) as
we derive a recursive equation for mj in (12.14) and its normalized u
follows.
X
j
T
e jþ1
mjþ1 ¼ m e jþ1 ¼ m
ui uiT m e jþ1
e jþ1 Uj Uj m
i¼2 ð12:26Þ
T
¼ I Uj Uj me jþ1 :
For the processed data given by u j and a new input data sample vector mj+1
Xj1
e j; m
m ei
e jþ1 ¼ mjþ1 m1 , that is, mj ¼ m
m ej e i and uj ¼ Tmj 1=2 ¼ mj .
m
h e
m i ; e
m i i ðmj mj Þ jjmj jj
i¼2
Now, let Uj ¼ u2 u3 uj . Then mjþ1 can be found by
X
j
e jþ1
mjþ1 ¼ m ui uiT me jþ1
i¼2
ð12:27Þ
T
¼m e jþ1 ¼ I Uj UjT m
e jþ1 Uj Uj m e jþ1 ;
12.3 Recursive Hyperspectral Sample Processing of Geometric Simplex. . . 371
where
T
X
j
Uj Uj ¼ ui uiT ¼ Uj1 þ uj ujT : ð12:28Þ
i¼2
RHSP-GSGA
1. Initial conditions:
Find two data sample vectors with the maximal segment with two endpoints
specified by m1 and m2, denoted by mGSGA
1 and mGSGA
2 respectively.
m GSGA m GSGA
e 2GSGA ¼ m2GSGA m1GSGA , u1GSGA ¼
Define m 1
, and u2GSGA ¼ m2GSGA .
jjm1GSGA jj jj 2 jj
e j ¼ mj m1 and (12.27) to find
2. For each 3 j p use (12.13) to find m
GSGA GSGA T
mjGSGA ¼ I Uj2 Uj2 me jGSGA ð12:29Þ
and
mjGSGA
ujGSGA ¼ ;
ð12:30Þ
mjGSGA
h i
GSGA
where Uj2 ¼ u2GSGA u3GSGA uj2
GSGA
.
n o
3. Use (12.16) to find me jGSGA ¼ arg maxer e r T uj or (12.20) to find mjGSGA ¼ arg
maxr krj k .
4. Calculate hjGSGA ¼ kmjGSGA k.
5. Use (12.12) to find the SV
GSV S m2GSGA ; . . . ; mj1
GSGA
; mjGSGA
ð12:31Þ
¼ hjGSGA GSV S m2GSGA ; . . . ; mj1GSGA
:
and
T
GSGA GSGA T GSGA
Uj1 Uj1 ¼ Uj2 þ uj1
GSGA GSGA
uj1 : ð12:33Þ
It should be noted that the major difference between RHSP-GSGA and GSGA is
the use of recursive equations (12.32)–(12.33) in step 5.
Alternatively, from (12.26) we can further derive Kalman filte-like equations for
the GSGA using OSP as follows in a manner similar to that described in Chap. 3.
For j 2 a new endmember mjþ1 can be found by
X
j
T
e jþ1
mjþ1 ¼ m ui uiT me jþ1 ¼ m
e jþ1 Uj Uj me jþ1
i¼2
T
e jþ1 m
¼m ej þ m e jþ1 uj ujT m
e j Uj1 Uj1 m e jþ1
T
e jþ1 m
¼m ej þ m
e j Uj1 Uj1 me jþ1 m
ej þ m ej ð12:34Þ
e jþ1
uj ujT m
T
¼ ∇me jþ1 þ mj Uj1 Uj1 ∇m e jþ1
e jþ1 uj ujT m
¼ mj þ I Uj1 Uj1
T
∇m e jþ1 ;
e jþ1 uj ujT m
where
∇m e jþ1 m
e jþ1 ¼ m e j ¼ mjþ1 mj ð12:35Þ
is the innovation information obtained from mjþ1 , but not in mj , and with
T
X
j
Uj Uj ¼ ui uiT ¼ Uj1 þ uj ujT ð12:36Þ
i¼2
ju
simply updated by Uj1 and u Tj .
KF OSP-GSGA
1. Initial conditions:
Find two data sample vectors with the maximal segment with two endpoints
specified by m1 and m2, denoted by mGSGA
1 and mGSGA
2 respectively. Define
m GSGA m GSGA
m2GSGA ¼ m2GSGA m1GSGA , u1GSGA ¼ 1
, and u2GSGA ¼ m2GSGA .
jjm1GSGA jj jj 2 jj
2. For each 3 j p use (12.14) to find e
r j ¼ r m1 and (12.17) to find
T
GSGA
mjGSGA ¼ mj1
GSGA
þ I Uj2 GSGA
Uj2 e j1
∇m GSGA
T ð12:37Þ
ujGSGA ujGSGA e j1
m GSGA
and
12.3 Recursive Hyperspectral Sample Processing of Geometric Simplex. . . 373
mjGSGA
ujGSGA ¼ ;
ð12:38Þ
mjGSGA
h i
GSGA
where Uj2 ¼ u2GSGA u3GSGA uj2
GSGA
.
3. Find hjGSGA ¼ kmjGSGA k via (12.17).
4. Use (12.12) to find the SV
GSV S m2GSGA ; . . . ; mj1
GSGA
; mjGSGA
ð12:39Þ
¼ hjGSGA GSV S m2GSGA ; . . . ; mj1GSGA
:
and
T
GSGA GSGA T GSGA
Uj1 Uj1 ¼ Uj2 þ uj1
GSGA GSGA
uj1 : ð12:41Þ
m m
• uj ¼ T j 1=2 ¼ mj ,
ðmj mj Þ
jj j jj
• I Uj1 Uj1
T
:
X
j
T
rjþ1 ¼ e
r ui uiT er ¼e
r Uj Uj er
i¼2
T
¼e ej þ m
rm e j Uj1 Uj1e
r uj ujT e
r ð12:42Þ
T
¼ ∇erj þ mej Uj1 Uj1 ∇e
rj uj ujT e
r
T
¼m r j uj ujT e
e j þ I Uj1 Uj1 ∇e r;
KF OVP-GSGA
1. Initial conditions:
Find two data sample vectors with the maximal segment with two endpoints
specified by m1 and m2, denoted by mGSGA
1 and mGSGA
2 respectively.
m1GSGA m2GSGA
e 2GSGA ¼ m2GSGA m1GSGA , u1GSGA ¼
Define m , and u2GSGA ¼ .
jjm1GSGA jj jjm2GSGA jj
2. For each 3 j p find e
r j ¼ r m1 and use (12.43) and (12.44) to find
T
GSGA
mjGSGA ¼ GSGA
mj1 þ I Uj2 GSGA
Uj2 e j1
∇m GSGA
T ð12:46Þ
ujGSGA ujGSGA e j1
m GSGA
and
mjGSGA
ujGSGA ¼ ;
ð12:47Þ
mjGSGA
h i
GSGA
where Uj2 ¼ u2GSGA u3GSGA uj2
GSGA
.
12.3 Recursive Hyperspectral Sample Processing of Geometric Simplex. . . 375
and
T
GSGA GSGA T GSGA
Uj1 Uj1 ¼ Uj2 þ uj1
GSGA GSGA
uj1 : ð12:50Þ
2. New information:
• mj,
• rj :
3. Innovation information:
e j;
• ∇m
• ∇rj ¼ e e j.
rm
Table 12.1 provides a comparison of how information is used for updating the
GSGA using the KF OSP-GSGA and KF OVP-GSGA.
376 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
T Xj1
T
Uj1 Uj1 ¼ i¼2
ui uiT GSGA
I Uj2 GSGA
Uj2
T
mj ¼ I Uj1 Uj1 m ej
mj
uj ¼
mj
I Uj1 Uj1 T
j as
e jþ1 and U
Recalling (12.14), we can rewrite mjþ1 in terms of m
T T
mjþ1 ¼ m e jþ1 ¼ I Uj1 Uj1 m
e jþ1 Uj Uj m e jþ1 ; ð12:51Þ
T
where Uj ¼ u2 u3 uj , and Uj Uj ¼ Uj1 þ uj ujT is given by (12.28).
j 1; i ¼ j T
Since fui gi¼1 is an orthonormal vector, that is, ui uj ¼ , U U ¼ I, this
0; i 6¼ j j j
1
# T T T
implies that P⊥ U
¼ I U j Uj ¼ I Uj U j U j Uj ¼ I Uj Uj , which is exactly
j
mj
mjþ1 ¼ P⊥
U
e jþ1
m and ujþ1 ¼ : ð12:52Þ
j mj
e jþ1
However, (12.53) is the same as finding the ( j + 1)th endmember m GSGA
obtained
by GSGA that is given by (12.16), that is,
n o
e jþ1
m GSGA
¼ arg maxer e
r T ujþ1 ¼ arg maxr rT ujþ1 : ð12:54Þ
¼ e jþ1
m T
P⊥ e :
m
Uj jþ1
ð12:55Þ
e jþ1
ηjþ1 ¼ m T
P⊥
U
e jþ1 ;
m ð12:56Þ
j
which can be used to estimate the value of VD, nRHBP-GSGA, to determine how many
endmembers for RHSP-GSGA should be generated by letting m e RHSP
e jþ1 ¼ m -GSGA
jþ1
obtained by (12.54) or hjGSGA
¼ kmj GSGA
k derived from (12.29). In other words, the
ηj+1 obtained by (12.56) can be used as the ( j + 1)th signal source under the binary
hypothesis testing problem to determine the VD. Similarly, (12.56) can also be
modified by the residual strength of the signal source, denoted by ρj+1, similar to
(11.42) in Chap. 11, and given by
pffiffiffiffiffiffiffiffi T ⊥
1=2
ρjþ1 ¼ ηjþ1 ¼ m e jþ1
e jþ1 PU m ; ð12:57Þ
j
e jþ1 ¼ m RHSP-GSGA
e jþ1
with m in (12.54).
Let N denote the total number of pixels in a hyperspectral image, L the number of
bands, j the number of endmembers to be found during the process, and p the total
number of endmembers needed to be generated. Since the initial steps of these
methods are the same—either to find a data sample with maximum vector length
and then find another data sample with maximum distance from the first one or to
find two samples yielding the maximum distance—the complexity ignores the
complexity of the initial step to simplify the discussion.
12.6 Computational Complexity 379
The complexity
for finding the
second
endmember
is the same with each iteration of
OPSGA, O j3 þ jL2 þ NL2 O L2 þ NL2 . For j > 2, the projection operator
P⊥ ⊥ 2
e can be updated from Pe recursively with complexity O(NL ). Thus, the
Uj U j1
overall complexity of recursive OPSGA (ROPSGA) is O(pNL2).
To reduce the complexity of projecting all data sample vectors onto orthogonal
subspaces, the complexity required by GSOP is O( jL) for each data sample vector.
Thus, a complexity of O( jNL) is needed for the total number of N data sample
vectors.
X p This
implies that the complexity of GSGA using GSOP is
O j¼2
ðjNLÞ Oðp2 NLÞ.
The complexity for finding the second endmember is O(NL). For j > 2, the pro-
jections specified by (12.26)–(12.28) require complexity O(NL). This leads to a
complexity of RHSP-GSGA of O(pNL). A comparison among the complexities of
all the SGA-based methods is summarized in Table 12.2.
Two real hyperspectral image scenes in Figs. 1.10 and 1.14, which have been
studied extensively in the literature, are used for experiments.
12.7 Real Image Experiments 381
a b
p11, p12, p13
Fig. 12.2 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels
The image data to be studied are from the HYperspectral Digital Imagery Collec-
tion Experiment (HYDICE) image scene shown in Fig. 12.2a (also shown in
Fig. 1.10a), which has a size of 64 64 pixel vectors with 15 panels in the scene
and the ground truth map in Fig. 12.2b (Fig. 1.10b).
As noted, panel pixel p212, marked by yellow in Fig. 12.2b, is of particular
interest. Based on the ground truth, this panel pixel is not a pure panel pixel and is
marked by yellow as a boundary panel pixel. However, with our extensive and
comprehensive experiments this yellow panel pixel is always extracted as the one
with the most spectrally distinct signature compared to those R panel pixels in row
2. This indicates that a signature of spectral purity is not equivalent to a signature of
spectral distinction. As a matter of fact, in many cases panel pixel p212 instead of
panel pixel p221 is the one extracted by EFAs to represent the panel signature in row
2. Also, because of such ambiguity, the panel signature representing the panel
pixels in the second row is either p221 or p212, which is always difficult to find
using EFAs. This implies that the ground truth of R panel pixels in the second row
provided in Fig. 12.2b may not be as pure as was thought.
382 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
By virtue of (12.56) and (12.57), the energies [ηj+1 in (12.56)] and strengths [ρj+1
in (12.57)] of real target pixels found by DSGA and GSGA can be used to
determine the number of endmembers by VD with RHSP-OPSGA replaced by
RHSP-GSGA. Table 12.3 tabulates their values estimated by the maximal orthog-
onal complement algorithm (MOCA), which is equivalent to a Bayes detector as
well as an NPD with various false alarm probabilities.
Since DSGA found targets without performing data DR, it seemed to produce
larger values for nRHBP-GSGA than GSGA, which grows simplexes up to nRHBP-GSGA
vertices. In addition, 45 is the least upper bound on nRHBP-GSGA determined by
GSGA (Table 12.3), DSGA and GSGA were applied to find up to 45 target pixels as
potential endmembers as shown in Fig. 12.3 for HYDICE data. As we can see from
Fig. 12.3a, b, the first nine target pixels found in Fig. 12.3a already included three
panel pixels, p11, p312, p521, and the next 9 target pixels from 10 to 18 pixels in
Fig. 12.3b also picked up the remaining 2 panel pixels, p212 and p412. As shown in
Fig. 12.2b, p212 is more pure than its neighboring pixels p211 and p221. Although
both DSGA and GSGA found five panel pixels corresponding to the three panel
pixels p11, p312, and p521 in Fig. 12.3a and the remaining two panel pixels p212 and
p412 in Fig. 12.3b, their orders of appearance were different. Interestingly, both
required 18 target pixels to find each panel pixel in 5 different rows and the last
panel pixels were the 18th pixels, p412 by DSGA and p212 by GSGA.
Interestingly, according to Chang (2003a, b, 2013) and Chang and Du (2004),
the number of endmembers estimated for this scene by VD was nVD ¼ 9. Also, from
Chang et al. (2010a, 2011b), the number of signatures used for LSMA was
estimated as twice that of nVD, 2nVD ¼ 112. In this case, the first 9 target pixels in
Fig. 12.3a and 10–18 target pixels shown in Fig. 12.3b found by DSGA and GSGA
respectively can be used as desired endmembers, which include all 5 mineral
signatures.
Furthermore, the target pixels found in Fig. 12.3a, b were used to form 9 vertex
simplexes and 18-vertex simplexes from which we can calculate their SVs.
Table 12.4 tabulates their SVs, where DSGA-generated simplexes always yielded
larger volumes than GSGA-generated simplexes.
12.7 Real Image Experiments 383
Table 12.4 Simplex volumes calculated based on target pixels found by DSGA and GSGA
DGSA GSGA
nVD ¼ 9 1.4319E+24 1.0617E+24
2nVD ¼ 18 2.0509E+39 1.5677E+39
a b
Fig. 12.4 (a) Cuprite AVIRIS image scene. (b) Spatial positions of five pure pixels corresponding
to minerals alunite (A), buddingtonite (B), calcite (C), kaolinite (K), and muscovite (M). (c) Five
mineral reflectance spectra. (d) Five mineral radiance spectra
The image scene used was a real Cuprite image data (Fig. 12.4, also shown in
Fig. 1.4), which is available at the USGS Web site http://aviris.jpl.nasa.gov/. This
scene is a 224-band image with a size of 350 350 pixels and was collected over
the Cuprite Mining District site in Nevada in 1997. It is one of the most widely used
hyperspectral image scenes available in the public domain and has 20 m spatial
resolution and 10 nm spectral resolution in a range 0.4–2.5 μm. This scene has been
studied extensively because it is well understood mineralogically and has reliable
ground truth. Two data sets for this scene, reflectance and radiance data, are also
available for study. There are five pure pixels in Fig. 12.4a, b, which can be
identified as corresponding to five different minerals, alunite (A), buddingtonite
(B), calcite (C), kaolinite (K), and muscovite (M) labeled A, B, C, K, and M in
Fig. 12.4b, along with their spectral signatures plotted in Fig. 12.4c, d.
Once again, the energies [ηj+1 in (12.56)] and strengths [ρj+1 in (12.57)] of real
target pixels found by DSGA and GSGA are used to determine the number of
endmembers tabulated in Table 12.5.
12.7 Real Image Experiments 385
c
7000
Muscovite
6000
Alunite Kaolinite
5000
Reflectance
4000 Buddingtonite
3000
Calcite
2000
Alunite
Buddingtonite
1000 Calcite
Kaolinite
Muscovite
0
400 600 800 1000 1200 1400 1600 1800 2000 2200 2400
Wavelength(nm)
d
12000
Muscovite Kaolinite Alunite
Buddingtonite
10000 Calcite
Muscovite
Kaolinite
8000 Alunite
Radiance
6000
4000
Calcite
Buddingtonite
2000
0
400 600 800 1000 1200 1400 1600 1800 2000 2200 2400
Wavelength(nm)
Fig. 12.4 (continued)
Table 12.6 Target pixels found by EIDA for Cuprite reflectance data
Cuprite A B C K M
DSGA Reflectance tDSGA
11 tDSGA
13 tDSGA
44 tDSGA
64 tDSGA
8
GSGA Reflectance tGSGA
29 tGSGA
53 tGSGA
40 tGSGA
27 tGSGA
9
a b
a
A a A
b
b
B B
m M m M
K K
k k
c c
C C
Fig. 12.6 Target pixels found by EIDA compared to ground truth pixels for Cuprite reflectance
data. (a) DSGA. (b) GSGA
a b
6000 6000
5500 5500
5000 5000
4500 4500
4000
Reflectance
4000
Reflectance
3500 3500
3000 3000
2500 2500
c4000 d5500
5000
3500
4500
3000 4000
3500
Reflectance
Reflectance
2500
3000
2000
2500
1500 2000
e 7000
6000
5000
Reflectance
4000
3000
Groundtruth
2000
DSGA
GSGA
1000
0 20 40 60 80 100 120 140 160 180 200
Band
Fig. 12.7 Comparative plots of spectral signatures found by DSGA and GSGA on Cuprite
reflectance data. (a) “a” signatures against “A.” (b) “b” signatures against “B.” (c) “c” signatures
against “C.” (d) “k” signatures against “K.” (e) “m” signatures against “M”
Similarly, the same experiments as were conducted for Cuprite reflectance data
were also performed for Cuprite radiance data. Table 12.8 tabulates nRHBP-GSGA
determined by the energies [ηj+1 in (12.56)] and strengths [ρj+1 in (12.57)] of real
target pixels found by DSGA and GSGA.
12.7 Real Image Experiments 389
Table 12.7 SAM/SID of closest endmembers with ground truth by DSGA and GSGA on Cuprite
reflectance data
Once again, a value of 75 was used for DSGA and GSGA to generate 75 target
pixels. Figure 12.8a–c shows the first 22 target pixels (Fig. 12.6a), followed by
pixels 23–46 (Fig. 12.8b) and pixels 47–75 (Fig. 12.8c).
Figure 12.9 shows the five target pixels identified by EIDA from Fig. 12.8
against the five ground truth pixels, and Table 12.9 which tabulates these five target
pixels found by DGSA and GSGA with their orders of appearance.
From Table 12.9 we see that the most difficult mineral signature for DSGA and
GSGA to find are different, “K,” which requires 103 target pixels, for the DGSA
and “C,” which requires 68 target pixels, for GSGA.
In analogy with Figs. 12.9 and 12.10a–e, Fig. 12.4 shows profiles of the spectral
signatures of each of the five identified target pixels against their corresponding
ground truth pixels.
Once again, SAM and SID are used to calculate the spectral similarity values
among the identified target pixels in Fig. 12.9 against the ground truth pixels with
their results tabulated in Table 12.10, where both DGSA and GSGA found two
different sets of pixels with two common pixels that corresponded to two ground
truth pixels, “B” and “M.” Nonetheless, the spectral signatures of all the found
pixels are also very close.
The experiments conducted for Cuprite reflectance and radiance data sets dem-
onstrated an important fact: there are no unique target pixels corresponding to
ground truth pixels in the scene. In other words, many more pixels scattering around
in the scene can also correspond to ground truth pixels.
Finally, the target pixels in Figs. 12.5 and 12.8 were used to form 22/44-vertex
simplexes for reflectance data and 15/30-vertex simplexes for radiance data, from
which we can calculate their SVs. Table 12.11 tabulates their respective SVs.
Interestingly, unlike HYDICE data, GSGA produced larger SVs than DSGA did
for both the reflectance and radiance data sets.
Fig. 12.8 Target pixels
found by DSGA and GSGA
for Cuprite radiance data.
(a) 22 endmember pixels.
(b) 23–46 endmember
pixels. (c) 47–75
endmember pixels
Fig. 12.9 Endmember pixels found by EIDA compared to ground truth pixels for Cuprite
radiance data. (a) DSGA. (b) GSGA
Table 12.9 Target pixels found by EIDA for Cuprite radiance data
Algorithm Cuprite A B C K M
DSGA Radiance tDSGA
12 tDSGA
14 tDSGA
28 tDSGA
103 tDSGA
6
GSGA Radiance tGSGA
16 tGSGA
14 tGSGA
68 tGSGA
11 tGSGA
7
a b
12000
Groundtruth 9000 Groundtruth
DSGA DSGA
8000
10000 GSGA GSGA
7000
8000
6000
Radiance
Radiance
5000
6000
4000
4000 3000
2000
2000
1000
0 0
0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200
Band Band
c d
8000 10000
Groundtruth Groundtruth
7000 DSGA 9000 DSGA
GSGA GSGA
8000
6000
7000
5000
6000
Radiance
Radiance
4000 5000
4000
3000
3000
2000
2000
1000
1000
0 0
0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200
Band Band
e
12000
Groundtruth
DSGA
10000 GSGA
8000
Radiance
6000
4000
2000
0
0 20 40 60 80 100 120 140 160 180 200
Band
Fig. 12.10 Comparative plots of spectral signatures found by DSGA and GSGA on Cuprite
radiance data. (a) “a” signatures against “A.” (b) “b” signatures against “B.” (c) “c” signatures
against “C.” (d) “k” signatures against “K.” (e) “m” signatures against “M”
392 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
Table 12.10 SAM/SID of closest endmembers with ground truth by DSGA and GSGA on Cuprite
radiance data
a b
1000 4
DSGA Dist-SGA Dist-SGA
900 Dist-SGA OPSGA
OPSGA 3.5
RHSP-OPSGA
Accumulative computing time (sec)
Accumulative computing time (sec)
500 2 KF OVP-GSGA
400 GSGA
1.5 RHSP-OPSGA
300
1
200
0.5 KF OSP-GSGA
100 RHSP-GSGA
0 0
0 20 40 60 80 100 120 140 160 180 0 20 40 60 80 100 120 140 160 180
Number of endmembers (p) Number of endmembers (p)
Fig. 12.11 Cumulative computing time in seconds of DSGA without DR, Dist-SGA, OPSGA,
RHSP-OPSGA, GSGA, RHSP-GSGA, KF OSP-GSGA, and KF OVP-GSGA as p increases on
HYDICE data
This section summarizes the computer processing times required by DGSA and seven
SGA-based algorithms, four developed in this chapter, GSGA, RHSP-GSGA, KF
OSP-GSGA, and KF OVP-GSGA, two developed in Chap. 11, OPSGA and RHSP-
OPSGA, plus the Dist-SGA, developed by Wang et al. (2013a, b, c). Figure 12.11a
conducts an overall comparative analysis on cumulative computing time in seconds
among all eight algorithms, DSGA without DR, Dist-SGA, OPSGA, RHSP-OPSGA,
GSGA, RHSP-GSGA, KF OSP-GSGA, and KF OVP-GSGA applied to HYDICE
data as the number of target pixels, p, increases. Since DSGA requires a significant
amount of time, which overwhelmed the other seven SGA-modified algorithms, it is
impossible to see their relative performances. In this case, Fig. 12.11b only plots the
12.7 Real Image Experiments 393
Table 12.12 Comparison of computing time in seconds by various methods with DSGA on
HYDICE data
Methods
Dist- RHSP- RHSP- KF KF
p DSGA SGA OPSGA OPSGA GSGA GSGA OSP-GSGA OVP-GSGA
9 1.7086 0.1134 0.1136 0.1026 0.0697 0.0640 0.0673 0.1519
18 5.2592 0.2542 0.2292 0.2114 0.1521 0.1359 0.1457 0.3433
34 20.180 0.5202 0.4177 0.4014 0.3087 0.2639 0.2842 0.6843
45 43.770 0.7192 0.5482 0.5344 0.4236 0.3532 0.3763 0.9230
a b
12000 140
DSGA Dist-SGA
Dist-SGA OPSGA Dist-SGA
120
Accumulative computing time (sec)
0 0
0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200
Number of endmembers (p) Number of endmembers (p)
Fig. 12.12 Cumulative computing time in seconds of DSGA without DR, Dist-SGA, OPSGA,
RHSP-OPSGA, GSGA, RHSP-GSGA, KF OSP-GSGA, and KF OVP-GSGA as p increases on
Cuprite data
computing times required by these seven algorithms for better visual assessment.
According to Fig. 12.11b, clearly the best algorithm is RHSP-GSGA and the worst
are Dist-SGA in terms of computing time when the number of target pixels, nT, is
greater than 140 and KF OVP-GSGA when nT is less than 140. In addition, Fig. 12.11
also partitions all eight algorithms into four groups; the best group is composed of
RHSP-GSGA and KF OSP-GSGA, followed by the second best group made up of
GSGA, OPSGA, and RHSP-OPSGA, the third group consisting of Dist-SGA and KF
OVP-GSGA, and, finally, the worst group, DGSA, which is the original SGA
developed by Chang et al. (2006a, b, c) without DR.
To give a better idea about computing times, Table 12.12 tabulates the comput-
ing times required by all eight algorithms for p ¼ 9, 18, 34, and 45 as examples
where all seven algorithms except DSGA required less than 1 s to accomplish the
task.
Similarly, for an overall comparative analysis, Fig. 12.12 also plots the computer
processing times required by the eight algorithms, DSGA without DR, Dist-SGA,
OPSGA, RHSP-OPSGA, GSGA, RHSP-GSGA, KF OSP-GSGA, and KF
OVP-GSGA applied to Cuprite data as the number of target pixels, p, increases.
Conclusions similar to those drawn from the HYDICE data are also applied to the
394 12 Recursive Hyperspectral Sample Processing of Geometric Simplex Growing. . .
Table 12.13 Comparison of computing time in seconds by various methods with DSGA
Methods
Dist- RHSP- RHSP- KF OSP- KF OVP-
p DSGA SGA OPSGA OPSGA GSGA GSGA GSGA GSGA
22 262.31 10.150 7.6834 7.6477 6.0605 5.3360 5.5380 14.832
46 1465.5 12.971 16.3953 16.3544 13.510 12.3633 12.897 32.425
75 4431 39.894 26.9343 26.8683 12.343 112.6557 19.637 53.631
Cuprite data, except that KF OVP-GSGA was the worst and Dist-SGA was the
second worst algorithm (Fig. 12.12).
Like Table 12.10, Table 12.13 also tabulates the computing times required by all
eight algorithms for p ¼ 22, 46, and 75 as examples where all seven algorithms
required significantly less time than DGSA.
According to the computer processing times shown by Figs. 12.11 and 12.12,
when nT is less than approximately 140, the performance of DSGA, Dist-SGA,
OPSGA, RHSP-OPSGA, GSGA, RHSP-GSGA, KF OSP-GSGA, and KF
OVP-GSGA can be ranked as
RHSP-GSGA
KF OSP-GSGA
GSGA
KF OVP-GSGA
ð12:59Þ
RHSP-OPSGA
OPSGA
Dist-SGA
DSGA
RHSP-GSGA
KF OSP-GSGA
GSGA
RHSP-OPSGA OPSGA ð12:60Þ
Dist-SGA
KF OVP-GSGA
DSGA
e jþ1 but
specified by (12.35) to calculate the innovation information provided by m
e j , which is not required by RHSP-GSGA. Nonetheless, it is this equation
not in m
(12.35) that makes easy hardware design on a chip a real-time process. It is our
belief that when both RHSP-GSGA and KF OSP-GSGA are implemented on a
chip, KF OSP-GSGA will outperform RHSP-GSGA.
12.8 Conclusions
SVA
GSGA calculates GSV through a base formed by simplex edge vectors and the
height. Thus, it must start with a two-vertex simplex as its initial condition because
a degenerated simplex with a simplex edge vector is a simplex connected by two
endmembers. As a consequence, GSGA-produced simplexes have nothing to do
with the spatial locations of simplexes it finds, but OPSGA does. A second
difference is that OPSGA suppresses data sample vectors via OSPs, while GSGA
does not because GSGA orthogonalizes only height vectors via GSOP. Lastly but
most importantly is finding endmembers. Assume that at the ( j + 1)th stage we
j
already found j endmembers, fmk gk¼1 , and need to find the next endmember, the ( j
+1)th endmember, m j+1 , to form a new ( j + 1)-vertex simplex,
S m1 ; m2 ; . . . ; mjþ1 . There are three approaches to finding such an mj+1. The
first one is DSGA, which finds an mj+1 that yields the maximal DSV by calculating
the matrix determinant n specified o by (12.1). A second one is OPSGA, which finds mj
⊥
+1 via mjþ1 ¼ arg maxr PUj r specified by (11.1), where Uj ¼ m1 m2 mj . A
j
third approach is GSGA, which finds the maximal height via (12.12), where the
j
simplex formed by fmk gk¼1 , S(m1, m2, . . ., mj) is considered a base of
S m1 ; m2 ; . . . ; mjþ1 and height of mj+1. Because they avoid calculating DSV by
matrix determinants, OPSGA and GSGA turn out to be the best SGA-based
algorithms reported in the literature and outperform Wang et al.’s Dist-SGA in
the sense of having the least computational complexity and lowest computer
processing time and produces true SVs.
The diagram of Fig. 12.13 summarizes all the SVA-based approaches, MVT
and N-FINDR, DSGA, ATGP, GSGA and RHSP-GSGA, and OPSGA and RHSP-
OPSGA in Chap. 11, along with the Dist-SGA developed by Wang et al. (2013a, b, c).
Part IV
Sample Spectral Statistics-Based
Recursive Hyperspectral Band Processing
By far, all the major hyperspectral imaging algorithms described in this book have
been redesigned for implementation as real-time processing algorithms in Part II.
Specifically, Part II studied the band-interleaved-pixel/sample/line (BIP/BIS/BIL)
data acquisition format for hyperspectral target detection that uses a causal sample
correlation/covariance matrix (CSCRM/CSCVM) to perform real-time detection.
At this point we have not taken up the issue of designing and developing algorithms
according to the other data acquisition format, the band-sequential (BSQ) format,
which collects data band image by band image, shown in Fig. 1.2 and reproduced in
Fig. 4.1, where (x, y) indicates the spatial coordinate of a data sample vector or pixel
vector and λ is a parameter used to specify spectral bands.
Fig. 4.2 8-bit grayscale image encoded and decoded by bit plane image coding. (a) Progressive
8-bit grayscale processing of bit plane coding from least significant bit (bit plane 1) to most
significant bit (bit plane 8) (Gonzales and Woods 2008). (b) Progressive 8-bit grayscale recon-
struction processing from the two, three, and four most significant bit planes (bit planes 8, 7, 6, 5)
to 8-bit image (Gonzales and Woods 2007)
Chapter 13
Recursive Hyperspectral Band Processing for
Active Target Detection: Constrained Energy
Minimization
13.1 Introduction
l T
We first define xðlÞ ¼ r 1ll ; r 2ll ; . . . ; r Nl as an N-dimensional data sample vector
N
made up of data samples of fri gi¼1 in the lth band image. We also let Xl ¼
N
r1 ðlÞ r2 ðlÞ rN1 ðlÞrN ðlÞ be the data matrix formed by fri ðlÞgi¼1 given by
2 3
l
r 11 r ðlN1Þ1 l
r N1
6 7
6 7
6 ⋮ ⋱ ⋮ ⋮ 7
6 7
Xl ¼ ½r1 ðlÞr2 ðlÞ rN1 ðlÞrN ðlÞ ¼ 6 7; ð13:1Þ
6rl r ðlN1Þðl1Þ r ðN1Þ1 7
l
6 1ðl1Þ 7
4 5
r 1ll r ðlN1Þl l
r Nl
T
where ri ðlÞ ¼ r i1l ; r i2l ; . . . ; r ilðl1Þ ; r ill . In this case, Xl can be reexpressed as
X
l T
Xl ¼ Tl1 , where xðlÞ ¼ r 1ll ; r 2ll ; . . . ; r Nl is an N-dimensional data sample
x ðl Þ
N
vector in the last row in (13.1) consisting of all data sample components of fri gi¼1
in the lth band image. Then we define causal band correlation matrix (CBRCM),
denoted by Rll , as
" #
Xl1
Rll ¼ ð1=N ÞXl XlT ¼ ð1=N Þ T
Xl1 xðlÞ
xT ð l Þ
" # ð13:2Þ
T
Xl1 Xl1 Xl1 xðlÞ
¼ ð1=N Þ :
xT ðlÞXl1
T
xT ðlÞxðlÞ
Using a matrix identity in Settle (1996) and Chang (2003), we derive band-by-band
update equations for calculating the inverse of CBRCM, Rll , as follows:
" #1
T 1 UT U UT d
M M ¼
dT U dT d
2 1 T 3 ð13:3Þ
UT U þ βU# ddT U# βU# d
¼4 T
5;
βdT U# β
" #1
T
Xl1 Xl1 Xl1 xðlÞ
xT ðlÞXl1
T
xT ðlÞxðlÞ
2 1 1 h 1 iT 1 3
X XT þ β Xl1 Xl1
T
Xl1 xðlÞ Xl1 Xl1
T
Xl1 xðlÞ β Xl1 Xl1
T
Xl1 xðlÞ
6 l1 l1 7
¼4 h iT 5:
1
βxT ðlÞ Xl1 Xl1T
Xl1 β
ð13:4Þ
2 h iT 3
Rð1 1
l1Þðl1Þ Xl1 xðlÞ Rðl1Þðl1Þ Xl1 xðlÞ R1
ðl1Þðl1Þ Xl1 xðlÞ
6 R1 7
6 ðl1Þðl1Þ þ ð1=N Þ 7
6 xT ðlÞP⊥ T x ðl Þ xT ðlÞP⊥ T xðlÞ 1 7
6 Xl1 Xl1 7
¼66 ⊥
N x ðlÞPX T xðlÞ
T 7:
h iT 7
6 1
Rðl1Þðl1Þ Xl1 xðlÞ
l1
7
6 7
4 5
⊥
xT ðlÞPX T xðlÞ
l1
ð13:5Þ
R1
ðl1Þðl1Þ :
2. Innovation information:
1
Xl1 xðlÞ, xT ðlÞP⊥
XT
xð l Þ :
l1
3. Input information:
l T
xðlÞ ¼ r 1ll ; r 2ll ; . . . ; r Nl :
ð13:6Þ
1
where κl ¼ dT ðlÞR1
ll dðlÞ is a scalar, ν ¼ R1
ðl1Þðl1Þ Xl1 xðlÞ, and βlðl1Þ
l ðl1Þ
n h i o1
¼ xT ð l Þ P⊥
XT
xðlÞ ;
l1
ð13:7Þ
δRBPCEM ðrðlÞÞ
¼ ðκ l =κ l1 ÞδRBPCEM
ð r ð l 1Þ Þ
þ ð1=N Þκ l β d ðl 1Þν
T Ndl νT
rðl 1Þ Nr l ;
l ðl1Þ l ðl1Þ l ðl1Þ
ð13:8Þ
T
where r(l ) is an l-dimensional data sample vector given by rðlÞ ¼ ðr n
1 ; r 2 ; . .h. ; r l Þ i,
T
, ν
1
κl ¼ d ðlÞR1ll dðlÞ ¼R1
ðl1Þðl1Þ Xl1 xðlÞ, and β ¼ xT ðlÞ P⊥ XT
l ðl1Þ l ðl1Þ l1
1
xðlÞg .
13.4 Real Image Experiments 405
4. Innovation information:
n h i o1
β ¼ xT ðlÞ P⊥ T x ð l Þ ,
l ðl1Þ Xl1
ν ¼ R1
ðl1Þðl1Þ Xl1 xðlÞ:
l ðl1Þ
Two real hyperspectral image scenes were used for experiments to conduct a
performance evaluation of CEM.
The image data to be studied are from the HYperspectral Digital Imagery
Collection Experiment (HYDICE) image scene shown in Fig. 13.1a (also shown
in Fig. 1.10a), which has a size of 64 64 pixel vectors with 15 panels in the scene
406 13 RHBP for Active Target Detection: Constrained Energy Minimization
a b
c
7000
P1
6000 P2
P3
P4
5000 P5
Radiance
4000
3000
2000
1000
0
0 20 40 60 80 100 120 140 160 180
Band
Fig. 13.1 (a) HYDICE panel scene containing 15 panels. (b) Ground truth map of spatial
locations of the 15 panels. (c) Spectra of p1, p2, p3, p4, and p5
and the ground truth map in Fig. 13.1b (Fig. 1.10b). Figure 13.1c plots the five panel
spectral signatures obtained from Fig. 13.1b, where the ith panel signature, denoted
by pi, was generated by averaging the red panel center pixels in row i, as shown in
Fig. 13.1c (also shown in Fig. 1.11). These panel signatures will be used to
represent target knowledge of the panels in each row.
The image was acquired by 210 spectral bands with a spectral coverage from 0.4
to 2.5 μm. Low signal/high noise bands, bands 1–3 and 202–210, and water vapor
absorption bands, bands 101–112 and 137–153, were removed. Thus, a total of
169 bands were used in the experiments. The spatial resolution and spectral
resolution of this image scene are 1.56 m and 10 nm, respectively.
The panel signatures in Fig. 13.1c were used as required prior target knowledge
for the desired target signatures by CEM. Figure 13.2a–e shows that RHBP-CEM
13.4 Real Image Experiments 407
detected amounts of the 19 R panel pixels, p11, p12, p13, p211, p221, p22, p23, p311,
p312, p32, p33, p411, p412, p42, p43, p511, p521, p52, p53, with the desired spectral
signature d specified by one of the 5 panel signatures, p1, p2, p3, p4, p5, in Fig. 13.1c
respectively as the band number is progressively increased from 1 to 169. The
y-axis is the detected amount and the x-axis is the number of bands used to perform
RHBP-CEM. The displayed detected amounts correspond to the δRBPCEM ðrðlÞÞ for
the specified panel pixels. Each value was normalized so that the minimum value is
zero and the maximum value is one. Thus, the pixel with the strongest response to
the algorithm will have a value of one. Such δRBPCEM ðrðlÞÞ-generated abundance
fractional amounts are referred to as normalized abundance values.
According to Fig. 13.2, it is clear that over 75 % of all 19 R panel pixels located
in the first two columns required less than 50 bands to reach similar abundance
fractions with few changes detected by CEM using additional new bands beyond
50. Of particular interest are panel pixels p411, p412 and p511, p521 where p411 and
p412 needed more than 50 bands to reach values consistent with CEM, while p511
and p521 required less than half of 50 bands. To further demonstrate the differential
detected amounts by two consecutive band numbers, Fig. 13.3a–e shows the
difference in the detected abundance fractions of the 19 R panel pixels with the
desired signature d specified by one of the five panel signatures p1, p2, p3, p4, p5 as
the band number l is increased from 1 to 169. This can be written mathematically as
the difference between the detected abundance fractions by δPBPCEM ðrðlÞÞ and
δPBPCEM ðrðl 1ÞÞ, that is, δRHBP-CEM(r(l )) δRHBP-CEM(r(l 1)), where the
detected amounts were the newly added detected abundance fractions provided
by the lth band, Bl, minus the detected amounts from the previous (l 1)
l1
bands. This updates the results from the first (l 1) bands Bj j¼1 to the first
l
l bands, Bj j¼1 .
The plots in Fig. 13.3 provide a better picture of dynamic changes in the detected
amounts of the 19 R panel pixels as the band number is progressively processed. In
particular, these plots can also be used to observe fluctuation and saturation in the
detected amounts, which showed that there was no need to use the complete set of
bands to perform CEM. As a matter of fact, this could easily have been done with
less than one-third of the total number of bands. Although panel pixel p212 is also of
interest, its experiment is not included because it is not an R center panel pixel, as
shown in Fig. 13.1b. To have an even better view of RHBP-CEM detected abun-
dance fractional maps in Figs. 13.2 and 13.3, we show their three-dimensional
(3D) progressive plots of band-varying RHBP-CEM detection maps in Figs. 13.4,
13.5, 13.6, 13.7, and 13.8, where the x-, y-, and z-axes are specified by the 19 R
panels plus a yellow panel pixel, p212, which shows the number of bands being used
to process CEM and the detected abundance fractions.
Since we have the ground truth of the 19 R panel pixels, we can further calculate
their detection rates via a receiver operating characteristic (ROC) analysis for a
performance evaluation (Poor 1994), that is, the detection of panel pixels p11, p12,
p13 using p1, the detection of panel pixels p211, p221, p22, p23 using panel signature
408 13 RHBP for Active Target Detection: Constrained Energy Minimization
1 1 1
0.8 0.8 0.8
0 0 0
0 50 100 150 0 50 100 150 0 50 100 150
Bands Processed Bands Processed Bands Processed
panel pixel p11 panel pixel p12 panel pixel p13
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p211 panel pixel p221
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p22 panel pixel p23
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p311 panel pixel p321
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p32 panel pixel p33
Fig. 13.2 Normalized values of 19 R panel pixels in Fig. 13.1b specified by the five panel
signatures p1, p2, p3, p4, p5 by RHBP-CEM
13.4 Real Image Experiments 409
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p411 panel pixel p412
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p42 panel pixel p43
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p511 panel pixel p512
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 50 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p52 panel pixel p53
p2, the detection of panel pixels p311, p312, p32, p33 using panel signature p3, the
detection of panel pixels p411, p412, p42, p43 using panel signature p4, and the
detection of panel pixels p511, p521, p52, p53 using panel signature p5. To include
the band number as a parameter to see the progressive detection performance of
RHBP-CEM on these 19 R panel pixels band by band, we do not follow the
traditional approach to plotting an ROC curve of detection probability, PD, versus
0.4 0.4 0.4
0 0 0
0.2 0.2
0 0
-0.2 -0.2
-0.4 -0.4
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p211 panel pixel p212
0.4 0.4
0.2 0.2
0 0
-0.2 -0.2
-0.4 -0.4
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p22 panel pixel p23
0.4 0.4
0.2 0.2
0 0
-0.2 -0.2
-0.4 -0.4
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p311 panel pixel p312
0.4 0.4
0.2 0.2
0 0
-0.2 -0.2
-0.4 -0.4
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p32 panel pixel p33
Fig. 13.3 Difference in normalized values of the 19 R panel pixels in Fig. 13.1b specified by the
five panel signatures p1, p2, p3, p4, p5 between δRHBP-CEM(r(l )) and δRHBP-CEM(r(l 1))
13.4 Real Image Experiments 411
0.4 0.4
0.2 0.2
0 0
-0.2 -0.2
-0.4 -0.4
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p411 panel pixel p412
0.4 0.4
0.2 0.2
0 0
-0.2 -0.2
-0.4 -0.4
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p42 panel pixel p43
0.4 0.4
0.2 0.2
0 0
-0.2 -0.2
-0.4 -0.4
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p511 panel pixel p512
0.4 0.4
0.2 0.2
0 0
-0.2 -0.2
-0.4 -0.4
0 50 100 150 0 50 100 150
Bands Processed Bands Processed
panel pixel p52 panel pixel p53
the false alarm probability, PF. Instead, we choose the common practice used in
medical diagnosis to calculate the area under an ROC curve, referred to as the area
under curve (AUC) in Chang et al. (1998b). When such an AUC is used, an
alternative ROC curve can be plotted in terms of AUC on the y-axis versus band
number on the x-axis. Figure 13.9 plots five ROC curves of (number of bands, area
412 13 RHBP for Active Target Detection: Constrained Energy Minimization
Fig. 13.4 3D progressive plots of band-varying RHBP-CEM detection maps of 19 R panel pixels
plus p212 specified by the signature, p1
Fig. 13.5 3D progressive plots of band-varying RHBP-CEM detection maps of 19 R panel pixels
plus p212 specified by the signature, p2
Fig. 13.6 3D progressive plots of band-varying RHBP-CEM detection maps of 19 R panel pixels
plus p212 specified by the signature, p3
Fig. 13.7 3D progressive plots of band-varying RHBP-CEM detection maps of 19 R panel pixels
plus p212 specified by the signature, p4
Fig. 13.9 demonstrates the fact that in order for CEM to be effective, there is no
need to use full bands. As a matter of fact, less than one-third of the bands are
sufficient for good CEM. In addition, the plots in Fig. 13.9 also provide information
on how a band affects CEM detection performance. This is particularly interesting
in Fig. 13.9a, where the progressive band-varying detected abundance fractions had
ups and downs between 8 and 50. Such valuable information is not offered by
detection techniques that use full bands to process data.
414 13 RHBP for Active Target Detection: Constrained Energy Minimization
Fig. 13.8 3D progressive plots of band-varying RHBP-CEM detection maps of 19 R panel pixels
plus p212 specified by the signature, p5
The data to be studied were collected by the Hyperion sensor mounted on the
Earth Observer 1 (EO-1) satellite. Hyperion uses a high-resolution hyperspectral
imager to record Earth surface images of approximately 7.5 100 km with a
30 m spatial resolution and a 10 nm spectral resolution. The image scene
EO1H0150322011201110K3 was downloaded from the USGS Earth Explorer
Web site (http://earthexplorer.usgs.gov/) and is shown in Fig. 1.13. The image
used is a Hyperion L1 data product that includes 198 channels of calibrated
spectral information. This image is reproduced in Fig. 13.10 and was then
cropped to select a region of interest (ROI) covering the western side of the
Chesapeake Bay Bridge. The resulting image cube is a square with spatial
dimension of size 64 64 and 198 spectral bands ranging from 426 to
2396 nm. The image was then compared with higher-resolution areal imagery
to identify five distinct areas of interest (AOIs). The areas include a beach,
Westinghouse Bay, Mezick Ponds, farmland, and a large corporate building.
Four pixels were selected at random from each of these areas and used to generate
an average spectral signature for the material contained in the area, with the
exception of the farmland, which used seven pixels owing to the fact that it is
composed of four disjoint regions in the image where four pixels were selected
from the largest farmland region and a single pixel was chosen from the
remaining three regions. A color image of the region and the spectral signatures
of each of the five AOIs are shown in Fig. 13.10.
Using the spectral signatures of five AOIs plotted in Fig. 13.10c as the desired
target signatures for CEM, Fig. 13.11a–e shows detected amounts of five AOIs, a
beach, Westinghouse Bay, Mezick Ponds, farmland, and a corporate building, by
13.4 Real Image Experiments 415
a b
1 1
0.95 0.95
0.9
0.9
0.85
0.85 0.8
AUC
AUC
0.8 0.75
0.75 0.7
0.65
0.7
0.6
0.65 0.55
0.5
0 20 40 60 80 100 120 140 160 180 0 20 40 60 80 100 120 140 160 180
Band Band
c 1
d
1
0.95 0.95
0.9 0.9
0.85 0.85
0.8 0.8
AUC
0.75
AUC
0.75
0.7 0.7
0.65 0.65
0.6 0.6
0.55 0.55
0.5 0.5
0 20 40 60 80 100 120 140 160 180 0 20 40 60 80 100 120 140 160 180
Band Band
e
1
0.95
0.9
0.85
0.8
AUC
0.75
0.7
0.65
0.6
0.55
0.5
0 20 40 60 80 100 120 140 160 180
Band
Fig. 13.9 ROC curves of (number of bands, area under curve (AUC)) for band-varying progres-
sive detection performance of RHBP-CEM. (a) ROC curve for panels in row 1. (b) ROC curve for
panels in row 2. (c) ROC curve for panels in row 3. (d) ROC curve for panels in row 4. (e) ROC
curve for panels in row 5
a b
c
10000
Beach
9000 Westinghouse Bay
Mezick Ponds
8000 Farm Land
Corporate Building
7000
6000
5000
4000
3000
2000
1000
0
0 50 100 150
Fig. 13.10 (a) Cropped Hyperion image scene EO1H0150322011201110K3. (a) Cropped Hype-
rion image scene EO1H0150322011201110K3 with five areas of interest (AOIs): beach (dark
blue), Westinghouse Bay (green), Mezick Ponds (orange), farmland (light blue), and a corporate
building (magenta). (c) Average spectral signatures for each of the five AOIs
what they represented. The pixels used to generate the desired spectral signatures
were excluded, with the exception of the corporate building, which only consisted
of the four pixels used to generate the spectral signature. This allows a comparison
of the response for each region relative to each of the five corresponding signatures.
Unlike the HYDICE data, there is no complete ground truth about specific targets
other than the AOIs selected. In addition, the purity of the used desired signatures
d was smeared by the 30 m spatial resolution so that the CEM-detected abundance
fractions fluctuated. Nevertheless, analogous to HYDICE experiments, the plots in
Fig. 13.11 also indicated that there was no need to use the complete set of bands,
and in most cases less than 50 bands were required to perform CEM.
13.4 Real Image Experiments 417
a b
1 Beach 1 Beach
Westinghouse Bay Westinghouse Bay
Mezick Ponds Mezick Ponds
Relative Abundance
Relative Abundance
0.6 0.6
0.4 0.4
0.2 0.2
0 0
50 100 150 50 100 150
Bands Processed Bands Processed
c d
1 Beach 1 Beach
Westinghouse Bay Westinghouse Bay
Mezick Ponds Mezick Ponds
Relative Abundance
Relative Abundance
0.6 0.6
0.4 0.4
0.2 0.2
0 0
50 100 150 50 100 150
Bands Processed Bands Processed
e
1
Relative Abundance
0.8 Beach
Westinghouse Bay
Mezick Ponds
0.6 Farm Land
Corporate Building
0.4
0.2
0
50 100 150
Bands Processed
Fig. 13.11 Normalized values of five AOIs: beach, Westinghouse Bay, Mezick Ponds, farmland,
corporate building by RHBP-CEM with desired material signatures. (a) d ¼ beach signature, (b)
d ¼ Westinghouse Bay signature, (c) d ¼ Mezick Ponds signature, (d) d ¼ farmland signature, and
(e) d ¼ corporate building signature
418 13 RHBP for Active Target Detection: Constrained Energy Minimization
a b c d
e f g h
Fig. 13.12 Result of RHBP-CEM using Westinghouse Bay signature after receiving: (a) 1 band,
(b) 2 bands, (c) 10 bands, (d) 20 bands, (e) 30 bands, (f) 40 bands, (g) 60 bands, and (h) all bands
Since the plots in Fig. 13.11 do not illustrate how detection is performed by
CEM in a progressive manner, Fig. 13.12a–g visually depicts the progression in the
detected amounts of abundance for the entire image using the Westinghouse Bay
spectral signature for the desired signature d as the number of bands processes
increased. Similar results are also obtained for each spectral signature, with the
exception of the Mezick Ponds signature, which resulted in poor discrimination;
however, they are not included here. It is worth noting that plots similar to Fig. 13.9
were not applicable owing to a lack of specific ground truth in the AOIs.
Interestingly, as shown in Fig. 13.12, after 20 bands, CEM was able to outline
the bay area in Fig. 13.12d and to improve the detection results up to 40 bands in
Fig. 13.12e–g, and then its silhouette began to gradually deteriorate until full bands
were used (Fig. 13.12h). This experiment further demonstrates that progressive
band processing of CEM offers unique advantages regarding how a detected target
varies its detected abundance in a progressive manner. Since no ground truth is
provided by this Hyperion data, we are not able to perform ROC analysis for
quantitative study.
The computational time of RHBP-CEM was compared to that of CEM. Each
algorithm was run and executed in MATLAB R2012A with an Intel Core i7-3770
running at 3.4 GHz with 16 GB RAM on the HYDICE image five times to produce
an average computing time. While CEM required 34.3 ms to complete processing,
RHBP-CEM required 106.9 ms to process the first band and 380.2 ms to process all
bands. This is mainly due to the fact that CEM only calculates the global sample
correlation matrix R prior to its processing, and once it is done, it does not have to
do it again. By contrast, RHBP-CEM must recalculate the sample-varying correla-
tion matrix via a recursive equation specified by (13.4). Of course, RHBP-CEM
13.5 Graphical User Interface Design 419
requires more time for data processing than CEM. However, such a disadvantage is
offset by several benefits. One is progressive detection performance as bands are
progressively processed. Second, with such progressive performance we can learn
the impact of each band on CEM performance. Third, the recursive equation (13.4)
provides a feasible implementation of hardware design. Finally, RHBP-CEM can
begin as soon as the first band has been received, as opposed to CEM, which must
wait for data to be completed before it can calculate matrix R.
A graphical user interface (GUI) with a screenshot shown in Fig. 13.13 was
developed using MATLAB’s Guide to aid in algorithm performance analysis.
The GUI allows the user to load different data sources for analysis. At the bottom
of the GUI are three image display windows. The images displayed from left to
right are a color image of the scene, a grayscale image of the last band processed,
and the current abundance image. Once the data are loaded, the user is given an
option to select a pixel from the image to be the target material d. Once the desired
pixel is selected, the user can start processing by clicking the start button. The
program then simulates data transmission using the MATLAB timer function. At
each timer tick, one complete band of data is considered to be received, and RHBP-
CEM processes the newly received data. Upon completion of processing, the
abundance image and the most recently processed band images are updated to
allow the user to observe the results. A textbox indicating the current band is
displayed at the top center so that the user can easily tell how many bands have
been processed. The cycle repeats until all bands have been processed or the user
clicks on the stop button. If the processing is stopped, it may be continued from the
current band or reset to begin processing at the first band. The user may adjust the
simulated transmission rate by adjusting a slide bar in the upper left-hand corner of
the GUI. The range of data rates varies from 66 kbps to 66 Mbps and assumes the
data consist of 16 bit numbers. The user is also given an opportunity to observe the
spectral signatures of both the desired material and any pixel in the abundance
image. The two spectral signatures are located above their corresponding images.
Their corresponding pixels are highlighted by a red circle for the desired material
d and green for the observed pixel r. In the case of the desired material spectral
signature, the complete signature is displayed. In the case of the observed pixel
spectral signature, only the bands that have been received and processed are
displayed. Finally, the normalized value for the observed pixel is displayed in the
center of the GUI.
According to Fig. 13.13, the developed GUI performs RHBP-CEM whenever
new bands are added. The processing can be carried out band by band in real time.
13.6 Conclusions
This chapter presented a new look at CEM, called recursive hyperspectral band
processing of CEM (RHBP-CEM), which makes use of the causal band correlation
matrix (CBCRM) to adapt CEM to different bands. Since CBCRM varies with
bands, a recursive innovation information update equation was specifically derived
to avoid repeatedly recalculating CBCRM where only information provided by a
new incoming band is needed to be processed for updating results. As a result,
RHBP-CEM can be performed as if it were a real-time processing algorithm with
respect to data transmission and communication when ground stations receive data
sequentially from different bands. In this case, RHBP-CEM can be performed
instantly as the information in new bands is received. To further include its practical
value in applications, a GUI software program was also developed to allow users to
process CEM band by band progressively in real time according to a BSQ format.
Chapter 14
Recursive Hyperspectral Band Processing for
Passive Target Detection: Anomaly Detection
14.1 Introduction
The causal band sample correlation matrix (CBCRM) introduced in Sect. 13.2 and
XN
defined by Rll ¼ ð1=N Þ i¼1 ri ðlÞ½ri ðlÞT in (13.1) is formed by image pixel
N
vectors fri ðlÞgi¼1 using image pixels in all previously visited band images, Ωl1
¼ fBi gl1 , up to the current lth band image, Bl. Note that each band image Bj has a
i¼1
N
total number of N pixels denoted by r ij i¼1, where rij is the ith pixel in the jth band
image. According to (13.5), the inverse of Rll can be calculated by
424 14 RHBP for Passive Target Detection: Anomaly Detection
" #1
T 1
T
Xl1 Xl1 Xl1 xðlÞ
R1¼
ll ð1=N ÞXl X l ¼ N
xT ðlÞXl1
T
xT ðlÞxðlÞ
2 1 h 1 iT 3
1
ð1=N ÞXl1 Xl1T
Xl1 xðlÞ ð1=N ÞXl1 Xl1 T
Xl1 xðlÞ
6 ð1=N ÞX X T 1 þ ð1=N Þ ð1=N ÞXl1 Xl1 Xl1 xðlÞ 7
T
6 l1 l1 7
6 xT ðlÞP⊥ T xðlÞ xT ðlÞP⊥ T xðlÞ
7
6 Xl1 Xl1 7
¼66 h iT
7
7
6 1 7
6 ð 1=N ÞX X
l1 l1
T
X l1 xð lÞ 1 7
4 ⊥
N x ðlÞPX T xðlÞ
T 5
⊥
x ðlÞPX T xðlÞ
T l1
l1
2 h iT 3
R1 1
ðl1Þðl1Þ Xl1 xðlÞ Rðl1Þðl1Þ Xl1 xðlÞ R1
ðl1Þðl1Þ Xl1 xðlÞ
6 R1 þ ð Þ 7
6 ðl1Þðl1Þ 1=N 7
6 xT ðlÞP⊥ T xðlÞ xT ðlÞP⊥ T xðlÞ 1 7
6 Xl1 Xl1 7
¼66 ⊥
N x ðlÞPX T xðlÞ
T 7:
h iT 7
6 1
Rðl1Þðl1Þ Xl1 xðlÞ
l1
7
6 7
4 5
xT ðlÞP⊥X T xðlÞ
l1
ð14:1Þ
R1
ðl1Þðl1Þ ;
2. Innovation information:
1
Xl1 xðlÞ, xT ðlÞP⊥
X T xð l Þ ;
l1
3. Input information:
xðlÞ ¼ ðr 1l ; r 2l ; . . . ; r Nl ÞT :
Using (14.1), a causal band-based R-AD (CBR-AD), δCBRAD ðrðlÞÞ can be
derived as follows:
δCBR-AD ðrðlÞÞ
¼ rT ðlÞðRll Þ1 rðlÞ
¼ ðrT ðl 1Þ,r l Þ
2 1 T 3
RðTl1Þðl1Þ þ ð1=N Þβ
1
Rðl1Þðl1Þ Xl1 xðlÞ Nβ
l ðl1Þ l ðl1Þ
!
rðl 1Þ
rl
1 T
¼ r ðl 1Þ Rðl1Þðl1Þ rðl 1Þ þ ð1=N Þβ
T
rT ðl 1ÞR1 1
ðl1Þðl1Þ Xl1 xðlÞ Rðl1Þðl1Þ Xl1 xðlÞ rðl 1Þ
l ðl1Þ
T
r l β
Rð1
l1Þðl1Þ Xl1 xl rðl 1Þ β
rT ðl 1ÞR1
ð14:2Þ
14.3 Recursive Hyperspectral Band Processing of Anomaly Detection 425
n h i o1
where η
¼ rT ðl 1ÞR1
As noted in (14.2), CBR-AD, δCBR ‐ AD(r(l )) operating on the lth band, can be
updated recursively via η
, β
, and R1
ðl1Þðl1Þ , but not the previous result,
l ðl1Þ l ðl1Þ
CBR-AD
δ ðrðl 1ÞÞ, obtained from the previous (l 1)st band.
14.3.1 RHBP-R-AD
In fact, we further show that δCBR ‐ AD(r(l )) can be easily updated by the previous
anomaly detector δCBR-AD ðrðl 1ÞÞ obtained using the first (l 1) bands, which is
already available, and new band information rl, along with two pieces of innovation
information, β
and η
δRHBP-R-AD ðrðlÞÞ
¼ δRHBP-R-AD ðrðl 1ÞÞ þ ð1=N Þβ
η2
2β
η
r l þ Nβ
r2
l ðl1Þ l
ðl1Þ l ðl1Þ l
ðl1Þ l ðl1Þ l
2
¼δ RHBP-R-AD
Nr l η
:
l ðl1Þ l ðl1Þ
ð14:3Þ
¼ r T ð l 1Þ
l ðl1Þ
n h i o1
R1 ð Þ, β
¼ T
ð Þ ⊥
ð Þ
l
ðl1Þ
X
ðl1Þðl1Þ l1 x l and x l P T
Xl1
x l .
In regard to computing R1
ll via (14.1) the recursive structure derived in (14.2) has
many significant advantages over (14.1) in terms of how to effectively use informa-
tion summarized as follows. In particular, there is no innovation information derived
in (14.1). Theoretically speaking, (14.2) can also be implemented in real time, but its
computational complexity involved may be at the cost of its real-time processing
capability. In fact, (14.3) is a much better alternative to δCBR ‐ AD(r(l )) in (14.2)
because it makes AD more effective and efficient in real time implementation by
only updating δRBP-R-AD ðrðl 1ÞÞ, along with the innovation information:
426 14 RHBP for Passive Target Detection: Anomaly Detection
1. Initial conditions:
r(1), R(1) ¼ (1/N )r(1)rT(1);
2. Newly received band information: input information from the lth band image of
N
all data sample vectors: fr il gi¼1 to form x(l );
3. Processed information produced by the previously visited l 1 band images r
(l 1) and Xl1:
R1
ðl1Þðl1Þ ;
¼ xT ðlÞ P⊥ T x ð l Þ ;
l ðl1Þ X l1
¼ rT ðl 1ÞR1
ðl1Þðl1Þ Xl1 xðlÞ:
l ðl1Þ
14.3.2 RHBP-K-AD
Despite the fact that R-AD is derived from K-AD defined in (6.1), finding the
inverse of a causal sample covariance matrix for K-AD is more complicated than
that for a causal correlation matrix since there is no sample mean involved in R-AD.
However, once R is calculated, K can be calculated by K ¼ R μμT .
Unlike the CBR-AD, which uses the autocorrelation matrix without
calculating the sample mean, a real-time CBK-AD requires the causal sample
mean prior to calculating the causal covariance matrix. Assume that
XN
μðlÞ ¼ ðμ1 ; . . . ; μl ÞT , with μj ¼ ð1=N Þ i¼1 r ji for 1 j l and Kll ¼ ð1=N Þ
XN
ðri ðlÞ μðlÞÞðri ðlÞ μðlÞÞT . Then Kll ¼ Rll μðlÞμT ðlÞ,
i¼1
XN
μðl 1Þ ¼ ð1=N Þ i¼1 ri ðl 1Þ, and
14.3 Recursive Hyperspectral Band Processing of Anomaly Detection 427
μðl 1Þ T μðl 1ÞμT ðl 1Þ μðl 1Þμl
μðlÞμ ðlÞ ¼
T
μ ðl 1Þ μl ¼ :
μl μl μT ðl 1Þ μ2l
K1 1
ll can be calculated through Rll by letting A ¼ Rll , u ¼ μðlÞ, and v ¼ μðlÞ
in (14.4) as follows:
T
1 R1 1
ll μðlÞ μ ðlÞRll
K1 ¼ Rll μðlÞμ ðlÞ T
¼ R1 þ : ð14:5Þ
ll ll
1 μT ðlÞR1
ll μðlÞ
Then R1 1
ll μðlÞ and μ ðlÞRll μðlÞ in (14.5) can be further calculated as
T
T
Φll ¼ R1ll μðlÞ μ ðlÞRll
1
1 μðl 1ÞμT ðl 1Þ μðl 1Þμl
¼ Rll R1
ll ; ð14:6Þ
μ l μ T ð l 1Þ μ2l
with
Aðl1Þðl1Þ b μðl 1Þ
R1
ll μðlÞ ¼
bT d μl
Aðl1Þðl1Þ μðl 1Þ þ bμl
¼ ð14:7Þ
bT ðl 1Þμðl 1Þ þ dμl
and
T
1 μðl 1Þ Aðl1Þðl1Þ b μðl 1Þ
ρ
¼ μ ðlÞRll μðlÞ ¼
T
l ðl1Þ μl bT d μl ð14:8Þ
¼ μ ðl 1ÞAðl1Þðl1Þ μðl 1Þ þ 2μ ðl 1Þbμl þ dμ2l ;
T T
R1
ðl1Þðl1Þ Xl1 xðlÞ,
l ðl1Þ
n h i o1
β
¼ xT ð l Þ P⊥
X T xðlÞ ; ð14:9Þ
l ðl1Þ l1
d ¼ Nβl|(l1), and
428 14 RHBP for Passive Target Detection: Anomaly Detection
1
Aðl1Þðl1Þ ¼ RðTl1Þðl1Þ
T ð14:10Þ
þ ð1=N Þβ
R1 1
ðl1Þðl1Þ Xl1 xðlÞ Rðl1Þðl1Þ Xl1 xðlÞ ;
l ðl1Þ
Φll
K1 1
ll ¼ Rll þ : ð14:11Þ
1 ρ
l ðl1Þ
l ðl1Þ
2 3
T4 Φll 5ðrðlÞ μðlÞÞ
¼ ðrðlÞ μðlÞÞ T
R1
ll ðrðlÞ μðlÞÞ þ ðrðlÞ μðlÞÞ
1 ρ
l ðl1Þ
T
rðl 1Þ μðl 1Þ Aðl1Þðl1Þ b rðl 1Þ μðl 1Þ
¼
r l μl bT d r l μl
T
1 rðl 1Þ μðl 1Þ rðl 1Þ μðl 1Þ
þ Φ ;
1 ρ
r l μl r l μl
ll
l ðl1Þ
ð14:12Þ
where ρ
¼ μT ðlÞR1
ll μðlÞ is given by (14.8).
l ðl1Þ
According to (14.12), CBK-AD can also be considered a recursive hyperspectral
band processing of K-AD (RHBP-K-AD) since it also implements R1 ll μðlÞ via
(14.7) recursively. However, implementing RHBP-K-AD in real time involves more
computation than RHBP-R-AD because it requires additional calculation of the
causal sample mean, μ(l ). Specifically, to do so we need another matrix identity,
Woodbury’s identity, which is not required for RHBP-AD. Also, (14.12) is not
recursive like (14.3) because it does not update δRHBP ‐ K ‐ AD(r(l )) through
δRHBP-K-AD ðrðl 1ÞÞ but rather calculates R1
ðl1Þðl1Þ through A and b in (14.7).
Finally, note that when the previously derived RHBP-R-AD and RHBP-K-AD
are applied to processing AD band by band according to BSQ progressively as well
as recursively, they are different from the real-time causal anomaly detectors
developed in Chang et al. (2014), which require full band information for each of
the data sample vectors according BIP/BIS.
The image data to be studied are the HYDICE image scene shown in Fig. 14.1a (and
Fig. 1.10a), which has a size of 64 64 pixel vectors with 15 panels in the scene and
the ground truth map in Fig. 14.1b (Fig. 1.10b).
14.4 Real Image Experiments 429
a b
p11, p12, p13
Fig. 14.1 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels
The image was acquired by 210 spectral bands with a spectral coverage from 0.4
to 2.5 μm. Low signal/high noise bands, bands 1–3 and 202–210, and water vapor
absorption bands, bands 101–112 and 137–153, were removed. Thus, a total of
169 bands were used in the experiments. The spatial resolution and spectral
resolution of this image scene are 1.56 m and 10 nm, respectively. As noted in
previous chapters, the panel pixel p212, marked yellow in Fig. 14.1b, is of particular
interest. It is always extracted as the one with the most spectrally distinct signature
compared to the R panel pixels in row 2. This indicates that a signature of spectral
purity is not equivalent to a signature of spectral distinction. Also, because of such
ambiguity, the panel signature representing panel pixels in the second row is either
p221 or p212, which is always difficult to determine by endmember finding algo-
rithms. This implies that the ground truth of R panel pixels in the second row in
Fig. 14.1b may not be as pure as was thought.
Figures 14.2 and 14.3 plot progressive changes in the detection of each of 19 R
panel pixels plus a yellow panel pixel, p212, as shown in Fig. 14.1b, produced by
RHBP-R-AD and RHBP-K-AD, respectively, where the x-axis is the number of
bands processed and the y-axis is the amount of detected abundance fractions. Note
that, based on the number of processed bands, say, n in Figs. 14.2 and 14.3, the first
n bands are used to process the data.
For a better visual assessment of Fig. 14.3 we plot Figs. 14.4 and 14.5 in three-
dimensional plots, where the x-axis specifies 19 R panel pixels plus a yellow panel
pixel denoted by p212, the y-axis denotes the number of bands being processed, and
the z-axis is R-AD or K-AD detected abundance fractions.
These figures clearly show three significant changes in the detected abundances
in ranges of processed bands around 20, 50, and 110 bands. These three major
ranges of processed bands correspond to Figs. 14.4 and 14.5 where the first 16–21
bands, the first 43–57 bands, and the first 103–113 bands are highlighted in pink to
show the dramatic changes in the amounts of detected abundance fractions for both
RHBP-R-AD and RHBP-K-AD.
Using the information of the three ranges of processed bands provided by
Figs. 14.2, 14.3 and 14.4, Figs.14.5, 14.6, and 14.7 show the progressive detection
430 14 RHBP for Passive Target Detection: Anomaly Detection
p11: R-AD Value vs. Bands p12: R-AD Value vs. Bands p13: R-AD Value vs. Bands
2000 2000 2000
R-AD Value
R-AD Value
R-AD Value
1000 1000 1000
0 0 0
0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands
p211: R-AD Value vs. Bands p212: R-AD Value vs. Bands p221: R-AD Value vs. Bands
2000 2000 2000
R-AD Value
R-AD Value
0 0 0
0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands
p22: R-AD Value vs. Bands p23: R-AD Value vs. Bands
2000 2000
1500 1500
R-AD Value
R-AD Value
1000 1000
500 500
0 0
0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands
p311: R-AD Value vs. Bands p312: R-AD Value vs. Bands p32: R-AD Value vs. Bands p33: R-AD Value vs. Bands
2000 2000 2000 2000
R-AD Value
R-AD Value
R-AD Value
0 0 0 0
0 50 100 150 0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands Number of Processed Bands
p411: R-AD Value vs. Bands p412: R-AD Value vs. Bands p42: R-AD Value vs. Bands p43: R-AD Value vs. Bands
2000 2000 2000 2000
R-AD Value
R-AD Value
R-AD Value
0 0 0 0
0 50 100 150 0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands Number of Processed Bands
p511: R-AD Value vs. Bands p521: R-AD Value vs. Bands p52: R-AD Value vs. Bands p53: R-AD Value vs. Bands
2000 2000 2000 2000
R-AD Value
0 0 0 0
0 50 100 150 0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands Number of Processed Bands
Fig. 14.2 Detected abundance fractions of 19 R panel pixels plus p212 by RHBP-R-AD versus
number of bands processed
14.4 Real Image Experiments 431
p11: K-AD Value vs. Bands p12: K-AD Value vs. Bands p13: K-AD Value vs. Bands
2000 2000 2000
K-AD Value
K-AD Value
K-AD Value
1000 1000 1000
0 0 0
0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands
p211: K-AD Value vs. Bands p212: K-AD Value vs. Bands
p221: K-AD Value vs. Bands
2000 2000
2000
1500 1500
1500
K-AD Value
K-AD Value
K-AD Value
1000 1000
1000
500
500 500
0
0 0 50 100 150 0
0 50 100 150 0 50 100 150
Number of Processed Bands
Number of Processed Bands Number of Processed Bands
p22: K-AD Value vs. Bands p23: K-AD Value vs. Bands
2000 2000
1500 1500
K-AD Value
K-AD Value
1000 1000
500 500
0 0
0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands
p311: K-AD Value vs. Bands p312: K-AD Value vs. Bands p32: K-AD Value vs. Bands p33: K-AD Value vs. Bands
2000 2000 2000 2000
K-AD Value
K-AD Value
0 0 0 0
0 50 100 150 0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands Number of Processed Bands
p411: K-AD Value vs. Bands p412: K-AD Value vs. Bands p42: K-AD Value vs. Bands p43: K-AD Value vs. Bands
2000 2000 2000 2000
K-AD Value
K-AD Value
K-AD Value
0 0 0 0
0 50 100 150 0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands Number of Processed Bands
p511: K-AD Value vs. Bands p521: K-AD Value vs. Bands p52: K-AD Value vs. Bands p53: K-AD Value vs. Bands
2000 2000 2000 2000
K-AD Value
K-AD Value
K-AD Value
0 0 0 0
0 50 100 150 0 50 100 150 0 50 100 150 0 50 100 150
Number of Processed Bands Number of Processed Bands Number of Processed Bands Number of Processed Bands
Fig. 14.3 Detected abundance fractions of 19 R panel pixels plus p212 by RHBP-K-AD versus
number of bands processed
432 14 RHBP for Passive Target Detection: Anomaly Detection
R-AD Value
2000
1500
1000
500
0
10
16
21
30
43
Nu
50
mb
57
er
70
of
Pro
90
ce
ss
103
ed
110
Ba
113
nd
s
130
P521 P52 P53
P43 P511
150 P412 P42
P33 P411
P312 P32
P23 P311
170 p212 P221 P22 ne l Pi xels
P13 P211 20 Pa
P11 P12
Fig. 14.4 3D plots of 19 R panel pixels plus p212 detected by RHBP-R-AD versus number of
bands processed
2000
K-AD Value
1500
1000
500
0
10
16
21
30
43
Nu
50
mb
57
er
70
of
Pro
90
ce
ss
103
ed
110
Ba
113
nd
130
s
P53
P521 P52
P43 P511
150 P412 P42
P33 P411
P312 P32
P22 P23 P311 ls
p212P221 l Pixe
170 P13 P211 20 Pane
P11 P12
Fig. 14.5 3D Plots of 19 R panel pixels plus p212 by RHBP-K-AD versus number of bands
processed
Fig. 14.6 Results of RHBP-R-AD of Hydice Data after receiving: (a) 16 bands (b) 17 bands (c)
18 bands (d) 19 bands (e) 20 bands (f) 21 bands (g) 43 bands (h) 45 bands (i) 47 bands (j) 49 bands
(k) 51 bands (l) 53 bands (m) 103 bands (n) 105 bands (o) 107 bands (p) 109 bands (q) 111 bands
(r) 113 bands
detection of panel pixels in rows 4 and 5 improved, until the first 21 bands were
processed. Similarly, the panel pixels in rows 1–3 were detected by the second
range of processed bands from almost being picked up when the first 43 bands were
processed to clearly being detected using the first 51–53 bands. Finally, using the
third range of processed bands, the detection of 19 R panel pixels in the five rows
pretty much stayed the same, but the interferer at the left corner, which was not
detected by the previous first and second ranges of processed bands, was clearly
detected as a very bright pixel. Interestingly, this strong interferer is always the first
target detected by an unsupervised target detection algorithm, the automatic target
generation process (ATGP) developed by Ren and Chang (2003).
Since we have the ground truth of the 19 R panel pixels and the yellow panel
pixel p212, we can further calculate the detection rates of RHBP-R-AD and RHBP-
K-AD via a receiver operating characteristic (ROC) analysis for performance
evaluation (Poor 1994). To include the number of processed bands as a parameter
to see their progressive detection performance on these 19 R panel pixels, we
434 14 RHBP for Passive Target Detection: Anomaly Detection
calculated the area under an ROC curve, referred to as area under curve (AUC) in
Metz (1978). Using such an AUC, an alternative ROC curve can be plotted in
terms of AUC on the y-axis versus number of processed bands. Figure 14.8a–c plots
the AUC curves versus the number of processed bands with (a) an ROC curve of
(number of bands, area under curve (AUC)) for panel pixels in column 1, p11, p211,
p212, p221, p311, p312, p411, p412, p511, and p521; (b) an ROC curve of (band, AUC) for
p12, p22, p32, p42, p52; and (c) an ROC curve of (band, AUC) for p13, p23, p33, p43,
p53.
As shown in Fig. 14.8, RHBP-K-AD generally performed better than RHBP-R-
AD when the number of processed bands was smaller than 45 (Fig. 14.8a),
40 (Fig. 14.8b), and 30 (Fig. 14.8c). Nevertheless, after 45 bands were processed,
both yielded nearly same detection performance.
Similarly, Fig. 14.9a–e plots the AUC curves versus the number of processed
bands with (a) an ROC curve of (band, AUC) for panel pixels in row 1, p11, p22, p33;
(b) an ROC curve of (band, AUC) for panel pixels in row 2, p211, p212, p221, p22, p23;
(c) an ROC curve of (band, AUC) for panel pixels in row 3, p311, p312, p32, p33;
(d) an ROC curve of (band, AUC) for row 4, p411, p412, p42, p43; and (e) an ROC
curve of (band, AUC) for row 5, p511, p521, p52, p53.
14.4 Real Image Experiments 435
Fig. 14.7 Results of RHBP-K-AD of HYDICE data after receiving: (a) 16 bands, (b) 17 bands, (c)
18 bands, (d) 19 bands, (e) 20 bands, (f) 21 bands, (g) 43 bands, (h) 45 bands, (i) 47 bands, (j)
49 bands, (k) 51 bands, (l) 53 bands, (m) 103 bands, (n) 105 bands, (o) 107 bands, (p) 109 bands,
(q) 111 bands, (r) 113 bands
Finally, Fig. 14.10 plots the curves of (band, AUC) for the progressive detection
performance of RHBP-R-AD and RHBP-K-AD for all 19 R panel pixels plus the
p212 yellow panel pixel.
As a second example, an Airborne Visible/InfraRed Image Spectrometer
(AVIRIS) image scene shown in Fig. 14.11a (and Fig. 1.8) was used for experi-
ments. It is the Lunar Crater Volcanic Field (LCVF) located in Northern Nye
County, Nevada.
Figure 14.12a plots the detected amount of abundance fractions of six signatures
specified in Fig. 14.11b by RHBP-R-AD versus the number of processed bands,
where it is very clear that the detection curve of anomaly increased exponentially as
the number of processed bands grew, as with vegetation at a slightly slower pace.
By contrast, the detection curves of the four signatures seemed to be linear. A
comparison of Fig. 14.12a with Figs. 14.2 and 14.4 reveals that LCVF has stable
and slow-changing R-AD detected values as the number of processed bands
increased, as opposed to those in Fig. 14.2, which had more dynamic ranges with
clear cuts on detecting pure panel pixels. In particular, there seemed to be three
ranges of interest in the detection of rapid changes in R-AD detected values , which
are three ranges of band numbers, 20 to 30, 57 to 69 and 110 to 110. For a better
visual assessment, Fig. 14.12b also plots the RHBP-R-AD-detected amounts of
abundance fractions of six signatures versus the number of processed bands in a 3D
view, which also confirmed the results.
436 14 RHBP for Passive Target Detection: Anomaly Detection
a b
1 1
0.9 0.9
0.8 0.8
AUC
AUC
0.7 0.7
0.6 0.6
PBP-R-AD PBP-R-AD
PBP-K-AD PBP-K-AD
0.5 0.5
0 25 50 75 100 125 150 0 25 50 75 100 125 150
Number of Processed Bands Number of Processed Bands
c
1
0.9
0.8
AUC
0.7
0.6
PBP-R-AD
PBP-K-AD
0.5
0 25 50 75 100 125 150
Number of Processed Bands
Fig. 14.8 Plots of (number of bands, area under curve (AUC)) for progressive detection perfor-
mance of RHBP-R-AD and RHBP-K-AD: (a) ROC curve of (number of bands, AUC) for panel
pixels in column 1, p11, p211, p212, p221, p311, p312, p411, p412, p511, p521; (b) ROC curve of (number
of bands, AUC) for p12, p22, p32, p42, p52; (c) ROC curve of (number of bands, AUC) for p13, p23,
p33, p43, p53
As we can see, in the first range of processed band numbers, vegetation was
detected, while an anomaly was detected in the second range of processed bands.
Interestingly, adding the third range of processed bands did not much improve the
detection of vegetation and anomalies. Most importantly, since there are large areas
of cinders, rhyolite, playa (dry lake), and shade they are not considered anomalies.
As a result, it makes sense that none of these four areas was detected in Figs. 14.14
and 14.15.
a b
1 1
0.9 0.9
0.8 0.8
AUC
AUC
0.7 0.7
0.6 0.6
PBP-R-AD PBP-R-AD
PBP-K-AD PBP-K-AD
0.5 0.5
0 25 50 75 100 125 150 0 25 50 75 100 125 150
Number of Processed Bands Number of Processed Bands
c 1
d
1
0.9 0.9
0.8 0.8
AUC
AUC
0.7 0.7
0.6 0.6
PBP-R-AD PBP-R-AD
PBP-K-AD PBP-K-AD
0.5 0.5
0 25 50 75 100 125 150 0 25 50 75 100 125 150
Number of Processed Bands Number of Processed Bands
e 1
0.9
0.8
AUC
0.7
0.6
PBP-R-AD
PBP-K-AD
0.5
0 25 50 75 100 125 150
Number of Processed Bands
Fig. 14.9 Plots of (band, AUC) for progressive detection performance of RHBP-R-AD and
RHBP-K-AD: (a) ROC curve of (band, AUC) for panel pixels in row 1, p11, p12, p13; (b) ROC
curve of (band,AUC) for row 2, p211, p212, p221, p221, p22, p23; (c) ROC curve of (band,AUC) for
row 3, p311, p312, p32, p33; (d) ROC curve of (band,AUC) for row 4, p411, p412, p42, p43; (e) ROC
curve of (band,AUC) for row 5, p511, p521, p52, p53
14.5 Computing Time Analysis 439
0.8
AUC
0.7
0.6
PBP-R-AD
PBP-K-AD
0.5
0 25 50 75 100 125 150
Number of Processed Bands
The computing time of RHBP-AD was compared to the traditional AD. Each
algorithm was run and executed in MATLAB R2011B on an Intel Core i7-3770
running at 3.40 GHz with 16 GB RAM on the HYDICE images five times to
produce an average computing time. Figure 14.16 plots the computing time versus
the number of processed bands for both RHBP-R-AD and RHBP-K-AD running on
the HYDICE and LCVF data sets, where RHBP-R-AD required less time than
RHBP-K-AD owing to the fact that the former implements a recursive equation
specified by (14.6), while the latter implements a nonrecursive equation specified
by (14.12). As also shown in Fig. 14.16, the computing time increases as new bands
are added. This is because an increase in the size of R and K results in higher
computational complexity. The reason for the drastic changes in the second band is
the calculation of initial conditions. Once initial conditions are calculated, no
dramatic changes in computing time occur, as shown in Fig. 14.16. It is also
worth noting that the fluctuations of the plots resulted from numerical computations
by computer implementation.
In addition, R-AD required 0.44 s to complete the entire process, while RHBP-
R-AD spent 0.02 s to complete processing of the first band and a total of 32.43 s was
required to process the entire image cube with 169 bands. For the global K-AD,
0.34 s was needed to complete the process, compared to RHBP-K-AD, which took
0.28 s to process the first band and an additional 41.94 s to process the whole data
set. This is mainly because R-AD and K-AD only require calculation of the global
sample correlation matrix R/global sample covariance matrix K once prior to
440 14 RHBP for Passive Target Detection: Anomaly Detection
shade anomaly
rhyolite
dry lake
b
1800
Shade
Rhyolite
1500 Dry lake
Cinders
Vegetation
Anomaly
1200
Radiance
900
600
300
0
0.4 mm 2.5 mm
Wavelength (mm)
processing. Once it is calculated, R/K will be used during the course of the entire
process. By contrast, RHBP-R-AD and RHBP-K-AD must recalculate the sample-
varying correlation matrix and covariance matrix via equations specified by (14.6)
and (14.15). Consequently, they both require more time than R-AD and K-AD.
Nevertheless, such a disadvantage can be offset by several benefits. One is the
progressive detection performance as bands are progressively processed. Second,
with such progressive performance, the impact of each band on AD performance
can be analyzed. Third, (14.6) and (14.15) provide a feasible implementation for
hardware design. Finally, RHBP-R-AD and RHBP-K-AD can be performed when-
ever bands are available, as opposed to R-AD and K-AD, which must wait for data
to be completed to calculate the global matrix R and K. Table 14.1 summarizes the
computing times in seconds required by K-AD, R-AD, RHBP-K-AD, and RHBP-
R-AD to run HYDICE and LCVF data sets.
14.5 Computing Time Analysis 441
a
Shade
Rhyolite
600 Dry lake
Cinders
Vegetation
Anomaly
R-AD Value
400
200
0
0 20 30 50 57 69 100 110 150
Number of Processed Bands
600
500
Value of R-AD
400
300
200
100
10
20
3035
Nu
m 57
be
ro 69
fB
an 85 Anomaly
ds
Pr 100 Vegetation
oc 110
es Cinders
se
d Dry lake
135
Rhyolite es
Shade 6 Signatur
160
Fig. 14.12 (a) Detected amounts of abundance fractions of six signatures by RHBP-R-AD versus
number of processed bands. (b) 3D plots of RHBP-R-AD-detected abundance fractions of six
signatures versus number of processed bands
Because the computing times for RHBP-K-AD and RHBP-R-AD were calcu-
lated by completing all bands, the times are significantly higher than their counter-
parts without RHBP. This is because RHBP must be repeatedly implemented for
each new incoming band with growing matrix sizes.
442 14 RHBP for Passive Target Detection: Anomaly Detection
a
Shade
Rhyolite
600 Dry lake
Cinders
Vegetation
Anomaly
K-AD Value
400
200
0
0 20 30 50 57 69 100 110 150
Number of Processed Bands
700
600
Value of K-AD
500
400
300
200
100
0
10
20
30
35
Nu
m 57
be
ro 69
fB
an 85 Anomaly
ds 100
Pr Vegetation
oc 110
es Cinders
se Dry lake
d 135
Rhyolite es
Shade 6 Signatur
160
Fig. 14.13 (a) Detected amounts of abundance fractions of six signatures by RHBP-K-AD versus
number of processed bands. (b) 3D plots of RHBP-R-AD-detected abundance fractions of six
signatures versus number of processed bands
14.5 Computing Time Analysis 443
Fig. 14.14 Results of RHBP-R-AD of LCVF data after receiving (a) 20 bands, (b) 22 bands, (c)
24 bands, (d) 26 bands, (e) 28 bands, (f) 30 bands, (g) 58 bands, (h) 60 bands, (i) 62 bands, (j)
64 bands, (k) 66 bands, (l) 68 bands, (m) 100 bands, (n) 102 bands, (o) 104 bands, (p) 106 bands,
(q) 108 bands, (r) 110 bands
444 14 RHBP for Passive Target Detection: Anomaly Detection
A graphical user interface (GUI) with a screenshot shown in Fig. 14.17 was
developed using MATLAB’s Guide to aid in algorithm performance analysis.
The GUI allows the user to load different data sources and choose different AD
methods (R-AD/K-AD) for analysis. The three images displayed at the tops of the
windows show a color image of the scene, a grayscale image of the current band
being processed, and the result of a detection map after the current band is
processed. At the bottom a window shows the changes in the R-AD (K-AD)
value in real time with detected abundance fractions highlighted by a pink line
with circles at the top as the number of processed bands increases. Once the data are
loaded, the user can start processing by clicking the start button. As a new band is
received at each iteration RHBP-AD (R-AD/K-AD) processes this newly received
data. Upon completion of processing, the current band and the resulting image are
updated to allow the user to observe the results. A final note on Fig. 14.17 is in
order. The plots shown in the bottom window represent a real-time progressive
detection version of Figs. 14.4 and 14.5/Figs. 14.12b and 14.13b. Unfortunately,
this attractive feature can only be demonstrated in a real-time process.
14.6 Graphical User Interface Design 445
Fig. 14.15 Results of RHBP-K-AD of LCVF data after receiving (a) 20 bands, (b) 22 bands, (c)
24 bands, (d) 26 bands, (e) 28 bands, (f) 30 bands, (g) 58 bands, (h) 60 bands, (i) 62 bands, (j)
64 bands, (k) 66 bands, (l) 68 bands, (m) 100 bands, (n) 102 bands, (o) 104 bands, (p) 106 bands,
(q) 108 bands, (r) 110 bands
446 14 RHBP for Passive Target Detection: Anomaly Detection
a b
0.4 16
14
Computational Time (Seconds)
0.3 12
10
0.2 8
0.1 4
RHBP-R-AD 2 RHBP-R-AD
RHBP-K-AD RHBP-K-AD
0 0
0 25 50 75 100 125 150 0 25 50 75 100 125 150
Number of Processed Bands Number of Processed Bands
Fig. 14.16 Plots of computing time versus number of processed bands for RHBP-R-AD and
RHBP-K-RD
Table 14.1 Summary of computing time (seconds) required by K-AD, R-AD, RHBP-K-AD, and
RHBP-R-AD to run HYDICE and LCVF data sets
R-AD K-AD RHBP-R-AD RHBP-K-AD
HYDICE 0.0528 0.0689 30.9614 41.9660
LCVF 0.3710 0.4552 1716.9 1854.4
14.7 Conclusions 447
14.7 Conclusions
15.1 Introduction
ATGP is an unsupervised target detection technique that has been widely used in
hyperspectral target detection (Ren and Chang 2003). It implements a succession of
orthogonal subspace projections (OSPs) to find targets that have maximal leaking
residuals from subspaces generated by previous targets. It is very simple but also
very effective. Recently, it has also been applied to finding endmembers and shown
to be effective because two popular endmember finding algorithms currently being
used in the literature, vector component analysis (VCA) (Nascimento and Bioucas-
Dias 2005) and simplex growing algorithm (SGA) (Chang 2006), can also be shown
to be equivalent to ATGP under certain assumptions (Chang et al. 2013, 2016b;
Greg 2010; Chap. 11 in Chang 2013).
This chapter takes a rather different approach by investigating ATGP in a very
interesting way. It considers ATGP to be performed band by band from a specific
data acquisition point of view. In general, there are two formats, band-interleaved-
pixel/sample/line (BIP/BIS/BIL) and band sequential (BSQ), which have been used
for the acquisition of remote sensing imagery (Schowengerdt 1997). While the
former has been widely used for real-time processing sample by sample (Du and
Ren 2003; Du and Nevovei 2005, 2009; Tarabalka et al. 2009; Chen et al. 2014), the
latter has not received much attention to date.
The concept of using the BSQ format to implement ATGP is novel and derived
from bit plane coding (Gonzalez and Woods, 2007). Suppose that the use of a band
is coded by “1” and “0” otherwise. By virtue of this interpretation, the total number
of bands to be used for data acquisition is the same as the total number of bits to be
used for image coding. The bit plane coding encodes an image bit by bit from the
most significant bit to the least significant bit, with image quality gradually
improved in a progressive manner until all bits are fully used for encoding. Such
an image processing is generally referred to as progressive image processing.
Following the same treatment we can consider implementing ATGP on an L-band
hyperspectral image band by band as the way the bit plane coding does for an L-bit
image bit by bit. In other words, if a band is used by ATGP, the band is specified by
“1” and “0” otherwise. As a result, all the advantages resulting from the bit plane
coding can also be applied to ATGP provided that ATGP can be implemented
progressively band by band. Inspired by this striking similarity, it is highly desir-
able to develop a progressive band processing version for ATGP that can be used
for progressive target detection. Such progressive hyperspectral band processing of
ATGP (PHBP-ATGP) is a result of this need. More specifically, if the total number
of bands is L, for each recursive hyperspectral band processing of ATGP (RHBP-
n oL
ðlÞ
ATGP)-detected target tp, we are able to find L targets tp migrated from band
l¼1
ðlÞ
1 to band L, where tp is the pth target detected by RHBP-ATGP using the first
n oL
ðlÞ ðLÞ ðlÞ
l bands, that is, tp ! tp ¼ tp as l ! L. The set of these targets, tp , provides
l¼1
ð1Þ
important band-to-band (interband) transition information from tp obtained by
15.2 ATGP 453
ðLÞ
using only the first band on tp obtained by using full bands, a benefit that cannot be
offered by ATGP using full bands, as originally proposed in Ren and Chang (2003).
Many advantages can be gained from PHBP-ATGP. First and foremost is that
PHBP-ATGP provides profiles of progressive band-varying target detection maps
that allow users to see changes in target detection maps by PHBP-ATGP in each
band. This benefit can be offered by a one-shot operation ATGP using all spectral
bands. Intuitively, we can view PHBP-ATGP as a slow motion version of ATGP in
the sense that slow varying changes in target detection maps can be observed and
dictated from band to band. Second, using progressive target detection map profiles
enables users to find weak targets that may be captured by certain bands and later be
overwhelmed and dominated by strong targets subsequently detected by ATGP.
Targets of this type, such as moving targets, are generally not shown in the final
ATGP detection map, that is, detection maps produced by ATGP using full bands.
Third, PHBP-ATGP can help identify significant bands for target detection without
performing band selection. This is a significant advantage since band selection
requires prior knowledge of how many bands need to be selected and as well
intelligent selection of appropriate bands. Finally, PHBP-ATGP can be used for
data communication and transmission by progressively processing data band by
band in the same manner that progressive image processing does for image trans-
mission and communication. However, in order for PHBP-ATGP to be
implemented in real time for data communication and transmission, a recursive
version of PHBP-ATGP, called recursive hyperspectral band processing of ATGP
(RHBP-ATGP), was further derived by Chang and Li (2016). It can be considered a
Kalman filter-like detector that can be carried out in a manner similar to that of a
Kalman filter by only updating innovation information provided by the new incom-
ing band via recursive equations band by band without reprocessing previous
bands. Most importantly, the computer processing time of RHBP-ATGP is inde-
pendent of the number of bands being processed. In other words, RHBP-ATGP
requires nearly linear time for each band compared to PHBP-ATGP whose com-
puting time is accumulated with the number of processed bands and results in
exponential growth in computation.
15.2 ATGP
The development of ATGP primarily stems from the need to find targets of interest
in image data when no prior knowledge is available. It utilizes a sequence of
orthogonal complement subspaces from which a succession of targets of interest
can be extracted by finding maximal orthogonal projections. Interestingly, as shown
in Greg (2010), Chang et al. (2013), and Chang (2013), most such ATGP-generated
target pixels turn out to be endmembers. This is certainly not a coincidence because
the concept behind ATGP is actually the same as that behind the pixel purity index
(PPI), except for two key differences. PPI requires a very large number of random
454 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
to find targets of interest directly from data without prior knowledge. Its step-by-
step implementation can be described as follows.
where Up1 ¼ t1 t2 tp1 is the target matrix generated at the ( p–1)st stage.
It should be pointed out that (15.2) can be carried out using the following
procedure:
N
(a) Assume that fri gi¼1 are the set of total data sample vectors. Let i ¼ 1 and
rmax
p r1 .
T 2
(b) If i > 1, calculate P⊥ P⊥ ⊥ and P⊥ ri >
Up1 ri Up1 ri ¼ PUp1 ri Up1
⊥ max
P r , then rmax ri , and go to step (c). Otherwise, continue.
Up1 p p
(c) If i < N, then go to step 2. Otherwise, continue.
(d) tp rmax
p , and go to step 3.
15.3 Recursive Equations for RHBP-ATGP 455
3. Stopping rule:
If tpT P⊥
Up1 tp > ε; ð15:3Þ
which is the same as (4.38), then augment Up ¼ Up1 tp ¼ t1 t2 tp by
adding tp to Up1 and go to step 2. Otherwise, continue.
4. At this stage, ATGP is terminated
and the final set ofproduced target
pixel vectors
comprises p target pixel vectors, t0 ; t1 ; t2 ; . . . ; tp1 ¼ ft0 g [ t1 ; t2 ; . . . ; tp1 .
It is worth noting that replacing tp in (15.3) with t0 yields
t0T P⊥
Up1 t0 ; ð15:4Þ
In order for ATGP to be processed band by band, a key issue is to implement step
2 of ATGP progressively band by band. First, let
2 3
t11 t21 tðp1Þ1 tp1
6 t12 t22 tðp1Þ2 tp2 7
6 7
Ulp ¼ t1 t2 tp ¼ 6 ⋮ ⋱ ⋱ 7 ¼ Uðl1 Þp
;
6 7 tT ðlÞ
4 t1ðl1Þ t2ðl1Þ tðp1Þðl1Þ tpðl1Þ 5
t1l t2l tðp1Þl tpl
ð15:5Þ
h i1 h i 1 1
Uðl1Þp
T
Ulp Ulp ¼ UðTl1Þp t ðlÞ ¼ UðTl1Þp Uðl1Þp þ tðlÞtT ðlÞ ;
t T ðlÞ
ð15:6Þ
2 3
t11 t21 tðp1Þ1 tp1
6 t12 t22 ⋮ tðp1Þ2 tp2 7
6 7
where Uðl1Þp ¼ 6 ⋮ ⋮ ⋱ ⋮ ⋮ 7 is a ðl 1Þ p
4 5
t1ðl1Þ
t2ðl1Þ tðp1Þðl1Þ t
p ðl1Þ
T
matrix and tðlÞ ¼ t1l ; t2l ; . . . ; tpl is a p-dimensional vector.
Applying Woodbury’s identity in Appendix A,
T 1 1 A1 u vT A1
A þ uv ¼A ; ð15:7Þ
1 þ vT A1 u
to (15.6) yields
456 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
h i1 1
T
Ulp Ulp ¼ UðTl1Þp Uðl1Þp þ tðlÞtT ðlÞ
h i1
¼ UðTl1Þp Uðl1Þp
1 1
UðTl1Þp Uðl1Þp tðlÞ tT ðlÞ UðTl1Þp Uðl1Þp
1 ;
1 þ tT ðlÞ UðTl1Þp Uðl1Þp tðlÞ
ð15:8Þ
P⊥ ¼ Ill Ulp U#lp
Ulp
0 1 1
B UðTl1Þp Uðl1Þp UðTl1Þp ρ C
l ðl1Þ
Uðl1Þp B h iC
¼ Ill B
t ð lÞ @
T 1 ρ ρ T
U T
ρ ρ T
tðlÞ CA
l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
1 þ tT ðlÞρ
l ðl1Þ
2 1 3
T T
ρ
Iðl1Þðl1Þ 0 6 U ð l1 Þp U ð l1 Þp U ðl1 Þp U ð l1 Þp U ð l1 Þp
l ðl1Þ 7
¼ 4 5
0 1 ρ T UðTl1Þp tT ðlÞρ
l ðl1Þ l ðl1Þ
h i
T
þ
1 Uðl1Þp ρ
l\ ðl1Þ ρ UðTl1Þp ρl\ ðl1Þ ρ T tðlÞ
1 þ t ðlÞρ
T t ðlÞ
T l ðl1 Þ l ð l1 Þ
l ðl1Þ
2 3
Uðl1Þp U#ðl1Þp Uðl1Þp ρ
Iðl1Þðl1Þ 0 l ðl1Þ
¼ 4 T 5
0 1 ρ UðTl1Þp tT ðlÞρ
l ðl1Þ l ðl1Þ
2 3
Uðl1Þp ρ ρ T UT Uðl1Þp ρ ρ T tðlÞ
1 l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
þ 4 5
1 þ tT ðlÞρ tT ðlÞρ ρ T UT tT ðlÞρ ρ T tðlÞ
l ðl1Þ l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
2 3
P⊥Uðl1Þp Uðl1Þp ρ
l ðl1Þ
6 7
¼4 1 T 5
Uðl1Þp Uðl1Þp Uðl1Þp
T
tðlÞ 1 t ðlÞρ
T
l ðl1Þ
2 3
Uðl1Þp ρ ρ T UðTl1Þp Uðl1Þp ρ ρ T tðlÞ
1 l ðl1Þ l ðl1Þ l ðl1Þ lðl1Þ
þ 4 5:
1 þ tT ðlÞρ tT ðlÞρ ρ T UT tT ðlÞρ ρ T tðlÞ
l ðl1Þ l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
ð15:9Þ
T
Let rl ¼ ðr 1 ; r 2 ; . . . ; r l1 ; r l ÞT ¼ rl1 ; r l . Then
T rl1
rlT P⊥
Ulp rl ¼ rl1 r l P⊥ Ulp rl
2 3
P⊥ Uðl1Þp Uðl1Þp ρ
l ðl1 Þ
T 6 T 7 rl1 1
¼ rl1 rl 4 5 þ T
T r l 1 þ t ð lÞρ
Uðl1Þp ρ
l ðl1Þ
1 t ðlÞρ
l ðl1Þ l ðl1Þ
2 T T T 3
T Uðl1Þp ρlðl1Þ ρlðl1Þ Uðl1Þp Uðl1Þp ρlðl1Þ ρlðl1Þ tðlÞ rl1
rl1 r l 4 T 5
t ðlÞρ ρ T UT tT ðlÞρ ρ T tðlÞ rl
l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
T
T T T rl1
¼ rl1 P⊥Uðl1Þp r l U ρ
ðl1Þp lðl1Þ r U
l1 ðl1Þp lðl1Þ ρ þ r l 1 t ð l Þρ
lðl1Þ rl
1
þ
1 þ tT ðlÞρ
l ðl1Þ
h T T
i ð15:10Þ
rl1 Uðl1Þp ρ ρ T UðTl1Þp þ r l tT ðlÞρ ρ T UðTl1Þp rl1 Uðl1Þp ρ ρ T tðlÞ þ rl tT ðlÞρ ρ T tðlÞ rl1
l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ
rl
T
T T T
¼ rl1 P⊥
Uðl1Þp rl1 r l Uðl1Þp ρlðl1Þ rl1 rl1 Uðl1Þp ρlðl1Þ r l þ r l 1 t ðlÞρlðl1Þ r l
1 T T T T T T
þ rl1 Uðl1Þp ρ ρ U rl1 þ rl t ðlÞρ ρ U rl1
1 þ tT ðlÞρ l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ ðl1Þp
l ðl1Þ
1 T T T T
þ r l1 U ðl1 Þp ρ ρ t ð l Þr l þ r l t ð l Þρ ρ t ðl Þr l
1 þ tT ðlÞρ lðl1Þ lðl1Þ lðl1Þ lðl1Þ
l ðl1Þ
T
T ⊥ T
¼ rl1 PUðl1Þp rl1 2rl Uðl1Þp ρ rl1 þ r l 1 t ðlÞρ rl
l ðl1Þ l ðl1Þ
1 T T T T T
þ rl1 Uðl1Þp ρ tðlÞrl U tðlÞr l ;
l ðl1Þ
þρ ρ rl1 þ ρ
1 þ tT ðlÞρ l ðl1Þ l ðl1Þ ðl1Þp l ðl1Þ
l ðl1Þ
458 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
1
where ρ ¼ UðTl1Þp Uðl1Þp tðlÞ is a p-dimensional vector.
l ðl1Þ
(1 )
(1 ) t2 (L) (L)
t1 (1 ) t1 t2
tj
x
(l )
tp (L) (L)
s1 tj
s2 tp
(1 )
tp
sp
ðlÞ
with tpðlÞ being the pth target generated by RHBP-ATGP using the first l bands, as
shown in Fig. 15.2, where the x- and y-axes denote the number of the first l bands
processed by nl, used to perform RHBP-ATGP, and sp is the pth target generated by
RHBP-ATGP.
Since the notation nl will be repeatedly used in the remaining discussions in this
chapter, a note on nl is necessary for clarification. It is defined as the number of
processed bands starting from the first and second bands and continuing on until the
lth band. Thus, both l and nl have dual interpretations. When nl is used, it indicates
that there are l bands being used to process RHBP-ATGP starting from the first
band and ending with the lth band, and the total number of these processed bands is
l, that is, nl ¼ l. On the other hand, when the notation of l is used, it can be an index
to specify the lth band or simply an integer number, l. Which interpretation is
applicable can be seen from the context without ambiguity.
ðlÞ
In general, the order of the target tpðlÞ specified by the subscript of p(l ) should be
a function of nl, which is used to produce the target. However, for simplicity and to
ðlÞ ðlÞ
avoid confusion, tpðlÞ is simplified as tp in Fig. 15.2 without including l as a
variable of p.
ðLÞ ð LÞ
Theoretically speaking, tpðLÞ ¼ tp ¼ tp , which is the pth target generated by
RHBP-ATGP using full bands, as it was originally designed in Ren and Chang
15.4 Algorithms for RHBP-ATGP 461
n op
ðLÞ
(2003). That is, RHBP-ATGP generates a set of p target signal vectors tj that
j¼1
can be arranged as a target vector, TV ¼ VRHBP-ATGP,
ðLÞ ðLÞ
TV ¼ VRHBP-ATGP ¼ t1 ; t2 ; ; tðpLÞ ; ð15:12Þ
The synthetic image simulated in Fig. 1.15 is shown in Fig. 15.3, with five panels in
each row simulated by the same mineral signature and five panels in each column
having the same size.
15.5 Synthetic Image Experiments 463
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
x 109
RBP-ATGP Detected Abundance Fractions
2 t (6)
1 = K(1, 1) t (6) = A(1, 1)
2
(15)
t 27 = M(1, 1)
1 t (28)
50 = B(1, 1)
0
6 0
S1 28 15
50
S2
S3 t (170) 100 ands
141 = C(1, 1) ed B
Sign
ature S4 Process
s
S5 150 ber of
170 Num
190
Fig. 15.4 3D plot of progressive detection maps of five signatures versus nl for TI
Among the 25 panels are five 4 4 pure pixel panels for each row in the first
column and five 2 2 pure pixel panels for each row in the second column, five
2 2 mixed pixel panels for each row in the third column, and five 1 1 subpixel
panels for each row in both the fourth and fifth columns, where the mixed and
subpanel pixels were simulated according to the legends in Fig. 15.2. Thus, a total
of 100 pure pixels (80 in the first column and 20 in the second column), referred to
as endmember pixels, were simulated in the data by the five endmembers A, B, C,
K, and M. The area marked “BKG” in the upper right corner of Fig. 1.14a was
selected to find its sample mean, that is, the average of all pixel vectors within the
area “BKG,” denoted by b and plotted in Fig. 1.14b, to be used to simulate the
464 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
Fig. 15.5 Progressive target detection maps by RHBP-ATGP using (a) nl ¼ 6, (b) nl ¼ 15, (c)
nl ¼ 28, (d) nl ¼ 170, (e) nl ¼ 189 (full 189 bands)
background (BKG) for the image scene with a size of 200 200 pixels in Fig. 15.2.
This b-simulated image background was further corrupted by an additive noise to
achieve a signal-to-noise ratio (SNR) of 20:1, which was defined as 50 % signature
(i.e., reflectance/radiance) divided by the standard deviation of the noise in Har-
sanyi and Chang (1994).
Once target pixels and background are simulated, two types of target insertion,
referred to as target implantation (TI) and target embeddedness (TE), can be
designed to simulate experiments for various applications. Since five mineral
signatures plus a background signature b are used to simulate the synthetic images,
there are six target signatures present in TI and TE.
15.5.1 TI Experiments
Figure 15.4 shows progressive detection profiles of these five mineral signatures, A,
B, C, K, and M, for TI, where the x-, y-, and z-axes denote extracted signatures, nl,
and RHBP-ATGP-detected abundance fractions, respectively. As shown in
Fig. 15.3, the signatures s1, s2, s3, s4, and s5 on the x-axis are the first five targets
found by RHBP-ATGP and the five mineral signatures A, B, C, K, and M are
represented by the panel pixels A(1,1), K(1,1), M(1,1), B(1,1), and C(1,1)
15.5 Synthetic Image Experiments 465
Table 15.2 Summary of minimal nl to identify signatures and their orders extracted in their
particular bands for TI
Minimal nl Order of signatures
Signatures found identifying found by RHBP- Order of signatures found using
by RHBP-ATGP signatures ATGP bands up to its particular band
ð6Þ 6 2 2
A ¼ t2
ð28Þ 28 50 3
B ¼ t50
ð170Þ 170 141 5
C ¼ t141
ð6Þ 6 1 1
K ¼ t1
ð15Þ 15 27 2
M ¼ t27
corresponding to the first panel pixel of the five 4 4 panels in five rows, each of
which is located in the upper left corner in each row in the first column in Fig. 15.3.
Figure 15.5 shows five RHBP-ATGP-extracted target maps using different
numbers of bands starting with the six bands in Fig. 15.5a progressively increasing
to 15 bands in Fig. 15.5b, 28 bands in Fig. 15.5c, 170 bands in Fig. 15.5d, and
finally reaching the full bands, 189 bands in Fig. 15.5e, each of which shows the
targets found in particular transition bands specified by Fig. 15.4.
More specifically, the bands below the figures in Fig. 15.5 were those actually
extracted panel pixels where the red crosses indicate ground truth pixels and
triangles highlight the targets found in the current band, with the numbers next to
the triangles indicating the order of the targets found. Also, the targets found in the
previous bands are indicated by magenta.
Table 15.1 summarizes 6 target pixels extracted by RHBP-ATGP, with nl
varying with 6 bands, 15 bands, 28 bands, 170 bands, and full 189 bands, to extract
panel pixels corresponding to five mineral signatures, A, B, C, K, and M. The panel
ðlÞ
pixels tj in the second column of the table were identified as the jth target among
the six target pixels found by RHBP-ATGP using the first l bands. The value of
j indicates the jth order of the target extracted by RHBP-ATGP when the particular
lth band was the new band added to process RHBP-ATGP.
According to Figs. 15.4 and 15.5 and Tables 15.1 and 15.2, all panel pixels found
by RHBP-ATGP correspond to the five mineral signatures using the minimal nl.
ðlÞ
The value of the subscript “j” in tj , listed in the third column, is the order in which
466 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
a b
6
2 C(1, 1)
5
5 M(1, 1)
3
1 1
0
0 25 50 75 100 125 150 175 200
Number of Processed Bands
Fig. 15.6 Six ATGP- and RHBP-ATGP-generated targets for TI. (a) Six ATGP-generated targets
for TI; (b) orders of panel pixels extracted by RHBP-ATGP
As noted, nl is defined as the number of the first l bands used to process RHBP-
ATGP, where these l bands start with the first band and end with the lth band, and
the subscript “l” is used as an index to track the lth band. Thus, if two different
15.5 Synthetic Image Experiments 467
ðlÞ ðmÞ
targets, tj and tk , are produced by RHBP-ATGP using nl and nm bands, then these
ðlÞ
targets will be ranked by nl and nm, respectively. If l < m, then tj is more
ðmÞ ðlÞ
significant than tk because tj can be extracted by RHBP-ATGP using a smaller
ðmÞ ðlÞ ðmÞ
nl than that used to extract tk . If both tj and tk are extracted with the same
ðlÞ ðmÞ ðlÞ
nl ¼ nm and j < k, then tj will be ranked higher than tk since tj is extracted ahead
ðlÞ ðlÞ
of tk . Therefore, the target sets, {tp }, found by RHBP-ATGP can be ranked by their
significance according to two indices, the superscript “l” and the subscript “p.” More
ðlÞ
specifically, the value of the superscript “l” in tp is first used to rank targets. If the
values of l are the same, the subscript “p” will then be used to prioritize their orders.
In other words, when two targets are found using the same “nl” bands, their
significance should be ranked by their extracted order specified by the subscript “p.”
For a given nl bands being used, RHBP-ATGP is implemented to find p targets.
ðlÞ ðlÞ
If two targets tj and tk with j < k are found by RHBP-ATGP using the same nl
ðlÞ ðmÞ ðlÞ
bands, it implies that tj is extracted prior to tk . This indicates that tj is more
ðmÞ
significant than tk . Thus, for example, according to Table 15.1 the two mineral
ð6Þ ð6Þ
signatures A and K were identified by Að1; 1Þ ¼ t2 ¼ Kð1; 1Þ ¼ t1 and both A
and K were extracted by RHBP-ATGP using the same nl ¼ 6 bands. Now if nl is
increased from 6 to 15, then the two mineral signatures K and M were identified by
ð15Þ ð15Þ
Kð1; 1Þ ¼ t1 , Mð1; 1Þ ¼ t2 , where K once again was picked up by RHBP-
ATGP using the first 15 bands while M was the first time to be found by RHBP-
ATGP as the second target. In this case, the minimal numbers of bands for A, K, and
M were 6, 6, and 15, respectively, but their extracted orders by RHBP-ATGP were
actually 1, 2, and 27, as tabulated in Table 15.2. Tables 15.1 and 15.2 show an
advantage of RHBP-ATGP over ATGP in that RHBP-ATGP keeps track of the
targets it detects and records these targets in its database.
15.5.3 TE Experiments
The same experiments conducted for TI were also performed for TE. Figure 15.7
shows the progressive detection profiles of these five mineral signatures, A, B, C, K,
and M, for TE, where the x-, y-, and z-axes denote extracted signatures, nl, and
RHBP-ATGP-detected abundance fractions, respectively.
Similar to Figs. 15.4, 15.7 also shows the five mineral signatures A, B, C, K, and
M extracted by the panel pixels, A(1,1), K(1,1), M(1,1), B(1,1), and C(1,1)
corresponding to the first panel pixel of the five 4 4 panels in the five rows,
each of which is located at the upper left corner in each row in the first column in
Fig. 15.3.
Figure 15.8 further shows five RHBP-ATGP-extracted target maps using differ-
ent sets of bands starting with nl ¼ 6 bands in Fig. 15.8a, progressively increasing to
468 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
x 109
RBP-ATGP Detected Abundance Fractions
12
10
6
(6)
t = K(1, 1)
1 (6)
4 t = A(1, 1)
2
(15)
t = M(1, 1)
24
2 t
(22)
= B(1, 1)
35
0
6 0
S1 28 15
(141)
50
S2 t = C(1, 1)
108 ds
S3 100 Ban
Sign ssed
ature
s
S4 141 fProce
S5 150 ber o
Num
190
Fig. 15.7 3D plot of progressive detection maps of five signatures versus nl for TE
Fig. 15.8 Progressive target detection maps by RHBP-ATGP with different numbers of processed
bands: (a) 6 bands, (b) 15 bands, (c) 22 bands, (d) 141 bands, (e) 189 bands (full bands)
15.5 Synthetic Image Experiments 469
Table 15.4 Summary of minimal nl to identify signatures and their orders extracted in their
particular bands for TE
Minimal nl Order of signatures
Signatures found identifying found by RHBP- Order of signatures found using
by RHBP-ATGP signatures ATGP bands up to its particular band
ð6Þ 6 2 2
A ¼ t2
ð22Þ 22 35 4
B ¼ t35
ð141Þ 141 108 5
C ¼ t108
ð6Þ 6 1 1
K ¼ t1
ð15Þ 15 24 2
M ¼ t24
15 bands in Fig. 15.8b, 22 bands in Fig. 15.8c, 141 bands in Fig. 15.8d, and
finally reaching the full 189 bands in Fig. 15.8e, where each detection map
shows the targets found in the transition bands in 4 ranges in Fig. 15.8b. Specifi-
cally, the bands below the figures were those actually extracting panel pixels
where the red crosses indicate ground truth pixels and triangles highlight the targets
found in the current band, with the numbers next to the triangles indicating the
order of the targets found. The targets found in the previous bands are also marked
by magenta.
According to Fig. 15.8, Table 15.3 summarizes 6 target pixels extracted by
RHBP-ATGP, when nl varies where the processed bands start from the first
6 bands, the first 15 bands, the first 28 bands, the first 170 bands, and 189 full
bands, to extract panel pixels corresponding to five panel signatures. The panel
ðlÞ
pixels tj in the second column of the table are identified as the jth target among
these six target pixels, with the first l bands being used to process RHBP-ATGP.
According to Fig. 15.8 and Tables 15.3, 15.4 tabulates all panel pixels found by
RHBP-ATGP corresponding to the five mineral signatures using the minimal nl,
ðlÞ
where the value of the subscript j in tj listed in the third column is its order in the
entire sequence of RHBP-ATGP-generated target pixels. The order of the five panel
pixels found by RHBP-ATGP is listed in the fourth column with the total number of
bands up to the particular band being used to process RHBP-ATGP. For example,
ð22Þ
panel pixel B(1,1) was found to be the fourth target in Fig. 15.8c, B ¼ t4 , when
nl ¼ 22 was used. But if we include the previous RHBP-ATGP-generated targets
470 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
a b
6
2 C(1, 1)
5
5 M(1, 1)
3
1 1
0
6 0 25 50 75 100 125 150 175 200
Number of Processed Bands
Fig. 15.9 Six ATGP and RHBP-ATGP generated targets for TE. (a) Six ATGP-generated targets
for TE; (b) orders of panel pixels extracted by RHBP-ATGP
found using the processed bands, nl < 22, panel pixel B(1,1) is actually extracted as
ð22Þ
the 35th target by RHBP-ATGP, that is, B ¼ t35 . Similarly, C(1,1) was picked up
as the 5th target by RHBP-ATGP in Fig. 15.8d when the minimal nl ¼ 141, but it
ð141Þ
was actually extracted as the 108th target, C ¼ t108 , by counting all previous
RHBP-ATGP-generated targets.
Figure 15.9a shows the first 6 distinct targets extracted by ATGP using all
189 bands from TE. According to the results in Fig. 15.7 and Table 15.2,
Fig. 15.9b shows how many bands are required to extract panel pixels
corresponding to the five mineral signatures A, B, C, K, and M, where the x- and
y-axes indicate nl used to process data and the order of the panel pixels A(1,1), K
(1,1), M(1,1), B(1,1), and C(1,1).
The image data to be studied here are from the HYperspectral Digital Imagery
Collection Experiment (HYDICE) image scene shown in Fig. 15.10a (and Fig. 1.
10a), which has a size of 64 64 pixel vectors with 15 panels in the scene and the
ground truth map in Fig. 15.10b (Fig. 1.10b). It was acquired by 210 spectral bands
with a spectral coverage from 0.4 to 2.5 μm. Low signal/high noise bands, bands
1–3 and 202–210, and water vapor absorption bands, bands 101–112 and 137–153,
were removed. Thus, a total of 169 bands were used in the experiments. The spatial
resolution and spectral resolution of this image scene are 1.56 m and 10 nm,
respectively.
15.6 Real Image Experiments 471
a b
p11, p12, p13
Fig. 15.10 (a) HYDICE panel scene containing 15 panels. (b) Ground truth map of spatial
locations of the 15 panels
It has been demonstrated that panel pixel p212, marked by yellow in Fig. 15.10b,
is of particular interest. It is always extracted as the one with the most spectrally
distinct signature compared to the R panel pixels in row 2. This indicates that a
signature of spectral purity is not equivalent to a signature of spectral distinction. In
fact, in many cases panel pixel p212 instead of panel pixel p221 is the first panel pixel
in row 2 extracted by endmember finding algorithms to represent the panel signa-
ture in row 2. Also, because of such ambiguity the panel signature representing
panel pixels in the second row is either p221 or p212, which is always difficult for
endmember finding algorithms to find. This implies that the ground truth of R panel
pixels in the second row in Fig. 15.10b may not be as pure as was thought. Thus, it
has a total number of 20 panel pixels of interest, 19 R panel pixels plus a yellow
panel pixel, p212.
According to some research (Chang et al. 2010a, 2011b), the number of targets
required for ATGP to find the five ground truth panel signatures in Fig. 1.11 was
shown to be 18. Figure 15.11 plots 3D detection maps from the 5th to the 18th
signatures on the x-axis versus nl on the y-axis from 1 to 169 and detected
abundance fractions specified by the z-axis since the first panel pixel extracted by
RHBP-ATGP was the 5th target pixel. The reason we started with the fifth signature
and not the first one is simply because the fifth and sixth signatures, p221 and p521,
were found to be the first two panel pixels representing two of the five panel
signatures p2 and p5 when the first six spectral bands, that is, l ¼ 6, were used to
process RHBP-ATGP.
To further track the targets found by RHBP-ATGP in the transition bands shown
in Fig. 15.11, these targets are also shown in Fig. 15.12, where the bands below the
figures were those bands actually picking up panel pixels. In Fig. 15.12, the red
crosses indicate the 19 R panel pixels and triangles highlight the targets found in the
current band, with the numbers next to the triangles indicating the order of the
targets being extracted by RHBP-ATGP. In addition, the targets found in the
previous bands are also marked by magenta.
According to Fig. 15.12, with 7 transition bands, Table 15.5 summarizes 18 tar-
get pixels extracted by RHBP-ATGP when nl varies, with 18 bands, 26 bands,
472 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
t(18) = p
5 521
RBP-ATGP Detected Abundance Fractions
t(27) = p
37 221
x 106
10 t(48) = p t(26) = p
55 311 35 212
8 0
t(46) = p 18
t(48) = p 53 312
26
56 11
6 27
46
4
s
48
d
an
B
2 t(97) = p
ed
98 411
ss
ce
0 97
ro
fP
ro
S5 S6
be
S7 S8
um
S9 S10
S11 S12
N
Target Si
gnatures S13 S14
Extracted S15 S16
by RBP-AT S17 S18 170
GP
Fig. 15.11 3D detection map plot of 5th to 18th signatures versus nl for HYDICE data
27 bands, 46 bands, 48 bands, 97 bands, and all 169 bands, to extract panel pixels
ðlÞ
corresponding to five panel signatures, where the panel pixels tj in the second
column of the table are identified as the jth target among these 18 target pixels, with
l bands being used to process RHBP-ATGP.
Comparing the results available in Chang (2013), two more panel pixels, p212
and p42, were found in Fig. 15.12f. These experiments demonstrated the advantages
of using RHBP-ATGP.
Furthermore, for comparative study, Fig. 15.13 also shows 18 targets produced
by ATGP using all 169 bands, with the number of targets to be generated, p ¼ 18,
determined by the work in Chang et al. (2010a, 2011b), where the red crosses
indicate the locations of the 19 R panel pixels and the open circle shows the targets
found by ATGP.
If we compare Fig. 15.13 to Fig. 15.12f–g, results using l ¼ 97 and 169 could
ð97Þ ð97Þ ð97Þ
extract 5 R panel pixels in 5 different rows, p11 ¼ t14 , p221 ¼ t16 , p312 ¼ t5 ,
ð97Þ ð97Þ ð169Þ ð169Þ
p411 ¼ t18 , and p521 ¼ t3 using the first 97 bands and p11 ¼ t6 , p212 ¼ t17 ,
ð169Þ ð169Þ ð169Þ
p312 ¼ t5 , p411 ¼ t18 , and p521 ¼ t3 using all 169 full bands. However, there
is a discrepancy between the 5 panel pixels generated by 97 bands and 169 bands in
ð169Þ
that the target pixel found by t17 is the yellow panel pixel p212, which is different
from the ground truth R panel pixel p221 found using 97 bands. In addition, the
extracted order of p11 was also different. If we further examine the panel pixels
found in Fig. 15.12 combined with Table 15.4, RHBP-ATGP actually found a total
of 7 panel pixels with 6 R panel pixels, p11, p221, p311, p312, p411, p521, and one
yellow panel pixel, p212, compared to only four R panel pixels, p11, p312, p411, p521,
plus a yellow panel pixel, p212, extracted in Fig. 15.13 by ATGP using all
Fig. 15.12 RHBP-ATGP progressive detection maps using various numbers of processed bands
for HYDICE data: (a) 18 bands, (b) 26 bands, (c) 27 bands, (d) 46 bands, (e) 48 bands, (f)
97 bands, (g) 169 bands
Table 15.5 Summary of panel pixels identified by RHBP-ATGP versus nl for HYDICE data
nl Panel pixels found by RHBP-ATGP
18 ð18Þ
p521 ¼ t5
26 ð26Þ ð26Þ
p212 ¼ t12 , p521 ¼ t3
27 ð27Þ ð27Þ
p221 ¼ t5 , p521 ¼ t3
46 ð46Þ ð46Þ ð46Þ
p221 ¼ t6 , p312 ¼ t17 , p521 ¼ t2
48 ð48Þ ð48Þ ð48Þ ð48Þ
p11 ¼ t16 , p221 ¼ t6 , p311 ¼ t8 , p521 ¼ t2
97 ð97Þ ð97Þ ð97Þ ð97Þ ð97Þ
p11 ¼ t14 , p221 ¼ t16 , p312 ¼ t5 , p411 ¼ t18 , p521 ¼ t3
169 ð169Þ ð169Þ ð169Þ ð169Þ ð169Þ
p11 ¼ t6 , p212 ¼ t17 , p312 ¼ t5 , p412 ¼ t18 , p521 ¼ t3
474 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
5 2
18
8
3
16
10 13 15
Table 15.6 Summary of minimal nl to identify signatures and their orders extracted in their
particular bands for HYDICE data
Minimal nl Order of signatures
Signatures found identifying found by RHBP- Order of signatures found using
by RHBP-ATGP signatures ATGP bands up to its particular band
ð48Þ 48 56 16
p11 ¼ t56
ð27Þ 27 37 5
p221 ¼ t37
ð26Þ 26 35 12
p212 ¼ t35
ð48Þ 48 55 8
p311 ¼ t55
ð46Þ 46 53 17
p312 ¼ t53
ð97Þ 97 98 18
p411 ¼ t98
ð18Þ 18 5 5
p521 ¼ t5
169 bands. This implies that RHBP-ATGP provides advantages over ATGP in
allowing users to find two more R panel pixels, p221 and p311.
Table 15.6 summarizes the results in Fig. 15.12 by tabulating the minimal nl that
first identified a signature and the order of this signature found in a particular band.
More specifically, all the 7 panel pixels found by RHBP-ATGP correspond to the
ðlÞ
five panel signatures using the minimal nl, where the value of the subscript j in tj
given in the third column is its extracted order in the entire sequence of RHBP-
ATGP-generated target pixels. The order of the 5 panel pixels to be found by
RHBP-ATGP is listed in the fourth column, with the total number of bands up to the
particular band were used to process RHBP-ATGP.
15.6 Real Image Experiments 475
RHBP-ATGP
5
p312
4
p221
3
p212
2
p521
1
0
0 25 50 75 100 125 150
Number of Processed Bands
of signatures found by
RHBP-ATGP versus nl for
HYDICE data
75
RBP-ATGP
50
25
0
0 25 50 75 100 125 150
Number of Processed Bands
According to Fig. 15.12 and Table 15.6, there are a total of 7 panel pixels,
including 6 R panel pixels, p11, p311, p312, p411, p42, and p521, and one yellow panel
pixel, p212, identified by RHBP-ATGP. Figure 15.14 further shows the order of
these 7 panel pixels found by RHBP-ATGP, indicated by the y-axis versus nl
indicated by the x-axis. For example, when nl ¼ 18, only one panel pixel, p521,
was found. Furthermore, when nl ¼ 26 and 27, two more panel pixels, p212 and p221,
were found, and then three panel pixels, p312, p11, and p311, were found before nl
reached 97. After 97 bands were processed, the last panel pixel, p411, was found.
Theoretically speaking, the target set found by PHBP-ATGP has the maximal
number of targets, p L, through band-by-band, progressive processing. That is,
for each different band, RHBP-ATGP can generate a different target. However,
476 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
200
p
Frequency
11
150
100
50 p
212
0 p
221 p
0 312
8
p
311
16
p p
411 511
24
Ro
32
w
In
de
40
x
64
48 56
48
40
56 32
24
16 Index
64 8 Column
0
a b
0.8 0.8
PBP-ATGP PBP-ATGP
RBP-ATGP RBP-ATGP
Computing Time (sec)
0.4 0.4
0.2 0.2
0 0
0 25 50 75 100 125 150 175 200 0 25 50 75 100 125 150 175 200
lth band lth band
c
0.2
PBP-ATGP
RBP-ATGP
Computing Time (sec)
0.15
0.1
0.05
0
0 25 50 75 100 125 150
lth band
Fig. 15.17 Computing time versus lth band by RHBP-ATGP for (a) TI, (b) TE, (c) HYDICE
A graphical user interface (GUI), with screenshot shown in Fig. 15.19, was devel-
oped to implement RHBP-ATGP using MATLAB for performance analysis. The
GUI allows users to load different data sources and choose different numbers of
targets for analysis. The three images displayed at the top of the windows show a
color image of the scene, a grayscale image of the band currently being processed,
and the result of subpixel detection after the current band is processed. At the
bottom, a window shows changes in the detected abundance fraction values in real
time with the cyan upward arrow at the top highlighting the results of the band
currently being processed. Once the data are loaded, users can start processing by
15.8 Graphical User Interface Design 479
a b
70 70
PBP-ATGP PBP-ATGP
60 RBP-ATGP 60 RBP-ATGP
50 50
Computing Time
Computing Time
40 40
30 30
20 20
10 10
0 0
0 25 50 75 100 125 150 175 200 0 25 50 75 100 125 150 175 200
Number of Processed Bands Number of Processed Bands
c
14
PBP-ATGP
12 RBP-ATGP
10
Computing Time
0
0 25 50 75 100 125 150
Number of Processed Bands
Fig. 15.18 Computing time versus nl by ATGP and RHBP-ATGP for (a) TI, (b) TE, (c) HYDICE
clicking the start button. When each band is received at each iteration, RHBP-
ATGP processes this newly received band to produce results with the current band
and the resulting image updated to allow users to observe the changes in target
detection by RHBP-ATGP.
480 15 Recursive Hyperspectral Band Processing of Automatic Target Generation Process
15.9 Conclusions
16.1 Introduction
Second, the progressive profiles in interband changes enable us to find bands that
are significant and crucial to mixed pixel classification without using band
selection. Finally and most importantly, PHBP-OSP provides missing details of
spectral band-to-band transition information that can be used to analyze interband
correlation in mixed pixel classification.
As PHBP-OSP is implemented, the computer processing time is increased and
accumulated by adding and processing new bands. Such computing time grows
exponentially by the number of processed bands and will become a major hurdle for
real-time processing. To resolve this dilemma, a Kalman filtering approach is used
to derive a recursive version of PHBP-OSP, called recursive hyperspectral band
processing of OSP (RHBP-OSP). The idea behind it is to update results only by new
information, referred to as innovation information, which is provided only by a new
incoming band but not in the processed information obtained from previous bands.
Such innovation information updating is done by recursive equations, and the time
of processing RHBP-OSP for each new band is nearly constant. As a result, RHBP-
OSP can implement not only progressively as bit plane coding does but also
recursively via recursive equations as a Kalman filter does. This benefit is very
significant when it comes to hyperspectral data communication and transmission as
the hyperspectral data must be processed on a timely basis.
Then OSP uses the signal-to-noise ratio (SNR) as a criterion for optimality to
derive
T ⊥
^ OSP
α p ðrÞ ¼ d PU r; ð16:2Þ
where
T 1 T
P⊥
U ¼ I UU ¼ I U U U
#
U ð16:3Þ
1
and U# is the pseudo-inverse of U given by UT U UT . It was shown in Tu et al.
(1997) and Chang et al. (1998) that OSP can also be derived as a least-squares OSP
(LSOSP) for αp, denoted by α^ pLSOSP ðrÞ, as
486 16 Recursive Hyperspectral Band Processing of Orthogonal Subspace Projection
1 OSP
^ pLSOSP ðrÞ ¼ dT P⊥
α Ud ^ p ðrÞ;
α ð16:4Þ
which turns out to be exactly the same as the pth abundance fraction of αp estimated
^ LS
by α p ðrÞ, as specified by
1
^ LS ¼ MT M MT r
α ð16:5Þ
^ LS ¼ α
where α ^ LS ^ LS
1 , ...,α p . Interestingly, it was shown in Chang (2007) that
^ LS
α ^ jLSOSP ðrÞ for all 1 j p. In other words, α
j ðrÞ ¼ α ^ LS
p ðrÞ and α^ pLSOSP ðrÞ can
T ⊥
1
be simply obtained by α ^ OSP
p ðrÞ with a scaled constant d PU d . Nonetheless,
^ OSP
α p ðrÞ has an advantage over α ^ LS ðrÞ in that α
^ OSP
p ðrÞ can be used to generate new
signal sources in an unsupervised manner, a task for which the LS approach
specified by (16.2) has difficulty with because it requires the prior knowledge of
all signatures for abundance estimation. One such algorithm is the automatic target
generation process (ATGP) developed by Ren and Chang (2003), which will be
used in our experiments to generate unsupervised target signal sources to form M in
(16.1) for OSP.
where
16.3 Recursive Equations for RHBP-OSP 487
2 3
m11 m21 mðp1Þ1 mp1
6 m12 m22 ⋮ mðp1Þ2 mp2 7
6 7
Uðl1Þp ¼ 6 ⋮ ⋮ ⋱ ⋮ ⋮ 7 is an ðl 1Þ p
4 5
m1ðl1Þ m2ðl1Þ mðp1Þðl1Þ m
p ðl1Þ
T
matrix and mðlÞ ¼ m1l ; m2l ; . . . ; mpl is a p-dimensional vector. Then applying
Woodbury’s identity in Appendix A,
T 1 1 A1 u vT A1
A þ uv ¼A ; ð16:8Þ
1 þ vT A1 u
h i1 h i
¼ UðTl1Þp Uðl1Þp UðTl1Þp mðlÞ
8 1 1 9
>
> T
mðlÞ m ðlÞ Uðl1Þp Uðl1Þp
T T >
>
< Uðl1Þp Uðl1Þp =h i
1 UðTl1Þp mðlÞ
>
> >
>
: 1 þ mT ðlÞ UðTl1Þp Uðl1Þp mðlÞ ;
1 1
¼ UðTl1Þp Uðl1Þp UðTl1Þp UðTl1Þp Uðl1Þp mðlÞ
h i
1 ρ ρ T UT ρ ρ T mðlÞ ;
l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
1 þ mT ðlÞρ
l ðl1Þ
ð16:10Þ
1
where ρ ¼ UðTl1Þp Uðl1Þp mðlÞ is a p-dimensional vector.
l ðl1Þ
488 16 Recursive Hyperspectral Band Processing of Orthogonal Subspace Projection
P⊥
Ulp ¼ Ill Ulp Ulp
#
0 1 1
" #B UðTl1Þp Uðl1Þp UðTl1Þp ρ C
l ðl1Þ
Uðl1Þp B C
¼ Ill B h i C
m ðlÞ @ B 1 ρ ρ T T
ρ ρ T
mðlÞ C
T U
l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
A
1 þ m ðlÞρ
T
l ðl1Þ
2 3
" # Uðl1Þp U#ðl1Þp Uðl1Þp ρ
Iðl1Þðl1Þ 0 l ðl1Þ
6 7
¼ 4 T 5
0 1 ρ UðTl1Þp mT ðlÞρ
l ðl1Þ l ðl1Þ
" #
Uðl1Þp h i
þ
1 ρl\ ðl1Þ ρ T UðTl1Þp ρl\ ðl1Þ ρ T mðlÞ
1 þ m ðlÞρ l ðl1Þ l ðl1Þ
T
l ðl1Þ m ðlÞ
T
2 3
" # Uðl1Þp U#ðl1Þp Uðl1Þp ρ
Iðl1Þðl1Þ 0 l ðl1Þ
6 7
¼ 4 T 5
0 1 ρ UðTl1Þp mT ðlÞρ
l ðl1Þ l ðl1Þ
2 3
Uðl1Þp ρ ρ T UT Uðl1Þp ρ ρ T mðlÞ
l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
1 6 7
þ 4 T 5
1 þ mT ðlÞρ ð Þρ ρ T T T
ð Þρ ρ T
ð Þ
lðl1Þ lðl1Þ ðl1Þp lðl1Þ lðl1Þ
l ðl1Þ m l U m l m l
2 3
P⊥ Uðl1Þp Uðl1Þp ρ
l ðl1Þ
6 7
¼64
1 T 7
5
Uðl1Þp UðTl1Þp Uðl1Þp mðlÞ 1 mT ðlÞρ
l ðl1Þ
2 3
Uðl1Þp ρ ρ T UT Uðl1Þp ρ ρ T mðlÞ
l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
1 6 7
þ 4 5;
1 þ mT ðlÞρ mT ðlÞρ ρ T UðTl1Þp mT ðlÞρ ρ T mðlÞ
l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ
ð16:11Þ
" #
-OSP ðr Þ ¼ d T P⊥ r ¼ d T rl1
α
^ dRHBP
l l l Ulp l l1 d l P⊥
Ulp
rl
2 3
P⊥
Uðl1Þp Uðl1Þp ρ " #
l ðl1Þ
6 7 rl1
¼ dl1
T
dl 6
4
T 7
5 r
Uðl1Þp ρ 1 mT ðlÞρ l
l ðl1Þ l ðl1Þ
2 3
Uðl1Þp ρ ρ T UT Uðl1Þp ρ ρ T mðlÞ " #
6 l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
1 7 rl1
þ T
dl1 dl 4 5
1 þ mT ðlÞρ mT ðlÞρ ρ T UT mT ðlÞρ ρ T mðlÞ rl
l ðl1Þ l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ
T
¼ dl1 P⊥ Uðl1Þp ρ rl þ d l 1 mT ðlÞρ
Uðl1Þp rl1 d l Uðl1Þp ρ rl1 dl1
T T
rl
l ðl1Þ l ðl1Þ l ðl1Þ
1
þ T
dl1 Uðl1Þp ρ ρ T UT rl1 þ d l mT ðlÞρ ρ T UT rl1
1 þ mT ðlÞρ l ðl1Þ lðl1Þ ðl1Þp l ðl1Þ lðl1Þ ðl1Þp
l ðl1Þ
1
þ T
dl1 Uðl1Þp ρ ρ T mðlÞr l þ dl mT ðlÞρ ρ T mðlÞr l
1 þ mT ðlÞρ l ðl1Þ lðl1Þ l ðl1Þ lðl1Þ
l ðl1Þ
T
^ dRHBP
¼α -OSP ðr Þ d U Uðl1Þp ρ rl þ d l 1 mT ðlÞρ
ðl1Þp ρ rl1 dl1
T
l1 l rl
l1 l ðl1Þ l ðl1Þ l ðl1Þ
1
þ T
dl1 Uðl1Þp ρ þ dl mT ðlÞρ ρ T UðTl1Þp rl1 þ ρ T mðlÞr l ,
1þ mT ðlÞρ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ
l ðl1Þ
ð16:12Þ
16.4 Recursive Hyperspectral Band Processing of OSP 489
ðrl1 Þ ¼ dl1
T
P⊥
Uðl1Þp rl1 using the first (l1) bands.
T
2. New information: the lth band, mðlÞ ¼ m1l ; m2l ; . . . ; mpl , dl and rl.
3. Innovation information, which is correlation information provided by Uðl1Þp
1
and m(l ): ρ ¼ UðTl1Þp Uðl1Þp mðlÞ .
l ðl1Þ
By virtue of (16.11), P⊥ ⊥ ⊥
U ¼ PUlp used in (16.3) can be updated by PUðl1Þp without
recalculating P⊥ T ⊥
Ulp , and the dl PUlp rl can also be updated by (16.12). Thus, using
(16.11) and (16.12), a recursive hyperspectral band processing of OSP (RHBP-
OSP) can be described as follows.
ðrl Þ is generated by simply replacing rl and rl1 with dl and dl1 in (16.12).
Figure 16.1 depicts a flow chart of RHBP-OSP.
As a concluding remark, we would like also point out that what Recursive
Hyperspectral Sample Processing of OSP (RHSP-OSP) in Chap. 8 is to RHBP-
OSP developed in this chapter is exactly what Recursive Hyperspectral Sample
Processing of ATGP (RHSP-ATGP) in Chap. 7 is to Recursive Hyperspectral Band
Processing of ATGP (RHBP-ATGP) in Chap. 15. That is, RHSP-OSP is designed to
490 16 Recursive Hyperspectral Band Processing of Orthogonal Subspace Projection
l=p
l l 1
Recursive
Band Recursive Process
Process #
Update U l U l 1t p , U l p and PU .
l p
Progressive process
RHBP - OSP RHBP - OSP
Update ˆ d (rl ) by ˆ d (rl 1)
l l 1
No
l =L
Yes
p p 1
Yes p>L No
Stop
The synthetic image simulated in Fig. 1.15 is shown in Fig. 16.2 with five panels in
each row simulated by the same mineral signature and five panels in each column
having the same size. Among 25 panels are five 4 4 pure-pixel panels for each row
in the first column, five 2 2 pure pixel panels for each row in the second column,
five 2 2 mixed pixel panels for each row in the third column, and five 1 1 subpixel
16.5 Synthetic Image Experiments 491
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
panels for each row in both the fourth and fifth columns, where the mixed and
subpanel pixels were simulated according to the legends in Fig. 16.2. Thus, a total
of 100 pure pixels (80 in the first column and 20 in the second column), referred to as
endmember pixels, were simulated in the data by the five endmembers A, B, C, K,
and M. An area marked “BKG” in the upper right corner of Fig. 1.14a was selected to
find its sample mean, that is, the average of all pixel vectors within the “BKG” area,
denoted by b and also plotted in Fig. 1.14b, to be used for simulating the background
(BKG) for the image scene with a size of 200 200 pixels in Fig. 16.2. This
b-simulated image background was further corrupted by an additive noise to achieve
a SNR of 20:1, which was defined in Harsanyi and Chang (1994)as a 50 % signature
(i.e., reflectance/radiance) divided by the standard deviation of the noise.
Once target pixels and background are simulated, two types of target insertion,
referred to as target implantation (TI) and target embeddedness (TE), can be
designed to simulate experiments for various applications. Since five mineral
signatures plus a background signature b are used to simulate the synthetic images,
there are six target signatures present in TI and TE.
16.5.1 TI Experiments
Let Ai, Bi, Ci, Ki, Mi for 1 i 4 denote the four panel pixels in the first row of
each of the five 4 4 pure panels in the first column in Fig. 16.2. In addition, let A5,
B5, C5, K5, and M5 also denote the first 2 2 panel pixel in the upper left corner in
the five panels in the second column in Fig. 16.2, respectively. Then Fig. 16.3
shows a 3D RHBP-OSP detection map of 25 pure panel pixels with each set of five
panel pixels Ai, Bi, Ci, Ki, Mi for 1 i 5 representing each of five mineral
signatures, A, B, C, K, M.
492 16 Recursive Hyperspectral Band Processing of Orthogonal Subspace Projection
x 106
Abundance
10
5
0
0
20
40
60
80
100
120
140 M4 M5
M1 M2 M3
160 K3 K4 K5
C3 C4 C5 K1 K2
180 B5 C1 C2
B2 B3 B4
A4 A5 B1
A1 A2 A3
To make discussions clear, we will use the notation nl defined in Chap. 15, where
nl is defined as the number of l bands being used to process OSP starting from the
first band and ending with the lth band. In other words, nl has a dual interpretation.
It not only provides information about how many bands are to be used for
processing OSP but also specifies the processed bands starting from the first band
and ending with the lth band. Similarly, when the notation of l is used, it can either
be used as an index to specify the lth band or simply be an integer number, l. Which
interpretation is applicable can be easily seen from the context.
Visual inspection of Fig. 16.3 reveals that the most distinct mineral signature
is M, followed by A, then K and B, with the least distinguishable mineral signature
being C. This unique progressive feature is a significant advantage of PHBP-OSP
not offered by OSP, which performs a one-shot operation using all the 189 bands as
shown in the last row corresponding nl ¼ 189 in Fig. 16.3. Figure 16.4 further shows
the detected abundance fractions of these 25 panel pixels with the 5 highest detected
abundance fractions highlighted on the plots, which are M1, M2, A1, M3, and M4.
Similarly, Fig. 16.5 shows a 3D RHBP-LSOSP estimation map of 25 pure panel
pixels with each set of 5 panel pixels Ai, Bi, Ci, Ki, and Mi for 1 i 5
representing each of 5 mineral signatures, A, B, C, K, and M.
It should be noted that, unlike Fig. 16.3, which shows the detected abundance
fractions of 25 panel pixels, the plots in Fig. 16.5 are actually abundance fractions
of the 25 panel pixels estimated by LSOSP. Since LSOSP is not fully abundance-
constrained, the peaks of 25 panel pixels in Fig. 16.5 that are supposed to be close to
1 may not be close to 1, as shown for the other panel pixels in Fig. 16.5. To see this,
Fig. 16.6 additionally shows the estimated abundance fractions of these 25 panel
pixels, where the abundance fractions of the 25 panel pixels were grouped into three
categories. One includes the 10 panel pixels A1, A2, B1, B2, C1, C2, K1, K2, M1, and
M2, with the correct estimated abundance fractions being one. Another is made up
of A3, A4, B3, B4, C3, C4, K3, K4, M3, and M4, which had estimated abundance
fractions of around 0.5, and the remaining panel pixels, A5, B5, C5, K5, and M5,
form the third category and had estimated abundance fractions of around 0.25.
16.5 Synthetic Image Experiments 493
x 106
A
2
M M2
1 K1
10 K
2
M
9 5
A3
8 A4
B1
7 B
2
Abundance
K4
6 M M4 K3
3
A A5
1
5
B3
B
4 4
K5
3 B5
C1
2 C
2
C
3
1 C4
C
0 5
Fig. 16.4 Detection of 25 panel pixels by RHBP-OSP versus nl for TI with top five ranked panels
indicated on plots
Abundance
1
0.5
0
0
20
40
Nu
60
m
be
80
ro
fp
100
ro
ce
120
ss
M4 M5
e
140
M1 M2 M3
d
K3 K4 K5
ba
160 C5 K1 K2
nd
C3 C4
B5 C1 C2
s
180 B2 B3 B4
A4 A5 B1
A1 A2 A3
Interestingly, these three groups also correspond to the first two panel pixels, Ai, Bi,
Ci, Ki, and Mi for 1 i 2, the next two panel pixels, Ai, Bi, Ci, Ki, and Mi for
3 i 4, and the last panel pixels, Ai, Bi, Ci, Ki, and Mi for i ¼ 5, respectively.
Furthermore, if we compare Fig. 16.3 to Fig. 16.5 we immediately notice that
there are significant differences in the results produced by RHBP-OSP and RHBP-
LSOSP due to the fact that OSP is performed as a detector and LSOSP is indeed an
estimator. More specifically, RHBP-OSP attempted to detect most distinct signa-
tures with the highest possible abundance fractions in Fig. 16.3, which are M1, M2,
494 16 Recursive Hyperspectral Band Processing of Orthogonal Subspace Projection
1.2
M1M2K1 K2 A1 A2 B1 B2 C1 C2
1
0.8
Abundance
0.6 C3 C4 B3 B4 A3 A4 M4M3K4 K3
0.4
C5 B5 A5 M5K5
0.2
-0.2
06 20 40 60 80 100 120 140 160 180 200
Number of processed bands
x 106
Abundance
10
5
0
0
20
40
Nu
60
m
be
80
ro
fp
100
ro
ce
120
ss
M3 M4 M5
ed
140
K5 M1 M2
K2 K3 K4
ba
160 C5 K1
nd
B5 C1 C2 C3 C4
s
180 B2 B3 B4
A4 A5 B1
A1 A2 A3
A1, M3, and M4, as opposed to LSOSP, which tried to accurately estimate abun-
dance fractions within the range between 0 and 1 in Fig. 16.5, where all 10 panel
pixels A1, A2, B1, B2, C1, C2, K1, K2, M1, and M2 were estimated accurately in
Fig. 16.6 by LSOSP with abundance fractions equal to 1 as pure pixels.
16.5.2 TE Experiments
The same experiments conducted for TI were also performed for TE. Figure 16.7
shows a 3D RHBP-OSP detection map of 25 pure panel pixels with each set of
5 panel pixels Ai, Bi, Ci, Ki, and Mi for 1 i 5 representing each of the 5 mineral
16.5 Synthetic Image Experiments 495
x 106
A
M1 M2 2
K1
10 K
2
M
9 A3
5
8 A4
B1
7 B
Abundance
2
K4
6 M M4
3
K3
A1 A5
5 B3
B
4 K5
4
3 B
5
C1
2 C
2
C
3
1 C4
C
0 5
1
0.5
0
0
20
40
Nu
60
m
be
80
ro
fp
100
ro
ce
120
ss
M3 M4 M5
ed
140
K5 M1 M2
K2 K3 K4
ba
160 C5 K1
nd
B5 C1 C2 C3 C4
s
180 B2 B3 B4
A4 A5 B1
A1 A2 A3
signatures, A, B, C, K, and M, along with Fig. 16.8, which shows the detected
abundance fractios of these 25 panel pixels with the 5 highest detected abundance
fractions highlighted on the plots, which are M1, M2, A1, M3, and M4.
Comparing Fig. 16.3 to Fig. 16.7 and Fig. 16.4 to Fig. 16.8, the results obtained
for TE were very close to those obtained for TI.
Analogous to Figs. 16.5 and 16.6, Figs. 16.9 and 16.10 also show a 3D RHBP-
LSOSP estimation map of 25 pure panel pixels with each set of 5 panel pixels Ai,
Bi, Ci, Ki, and Mi for 1 i 5 representing each of the 5 mineral signatures, A,
B, C, K, and M, along with their estimated abundance fractions. Once again, the
results obtained for TE were very close to those obtained for TI.
496 16 Recursive Hyperspectral Band Processing of Orthogonal Subspace Projection
1.5
Abundance
C1 C2 B 1 B 2 A 1 A 2 M1M2K 1 K 2
1
C3 C4 B 3 B 4 A 3 A 4 M4M3K 4 K 3
0.5
C5 B 5 A 5 M5K 5
The image data to be studied are the HYperspectral Digital Imagery Collection
Experiment (HYDICE) image scene shown in Fig. 16.11a (and Fig. 1.10a), which
has a size of 64 64 pixel vectors with 15 panels in the scene and the ground truth
map in Fig. 16.11b (Fig. 1.10b). It was acquired by 210 spectral bands with a
spectral coverage from 0.4 to 2.5 μm. Low signal/high noise bands, bands 1–3 and
202–210, and water vapor absorption bands, bands 101–112 and 137–153, were
removed. Thus, a total of 169 bands were used in the experiments. The spatial
resolution and spectral resolution of this image scene are 1.56 m and 10 nm,
respectively. It is worth noting that panel pixel p212, marked by yellow in
Fig. 16.11b, is of particular interest. It is a yellow panel pixel that is always
extracted as being the most spectrally distinct signature compared to the R panel
pixels in row 2. This implies that a signature of spectral purity is not equivalent to a
signature of spectral distinction. Also, because of such ambiguity, the panel signa-
ture representing panel pixels in the second row is either p221 or p212, which is
always the last one found by endmember finding algorithms. This implies that the
ground truth of the R panel pixels in the second row in Fig. 16.11b may not be as
pure as was thought.
According to Chang (2003) and Chang and Du (2004), the virtual dimensionality
(VD) estimated for this HYDICE scene was up to nVD ¼ 18. In this case, we used
ATGP to generate 18 targets, shown in Fig. 16.12, where panel pixels
corresponding to the five panel signatures were extracted, with p11 as the 6th target
t6, p212 as the 17th target t17, p312 as the 5th target t5, p411 as the 18th target t18, and
p521 as the 3rd target t3.
16.6 Real Image Experiments 497
a b
p11, p12, p13
Fig. 16.11 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial
locations of the 15 panels
Figures 16.13 and 16.14 show that RHBP-OSP and RHBP-LSOSP used the
18 target pixels extracted by ATGP in Fig. 16.12, where the z-axis in Figs. 16.13a
and 16.14a shows the progressive detection of changes in abundance fractions of
19 R panel pixels produced by RHBP-OSP and RHBP-LSOSP, respectively, and
Figs. 16.13b and 16.14b are their corresponding 3D plots of real-time progressive
profiles of abundance fractions of 19 R panel pixels detected by RHBP-OSP and
estimated by RHBP-LSOSP, respectively.
As we can see from Figs. 16.13a, b and 16.14a, b, the abundance fractions of all
panel pixels saturated and stayed nearly the same after n ¼ 118 bands were
processed. This implied that the remaining 51 bands provided no additional infor-
mation to improve the detection of desired targets. Figure 16.13a, b also clearly
shows that RHBP-OSP is a real-time progressive signal detection technique that
enhances detected signals in Fig. 16.13b, compared to RHBP-LSOSP, which is
indeed a real-time progressive signal estimation technique to estimate signal
498 16 Recursive Hyperspectral Band Processing of Orthogonal Subspace Projection
a
5
10 x 10 p312
p311
9 p32
p11
8 p33
p521
7 p12
p52
6
Abundance
p511
p212
5 p411
p221
4 p211
p42
3 p412
p22
2 p23
p43
1 p13
p53
0
0 9 18 29 40 60 80 100 120 140 160 180
Number of processed bands
b 5
x 10
10
Abundance
5
0
0
20
40
Nu
60
m
be
ro
80
fp
ro
ce
100
ss
ed
120
p52 p53
ba
p521
p43 p511
nd
140 p412p42
s
p33 p411
p312p32
160 p23 p311
p212 p221p22
p13 p211
p11 p12
Fig. 16.13 Detection abundance fractions of 19 R panel pixels plus panel pixel p212 by RHBP-
OSP. (a) RHBP-OSP-detected abundance fractions of 19 R panel pixels plus p212. (b) 3D plot of
progressive profile of 19 R panel pixels plus panel pixel p212 by RHBP-OSP
abundances accurately, where the peaks shown in Fig. 16.14b, highlighted by open
circles, represent significant changes in the abundance fractions estimated by
RHBP-LSOSP for each of the 19 R panel pixels in earlier band numbers less than
60, and Fig. 16.14a plots the abundances of 19 R panel pixels starting from nl from
the first 9 bands up to the total number of bands, 169.
These experiments demonstrate the advantages of RHBP-OSP and RHBP-
LSOSP, which are able to provide progressive changes in abundance fractions
band by band during real-time band processing.
16.7 Computing Time Analysis 499
a p11
5
p312
p521
4
p212
3 p411
p311
2 p221
p42
1 p412
Abundance
p211
0 p52
p511
-1 p22
p32
-2 p12
p23
-3 p33
p43
-4 p13
p53
-5
0 9 18 29 40 60 80 100 120 140 160 180
Number of processed bands
b
Abundance
2
0
20
40
Nu
60
m
be
ro
80
fp
ro
ce
100
ss
ed
120
p52 p53
ba
p521
p43 p511
nd
140 p412p42
s
Fig. 16.14 Detection abundance fractions of 19 R panel pixels plus p212 by RHBP-LSOSP versus
nl. (a) RHBP-LSOSP-estimated abundance fractions of 19 R panel pixels plus panel pixel p212. (b)
3D plot of progressive profile of 19 R panel pixels plus panel pixel p212 by RHBP-LSOSP
a b
0.06 0.06
PHBP-OSP PHBP-OSP
0.05 PHBP-LSOSP 0.05 PHBP-LSOSP
RHBP-OSP RHBP-OSP
0.04 RHBP-LSOSP 0.04 RHBP-LSOSP
Time (Sec)
Time (Sec)
0.03 0.03
0.02 0.02
0.01 0.01
0 0
06 20 40 60 80 100 120 140 160 180 200 06 20 40 60 80 100 120 140 160 180 200
Ith band Ith band
c
0.02
0.018 PHBP-OSP
0.016 PHBP-LSOSP
RHBP-OSP
0.014
RHBP-LSOSP
Time (Sec)
0.012
0.01
0.008
0.006
0.004
0.002
0
0 18 40 60 80 100 120 140 160 180
Ith band
Fig. 16.15 Computing time versus lth band by PHBP-OSP/LSOSP, RHBP-OSP/LSOSP for (a) TI,
(b) TE, (c) HYDICE scene
(16.12) to update the results when the lth band is a new incoming band. The
computer environment was as follows: operating system, Windows 7 64 bits;
CPU: Intel Core i7-3770 @ 3.40 GHz; memory: 16.0 GB; programming language:
MATLAB v8.0.
As can be seen, the computing time required for PHBP-OSP/LSOSP and RHBP-
OSP/LSOSP to run each individual band is nearly linear for TI and TE but shows
slow varying exponential curves for the HYDICE scene. Figure 16.16 further plots
the computing time required for PHBP-OSP/LSOSP without using recursive equa-
tions (16.11) and (16.12) to run using the first l bands where the x-axis is the running
time using the number of the first l bands, nl.
For a fair comparative study, we also include the cumulative computer
processing time of RHBP-OSP/LSOSP, which is obtained by summing up all the
computing times in Fig. 16.15 up to the lth band.
As shown in Fig. 16.16, the computing time increases as new bands are added.
This is due to an increase in the size of P⊥ U , which results in higher computational
16.7 Computing Time Analysis 501
a b
6 6
PHBP-OSP PHBP-OSP
Accumulative Time (Sec)
3 3
2 2
1 1
0 0
06 20 40 60 80 100 120 140 160 180 200 06 20 40 60 80 100 120 140 160 180 200
Number of processed bands Number of processed bands
c
1.5
PHBP-OSP
Accumulative Time (Sec)
PHBP-LSOSP
RHBP-OSP
1 RHBP-LSOSP
0.5
0
0 18 40 60 80 100 120 140 160 180
Number of processed bands
Fig. 16.16 Computing time versus nl by RHBP-OSP/LSOSP, RHBP-OSP/LSOSP for (a) TI,
(b) TE, (c) HYDICE
Table 16.1 Comparison of computing time in seconds required by ATGP and RHBP-ATGP
Method
Data set OSP PHBP-OSP RHBP-OSP LSOSP PHBP-LSOSP RHBP-LSOSP
TI 0.0286 5.4131 2.6752 0.0294 5.5565 2.7541
TE 0.0288 5.4365 2.7687 0.0296 5.5925 2.8565
HYDICE 0.0074 1.2583 0.7311 0. 0083 1.4040 0.7921
complexity. Table 16.1 calculates the final computer times required for RHBP-
OSP/LSOSP and PHBP-OSP/LSOSP to complete the entire data sets, where
RHBP-OSP/LSOSP achieved times twice as fast as PHBP-OSP/LSOSP. However,
we believe that this saving could be even greater once RHBP-OSP/LSOSP is
implemented on hardware thanks to its recursive structure.
502 16 Recursive Hyperspectral Band Processing of Orthogonal Subspace Projection
16.9 Conclusions
^ OSP
progressive band-varying detected abundance fractions of α dl ðrl Þ and RHBP-
h i
ðlÞ
ATGP produces a target matrix TM ¼ tp in (15.11). Second, RHBP-OSP/
LL
LSOSP needs to fix the number of signatures, p, at a constant, compared to RHBP-
ATGP, which varies the value of p and requires an automatic stopping rule to
determine an appropriate value of p, which can be determined by recursive ATGP
(R-ATGP) in Chap. 7. Third and most importantly, knowledge of the signature
matrix M in (16.1) must be provided a priori, while RHBP-ATGP is completely
unsupervised and requires no prior target knowledge.
Chapter 17
Recursive Hyperspectral Band Processing
of Linear Spectral Mixture Analysis
17.1 Introduction
Band selection (BS) is a commonly used data reduction technique to reduce the
original data space to a lower-dimensional band selected data space as an alterna-
tive to data compression, which reduces the original data space into a lower-
dimensional compressed data space. In other words, BS selects a subset of bands
of interest from a fully collected band set according to a certain criterion for
optimality. Linear spectral mixture analysis (LSMA) is a fundamental task that
has been widely used to perform linear spectral unmixing (LSU) for image analysis.
It is generally performed using full bands. Since hyperspectral imaging sensors use
hundreds of contiguous bands to acquire data, the provided spectral information
sometimes is more than what LSMA generally needs. This is particularly true for
LSMA, where only a small number of signatures is required for LSU. Most
recently, it was shown in Chang et al. (2010a, 2011b) that LSMA could be effective
even if only a small number of bands are used, according to a concept, virtual
dimensionality (VD), recently developed by Chang (2003a) and Chang and Du
(2004). More specifically, in the context of VD, if a signature used to unmix data by
LSMA can be effectively represented by a spectral band, then VD can be further
used to determine the number of signatures needed for LSMA to perform effec-
tively. The issue remaining is how to find appropriate signatures that correspond to
these spectral bands. Though this issue is very challenging and deserves further
investigation, this chapter explores an interesting approach different from BS,
called recursive hyperspectral band processing of LSMA (RHBP-LSMA).
The main idea of RHBP-LSMA is derived from progressive spectral band
dimensionality (PSBD), originally developed by Chang et al. (2011d) and
Chap. 21 in Chang (2013), so that LSMA can be performed progressively band
by band without resorting to BS. In other words, LSMA can be carried out band by
band without waiting for the completion of band collection. Such a resulting LSMA
is called progressive hyperspectral band processing of LSMA (PHBP-LSMA)
according to the band-sequential (BSQ) format; it can take place while band
collection is going on. It is important to recognize the major and crucial differences
between PHBP and BS. First, BS is a preprocessing technique that requires com-
plete knowledge of data prior to LSMA, while PHBP can be implemented by band
sequential (BSQ) format in conjunction with LSMA at the same time as bands are
being collected. Second, BS requires a specific optimal criterion to find a subset of
appropriate bands, while PHBP operates without selecting bands. Third, PHBP can
be implemented in such a way that the results unmixed by LSMA can be evaluated
by its progressive performance, while BS/LSMA gives one-shot unmixed results
produced by LSMA using selected bands where the notation “/” is used to indicate
BS and LSMA, which are carried out in sequence, with BS followed by LSMA, as
opposed to the notation “-”used in PHBP-LSMA, which indicates that both PHBP
and LSMA are jointly implemented simultaneously. Finally and most importantly,
PHBP-LSMA can be further extended to a recursive version of PHBP-LSMA,
called RHBP-LSMA, which can update LSMA results only through innovation
information via recursive equations in a manner similar to that of a Kalman filter. In
doing this we decompose the information required for RHBP-LSMA into three
pieces of information, processed information obtained by the previous bands, new
information provided by the current incoming band, and innovation information by
extracting discrepant information between processed information and new infor-
mation. Such information decomposition originating from Kalman filtering enables
us to derive a recursive equation that updates innovation information band by band
without reprocessing all past information. The derived equation includes two terms,
the unmixed results obtained by previous bands and the innovation information that
is provided by the new incoming band but cannot be predicted from the past
17.2 Linear Spectral Unmixing Via Recursive Hyperspectral Band Processing 507
e l to update P1
and the new information m l as follows:
1 T 1
T 1 P mel m e P
P1 e lm
¼ Pl1 þ m el 1
¼ Pl1 l1 T 1l l1 : ð17:5Þ
l
1þm el
e l Pl1 m
v T
#T !
l ðl1Þ l
ðl1Þ Ml1 rl1
^ ðlÞ ¼ @P1
α LS
l1
A
1 þ ρlðl1Þ e lT
m rl
0 1
v
v T
!
l ðl1Þ l
ðl1Þ rl1
¼ @P1
l1
A T
Ml1 el
m
1 þ ρ
rl
l ðl1Þ
v T
l ðl1Þ l
ðl1Þ
1
¼ Pl1 Ml1
T
e l rl
rl1 þ m T
Ml1 e l rl
rl1 þ m
1 þ ρ
l ðl1Þ
v T
l ðl1Þ l
ðl1Þ
1
¼α^ ðl 1Þ þ Pl1 m
LS
e l rl
T
Ml1 rl1 þ me l rl
1þρ
l ðl1Þ
0 1
v
v T
v
v T
l ðl1Þ l
ðl1Þ l ðl1Þ l
ðl1Þ
¼α^ LS ðl 1Þ ^ LS ðl 1Þ þ @P1
Pl1 α l1
Ame l rl
1 þ ρ
1 þ ρ
l ðl1Þ l ðl1Þ
0 1 0 1
v
v T
v
v T
l ðl1Þ l
ðl1Þ l ðl1Þ l
ðl1Þ
¼ @Ipp Pl1 Aα ^ LS ðl 1Þ þ @P1 l1
Ame l rl :
1 þ ρ
1 þ ρ
l ðl1Þ l ðl1Þ
ð17:6Þ
17.2 Linear Spectral Unmixing Via Recursive Hyperspectral Band Processing 509
e l r l defined by
2. New information: m
T
e l r l ¼ ml1 ; ml2 ; . . . ; mlðp1Þ ; mlp r l ;
m ð17:8Þ
e lT Ml1
¼m Ml1 m e lT P1
el ¼ m e lT v
el ¼ m
l1 m ; ð17:10Þ
l ðl1Þ l ðl1Þ
e l.
vl|(l1) and ρl|(l1), and new information, m
510 17 Recursive Hyperspectral Band Processing of Linear Spectral Mixture Analysis
v T
l ðl1Þ l
ðl1Þ
P1 ¼ P1 : ð17:11Þ
l l1
1 þ ρ
l ðl1Þ
v T
l ðl1Þ l
ðl1Þ
^ LS ðlÞ ¼ @Ipp
α Pl1 Aα
^ LS ðl 1Þ þ P1
l me l rl : ð17:12Þ
1 þ ρ
l ðl1Þ
According to (17.12), the recursive process of calculating α ^ LS ðlÞ can be carried out
by the following steps:
h i1
1. Initial condition: P1 T
1 ¼ ð m ð 1Þ Þ m ð 1Þ ^ LS ð1Þ ¼ P1
,α 1 M1 r 1 , where M1 ¼
T
m1 ð1Þm2 ð1Þ mp ð1Þ is a 1 p matrix reduced to a p-dimensional vector.
2. For each 1 < l L, with L being the total number of bands:
e l to compute vl|(l1) and ρl|(l1) by (17.9) and (17.10), respectively;
(a) Input m
v
T
l ðl1Þ l
ðl1Þ
(b) Use 2(a) to calculate 1þρ
;
l ðl1Þ
T
l ðl1Þ l
ðl1Þ
we can simply compute 1þρ
^ LS ðlÞ by (17.12).
and use it to calculate α
l ðl1Þ
^ LS ðlÞ
Table 17.1 Computational complexity of (17.12) to calculate α
h i1
Initial pp T One outer product of a p-dimensional vectors, m(1) ( p2
P1
1 ¼ mð1Þðmð1ÞÞ
conditions matrix P1
1
multiplications) + one matrix inverse requiring O( p3)
T p inner products of p-dimensional vectors ( p2 multiplications)
^ LS ð1Þ
α ^ LS ð1Þ ¼ P1
α 1 M1 r 1
+ p multiplications by r1
Previously processed T T Available from (l 1)st previous recursion, thus, no calculation
Pl1 ¼ Ml1 ^ LS ðl 1Þ ¼ P1
Ml1 , α l1 Ml1 rl1
information α ^ LS ðl 1Þ, is needed
Pl1 , P1
l1
Input m el m
e l rl p multiplications
Innovation information: v
¼ P1 el
l1 m
p inner products of p-dimensional vectors ( p2 multiplications)
l ðl1Þ
vl|(l1) by (17.9)
Innovation information: ρ
¼m
e lT v
v T
One outer product of vl|(l1) ( p2 multiplications + p2 division)
l by (17.11) ðP1 e l Þðm l ðl1Þ l
ðl1Þ
1 l1 m e lT P1l1 Þ 1
P1
l ¼ Pl1 ¼ P l1 1þρ
1þm e
l1 m l
e lT P1
l ðl1Þ
0 1
^ LS ðlÞ by (17.12)
Update α v
v T
p2 inner products ( p3 multiplications) + 2p inner products of
l ðl1Þ l
ðl1Þ
α Pl1 Aα ^ LS ðl 1Þ þ P1 e l rl p-dimensional vectors (2p2 multiplications)
Linear Spectral Unmixing Via Recursive Hyperspectral Band Processing
l ðl1Þ
511
512 17 Recursive Hyperspectral Band Processing of Linear Spectral Mixture Analysis
band number in (17.11) is not specified. It does not have to start with the first
band but rather any band of a desired wavelength. In this case, RHBP allows
users to tune particular bands as they desire.
2. Recently, a new approach to BS, referred to as progressive band selection (PBS),
was developed in Chang and Liu (2014). Although PBS performs LSMA in a
17.4 Synthetic Image Experiments 513
progressive manner, there are several distinct and salient differences between
PBS and RHBP. First and foremost is that RHBP is a recursive process using
(17.11) to update only innovation information band by band, while PBS is not
and must reimplement LSMA every time new bands are added. Second,
although both PBS and RHBP can be implemented progressively band by
band, only RHBP can be used in real time because the use of (17.11) allows
RHBP to be processed recursively without revisiting previous bands, whereas
PBS must recalculate the new signature matrix M as new bands are added. Third,
in order for PBS to work, band decorrelation and band prioritization are gener-
ally required and included as a part of PBS. Because band prioritization and band
decorrelation must be done prior to PBS, PBS generally cannot be implemented
in real time. Finally, by comparison with RHBP, which does not require prior
knowledge of the number of bands to be selected, PBS generally needs to know
in advance how many bands must be selected. To address this issue, PBS
includes the concept of band dimensionality allocation in conjunction with VD
to find the best possible band dimensionality for each signature to be unmixed
by LSMA.
3. As was pointed out, RHBP is a spectral band process, not BS, because RHBP,
using equations (17.6)–(17.11) to perform LSMA, does not involve determining
nBS and finding an optimal band set, ΩBS. However, RHBP can be considered a
generalized version of PBS if it is implemented in conjunction with band
prioritization and band decorrelation, along with VD. Since the value of VD,
nVD, provides a good estimate on bounding nBS from below and above, as shown
in Chang (2013) and Chang and Liu (2014) nVD nBS 2nVD . Using the range
of [nVD, 2nVD] RHBP can be performed progressively by starting the first band
indexed by band number, nVD, and ending up with the last band indexed by
2nVD. So, an optimal band subset can be found from the subset of fBl g2n VD
l¼nVD ,
where all bands in fBl g2n VD
l¼nVD have been prioritized by band prioritization and
decorrelated by band decorrelation. It was shown in Chang (2013) and Chang
p
and Liu (2014) that nBS varies with endmembers, mj j¼1 , used to unmix data
samples, but it does fall in a range of nVD nBS 2nVD .
Finally, it is worth noting that RHBP was developed primarily to perform LSMA
in a BSQ format, where LSMA-unmixed results can be progressively presented to
data analysts, who can determine when BSQ format should be terminated for data
acquisition. Thus, RHBP does not perform BS or data reduction or compression. It
is quite different from conventional BS, where bands must be selected according to
a criterion of optimality to achieve data reduction prior to LSMA.
The synthetic image simulated in Fig. 1.15 is shown in Fig. 17.1, with five panels in
each row simulated by the same mineral signature and five panels in each column
having the same size.
514 17 Recursive Hyperspectral Band Processing of Linear Spectral Mixture Analysis
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
Among the 25 panels are five 4 4 pure pixel panels for each row in the first
column, five 2 2 pure pixel panels for each row in the second column, five 2 2
mixed pixel panels for each row in the third column, and five 1 1 subpixel panels
for each row in both the fourth and fifth columns, where the mixed and subpanel
pixels were simulated according to the Fig. 17.1 legends. Thus, a total of 100 pure
pixels (80 in the first column and 20 in the second column), referred to as
endmember pixels, were simulated in the data by the five endmembers A, B, C,
K, and M. An area of background (BKG), marked “BKG” in the upper right corner
of Fig. 1.14a, was selected to find its sample mean, that is, the average of all pixel
vectors within BKG, denoted by b and plotted in Fig. 1.14b, to be used in the
simulation of BKG for an image scene with a size of 200 200 pixels in Fig. 17.1.
The reason for this background selection is empirical since the selected area “BKG”
seemed more homogeneous than other regions. Nevertheless, other areas can also
be selected for the same purpose. This b-simulated image background was further
corrupted by an additive noise to achieve a certain signal-to-noise ratio (SNR),
which was defined in Harsanyi and Chang (1994) as a 50 % signature (i.e., reflec-
tance/radiance) divided by the standard deviation of the noise.
Once target pixels and background are simulated, two types of target insertion
can be designed to simulate experiments for various applications. The first type of
target insertion is target implantation (TI), which can be simulated by inserting
clean target panels into a noisy image BKG by replacing their corresponding BKG
pixels, where the SNR is empirically set to 20:1. That is, TI implants clean target
panel pixels into noise-corrupted image BKG with SNR ¼ 20:1, in which case there
are 100 pure panel pixels in the first and second columns. A second type of target
insertion is target embeddedness (TE), which is simulated by embedding clean
target panels into a noisy image BKG by superimposing target panel pixels over the
BKG pixels, where the SNR is empirically set to 20:1. That is, TE embeds clean
target panel pixels into noise-corrupted image BKG with SNR ¼ 20:1, in which
17.4 Synthetic Image Experiments 515
case all 100 pure panel pixels in the first and second columns are no longer pure. In
other words, there is a difference between TI and TE worth noting. TE inserts
targets by adding target pixels to and superimposing them over background pixels
instead of replacing background pixels the way TI does. As a consequence, the
abundance fraction of the pixel into which a target pixel is embedded is not summed
to one.
Since both RHBP-NCLS and RHBP-FCLS unmixed all panel pixels in the first
three columns nearly 100 % accurately using nband ¼ number of bands from 6 to
12, starting with band ¼ band 1 and band step size nΔ ¼ 1, their unmixed results are
not included here. Instead, Fig. 17.2 only shows the progressive band-varying
performance of RHBP-NCLS and RHBP-FCLS-unmixed results for subpixel
panels in the fourth and fifth columns in TI with band step size nΔ ¼ 1 starting
with the band number nband ¼ 6–12, with band 1 as the first band and band 6/band
12 as the last bands, corresponding to nband ¼ 6 and 12, respectively.
Note that the reason for setting nband ¼ 12 in Figs. 17.2 and 17.3 as the upper
bound is because 2nVD, that is, twice the value of VD, is enough to have NCLS and
FCLS work effectively.
As for the TI scenario where the panel pixels were embedded in BKG pixels,
Fig. 17.2 shows that when nband is equal to or greater than 6, RHBP-FCLS
accurately unmixed all the subpixel panels in the fourth and fifth columns specified
by all the five mineral signatures A, B, C, K, and M. Similarly, RHBP-NCLS also
worked well as long as nband was equal to or greater than 8. However, compared to
RHBP-FCLS, RHBP-FCLS performed slightly better. Interestingly, the conclu-
sions drawn for the TE scenario where panel pixels were superimposed on BKG
pixels shows otherwise. Since TE did not satisfy ASC, there are no endmembers.
Consequently, the TE scenario is generally used for signal detection rather than
endmember extraction. In this case, finding endmembers becomes a matter of
detecting endmembers. This can be seen in Fig. 17.3, where RHBP-NCLS
performed better than RHBP-FCLS in that RHBP-NCLS produced more accurate
abundance fractions of subpanel pixels in the fourth and fifth columns than RHBP-
FCLS did. It is worth noting that while RHBP-NCLS worked reasonably well in
unmixing subpanel pixels specified by the M signature, RHBP-FCLS did not work
in this case. This may be due to the fact that the M signature’s spectral shape is close
to the spectral profile of the BKG signature according to Fig. 1.14b, in which case
FCLS could not unmix the subpanel pixels well, but as a detector RHBP-NCLS
could do so very effectively as long as nband 7. More surprisingly, RHBP-NCLS
could be used to unmix these subpanel pixels very accurately, a task that could not
by accomplished by RHBP-FCLS. For example, the worst case was subpanel pixels
unmixed by FCLS using the mineral signature A, where its RHBP-FLCS-unmixed
progressive performance seemed to achieve stability after nband 11, but the
progressive RHBP-FCLS-unmixed abundance fractions were still incorrect. Simi-
larly, the same phenomenon can be observed for the mineral signature B after nband
8 and for the mineral signature M after nband 7. The most stable RHBP-FCLS-
unmixed progressive performance was obtained from the signatures C and K, but
516 17 Recursive Hyperspectral Band Processing of Linear Spectral Mixture Analysis
Fig. 17.2 Plots of RHBP-NCLS and RHBP-FCLS-unmixed results of panel pixels in (a) fourth
column and (b) fifth column in TI from nband ¼ 6 ¼ nVD to nband ¼ 12 ¼ 2nVD by RHBP
their unmixed abundance fractions were still not accurate. All these problems were
caused by the simulated pixels, which did not satisfy ASC.
Interestingly, if we use the pure mineral signatures in Fig. 1.14b to form the
signature matrix M instead of the panel pixels directly extracted from TE, as we did
for Fig. 17.3, the unmixed results were similar to those obtained in Chang et al.
17.4 Synthetic Image Experiments 517
Fig. 17.3 Plots of RHBP-NCLS and RHBP-FCLS-unmixed results of panel pixels in (a) fourth
column and (b) fifth column in TE from band 6 (nVD) to band 12 (2nVD) by RHBP
(2010a, 2011b), where all abundance fractions were focused on a single mineral
signature because none of pixels in TE was pure and they all violated ASC. For
more details the interested reader is referred to Chang et al. (2010a, 2011b).
To compare computing times, Figs. 17.4 and 17.5 plots computing times in
microseconds of UCLS, NCLS, and FCLS and their progressive counterparts,
RHBP-UCLS, RHBP-NCLS, and RHBP-FCLS for TI and TE for processing a
518 17 Recursive Hyperspectral Band Processing of Linear Spectral Mixture Analysis
Fig. 17.4 Average computing time of UCLS, NCLS, FCLS, RHBP-UCLS, RHBP-NCLS, RHBP-
FCLS for TI
Fig. 17.5 Average computing time of UCLS, NCLS, FCLS, RHBP-UCLS, RHBP-NCLS, RHBP-
FCLS for TE
single pixel on average based on 10 trials where the y-axis is computing time in
microseconds (ms) and the x-axis is nband, starting from band 1 and ending with
189, the last band. The computer environment used for the computations is as
follows: operating system: Windows 7 Premium; CPU: i5 480 m; memory: DDRIII
8 GB; hard disk drive: 500 GB, 7200 rpm; compiler: Microsoft Visual C++ 2008
SP1.
As shown in Figs. 17.4 and 17.5, the computational cost required for processing
a single pixel by all progressive versions of LSMA techniques, RHBP-UCLS,
17.5 Real Image Experiments 519
RHBP-NCLS, and RHBP-FCLS, is nearly constant for each newly added band. By
contrast, the computing times required by UCLS, NCLS, and FCLS are linearly
increased with nband. This means that the more bands are involved in LSMA, the
more computing time for each data sample is required for LSMA techniques, and
the more significant are the savings in computing time. However, if a progressive
LSMA technique is used for LSMA, the computational cost is nearly constant for
each band and is independent of nband used for data unmixing, in which case
processing LSMA for each new incoming band requires the same computational
complexity.
The image data to be studied are from the HYperspectral Digital Imagery Collec-
tion Experiment (HYDICE) image scene shown in Fig. 17.6a (and Fig. 1.10a),
which has a size of 64 64 pixel vectors with 15 panels in the scene and the ground
truth map in Fig. 17.6b (Fig. 1.10b). It was acquired by 210 spectral bands with a
spectral coverage from 0.4 to 2.5 μm. Low signal/high noise bands, bands 1–3 and
a b
p11, p12, p13
7000 p1
p2 interferer
6000 p3
p4
5000 p5
4000 grass
3000
2000
1000
tree
0 road
0 20 40 60 80 100 120 140 160 180
Fig. 17.6 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels; (c) spectra of p1, p2, p3, p4, and p5; (d) ground truth map of 4 undesired signatures
520 17 Recursive Hyperspectral Band Processing of Linear Spectral Mixture Analysis
202–210, and water vapor absorption bands, bands 101–112 and 137–153, were
removed. Thus, a total of 169 bands were used in the experiments. The spatial
resolution and spectral resolution of this image scene are 1.56 m and 10 nm,
respectively.
Figure 17.7 shows RHBP-FCLS and RHBP-NCLS-unmixed abundance frac-
tions of the 19 R panel pixels on the y-axis and nband from the first 8 bands to all the
169 bands on the x-axis. Note that since there are at least nine spectral signatures
present in the data, nband was initialized by starting with the first eight bands. In
addition, no BS was used in these experiments, and bands were simply processed
band by band with a band step size nΔ ¼ 1 according to the original band order. In
this case, bands were tuned purely from their wavelength range beginning with the
first 8 bands and increasing one band at a time to observe the progressive RHBP-
FCLS and RHBP-NCLS-unmixed performance of the 19 R panel pixels in the
HYDICE image scene unmixed via RHBP.
As demonstrated in Fig. 17.7, the results of the abundance fractions of the
19 panel pixels unmixed by RHBP-NCLS seemed to be better than that unmixed
by RHBP-FCLS. However, this conclusion will be reversed (Fig. 17.8) when the
signatures used to unmixed are extracted in an unsupervised manner directly from
the real scene instead of using signatures in Fig. 17.6d by visual inspection.
Nevertheless, Fig. 17.7 demonstrates that, except for the panel pixels in row
1, nband required for RHBP-NCLS and RHBP-FCLS to achieve stable performance
is actually smaller than 80, which are fewer than half the total number of bands,
169. This simple example shows that there is no need to use full bands for NCLS
and FCLS to perform well. Of course, if BS is used to select the appropriate bands,
the number of bands to be selected, nBS, will certainly be fewer than that required
by the results presented here without BS.
It was shown in Chang et al. (2010a, 2011b) that unsupervised LSMA generally
performed better than supervised LSMA for hyperspectral imagery when many
spectrally distinct signatures could not be identified by visual inspection or pro-
vided by prior knowledge, as in the HYDICE data in Fig. 17.6d. The experiments in
Chang et al. (2010a, 2011b) suggested that implementing unsupervised LSMA
(ULSMA) was more effective than supervised LSMA (SLSMA). With the main
focus of this chapter being on least squares-based LSMA, the least squares-based
approach developed in Chang et al. (2010a) was selected over the component
analysis-based approach proposed in Chang et al. (2011b). First, VD was used to
estimate the number of signatures employed to unmix this scene, nVD, which was
nine, with a false alarm probability of PF 103 . Then an automatic target
generation process (ATGP) (Ren and Chang 2003) was used to find target pixels
of interest from the original HYDICE and sphered HYDICE scenes, respectively.
Figure 17.8a shows the 9 target pixels found in HYDICE scene by ATGP that can
n o9
be considered a set of BKG (BKG) pixels, SBKG ¼ bjATGP , in which three
j¼1
panel pixels from rows 1, 3, and 5 are included. Figure 17.8b shows the nine target
pixels extracted from the sphered HYDICE scene by ATGP, which included five
17.5 Real Image Experiments 521
Fig. 17.7 RHBP-NCLS and RHBP-FCLS-unmixed results of 19 R panel pixels in HYDICE scene
from band 8 to band 169
522 17 Recursive Hyperspectral Band Processing of Linear Spectral Mixture Analysis
panel pixels extracted from each of the five rows, and can be considered a set of
n o9
target pixels Starget ¼ tjATGP . Figure 17.8c singled out the five pixels that had
j¼1
n o
been identified as BKG pixels S eBKG ¼ b e ATGP by removing the 4 target pixels
i
with a similarity measure such as spectral angle mapper (SAM), and Fig. 17.8d
shows a total number of 14 pixels obtained by combining the BKG set S eBKG
target
in Fig. 17.8c with the target set S in Fig. 17.8b into a BKG-target merged set
eBKG [ Starget to be used for spectral unmixing where the numbers in the figures
S
17.5 Real Image Experiments 523
Fig. 17.8 ATGP-generated BKG and target pixels: (a) 9 BKG pixels in original data; (b) 9 target
pixels in sphered data; (c) 5 BKG pixels not identified as target pixels; (d) 14 pixels obtained by
combining the pixels in (b, c)
indicate the orders of pixels extracted by ATGP. Using these 14 target pixels to
form a signature matrix M, FCLS was then performed by RHBP band by band.
Since the total number of used bands was 14, the initial band number, nband, was set
to 13 and FCLS was then performed progressively band by band via (17.6) to unmix
the 19 R panel pixels in Fig. 17.6b until it reached the full band number, 169. In
analogy with Figs. 17.7, Fig. 17.9 also plots the progressive FCLS-unmixed results
of these 19 R panel pixels in terms of their unmixed abundance fractions on the y-
axis and nband from the first 13 bands to the full 169 bands on the x-axis.
The results presented in Fig. 17.9 show that only required half of the spectral
bands were required to achieve stable performance for all 19 R panel pixels in terms
of unmixed abundance fractions. Also, RHBP-constrained LS methods sensitive to
the number of endmembers if that number is not properly selected. However, once
it reaches a certain number, its performance tends to be robust, as shown in
Fig. 17.9. For example, for TI and TE this number is nVD ¼ 6 and nVD ¼ 18 for
the HYDICE data. This phenomenon is also observed in traditional UCLS (i.e.,
unsupervised unconstrained least-squares orthogonal subspace projection), NCLS,
and FCLS methods described in Chap. 4.
According to the ground truth, the pure panel pixels in the first column are
supposed to have 100 % purity in abundance. In fact, the progressive profiles of
unmixed abundance fractions showed otherwise for some panel pixels, such as p211,
p311, p412, and p511, which are actually not true pure pixels since p211 and p311 have
85–95 % purity, panel p412 has 41 % purity, and p511 has 72 % purity. This makes
sense because two panel pixels account for twice the spatial resolution, 1.56 m, that
is, 3.12 m compared to the panel size in the first column, which has a size of only
3 3m. Furthermore, comparing the results in Fig. 17.9 to those in Fig. 17.7, it is
clearly demonstrated that RHBP-FCLS did perform better than RHBP-NCLS in
terms of accuracy regarding unmixed abundance fractions of the 19 R panel pixels.
This conclusion is the complete opposite of that drawn from Fig. 17.7. The key
factor lies in the fact that the signatures used for data ummixing must be generated
directly from an unsupervised algorithm, unsupervised LSMA (ULSMA), without
human intervention, such as prior knowledge or visual inspection, supervised
LSMA (SLSMA). Moreover, nband required for ULSMA was much smaller than
524 17 Recursive Hyperspectral Band Processing of Linear Spectral Mixture Analysis
Fig. 17.9 Unmixed results of 19 R panel pixels in HYDICE scene by unsupervised RHBP-NCLS
and RHBP-FCLS from band 13 to band 169
17.5 Real Image Experiments 525
for some panel pixels, for example, pixels p11, p221, p312, p411, and p521, as shown
in Fig. 17.9.
Three comments on the experiments are worthwhile:
1. Despite the fact that only one real hyperspectral scene was used for the exper-
iments in this chapter, the conclusions drawn from these experiments for this
scene can also be applied to other real data sets. However, in this case, ground
truth must be provided for verification. The presented HYDICE data seemed to
be the best example we had for illustration. For quantitative analysis the exper-
iments conducted in Sect. 17.4 for two types of synthetic images, TI and TE, also
confirmed our conclusions for real data experiments.
2. It is unfortunate that a vivid real-time demonstration of real HYDICE experi-
ments could not be shown in this chapter to see how the progressive unmixed
results of each data sample vector gradually changed as bands were added band
by band progressively. However, as the band number, nband, reached a certain
limit, the progressive LSMA did not necessarily improve even if more bands
were added. Nevertheless, unmixed results did indeed progressively improve
before nband reached a specific number. But this number varies with different
signatures.
3. As noted, RHBP can be used to process specific bands of interest, such as a range
of wavelengths and visible or infrared bands, in which case RHBP can tune these
particular bands as desired. Using panel pixel p22 as an illustrative example, its
unmixed performance was stable with nband between 8 and 20 but became worse
between 20 and 33 and flat between 34 and 52, then began improving again
between 53 and 70 and fluctuating between 70 and 100. It finally reached
stability after nband was greater than 100. Thus, in this specific example, the
visible wavelengths did not improve but rather deteriorated the unmixed perfor-
mance of p22. Instead, the near infrared wavelengths actually provided useful
and valuable information for improving unmixed performance. Such similar
evidence were also found in other panel pixels.
To illustrate additional computational savings resulting from progressive band
processing of FCLS unmixing, Fig. 17.10 plots the computing time in microseconds
(ms) required for three LSMA techniques, UCLS, NCLS, and FCLS, along with
their corresponding progressive versions, RHBP-UCLS, RHBP-NCLS, and RHBP-
FCLS, on a Windows 7 Premium machine running an i5-480M processor, with
DDR3 RAM, 8 GB memory, 500 GB 7200 rpm hard disk drive, and Microsoft
Visual C++ 2008 SP1 compiler.
Similar to Figs. 17.4 and 17.5, Fig. 17.10 also shows the plots of, the three
progressive LSMA techniques which saved significant amounts of computing time
compared to their counterparts without taking advantage of RHBP, where the
computing time required by the former was nearly constant but that required by
the latter linearly increased with nband.
17.6 Conclusions 527
Fig. 17.10 Computing time in seconds of RHBP of UCLS, NCLS, and FCLS along with their
corresponding progressive versions, RHBP-UCLS, RHBP-NCLS, and RHBP-FCLS
17.6 Conclusions
This chapter developed a new concept of RHBP-LSMA that can perform LSMA
band by band recursively in real time, as if it were a Kalman filter, at the same time
bands are being collected. To the author’s best knowledge no similar technique has
been reported in the literature. Several interesting findings are noteworthy. First,
RHBP does not select bands. Instead, it allows users to tune bands of interest and
process these bands progressively band by band and recursively in real time. As a
result, RHBP does not require one to determine nBS or to find appropriate bands, as
is the case with BS. RHBP can be performed whenever bands are available without
waiting for full bands to be collected. Second, RHBP takes advantage of recursive
equations to update only innovation information without reprocessing previous
bands. As a consequence, RHBP can be implemented in real time band by band
progressively and recursively. With this real-time capability, RHBP is a very
promising technique in satellite data communication, transmission, and compres-
sion, none of which can be accomplished with BS. Third, RHBP-LSMA enables
users to evaluate progressive background suppression resulting from LSMA band
by band during data transmission in the BSQ format. In addition, RHBP also helps
users identify which bands are more effective and crucial in LSMA. Last but not
least, RHBP offers computational savings, which is a significant and tremendous
advantage. This is particularly important for hyperspectral imaging sensors oper-
ating on a space-based platform. As a concluding remark, it should also be noted
that RHBP is application-oriented, where its utility must be justified by applica-
tions. The RHBP-LSMA investigated in this chapter has demonstrated great poten-
tial in hyperspectral data communication between transmitting and receiving ends
once all necessary preprocessing tasks are accomplished. Most recently, the RHBP
idea was also applied by Wu et al. (2013) to the modified FCLS (MFCLS)
developed in Chang (2003, Chap. 10).
Chapter 18
Recursive Hyperspectral Band Processing
of Growing Simplex Volume Analysis
18.1 Introduction
As noted in Chap. 2, the main issue arising in the implementation of SGA is to use
the matrix determinant to calculate SVs, referred to as determinant-based SV
(GSV), by finding matrix inverses, which requires an enormous amount of com-
puting time. Fortunately, Chap. 11 develops an approach, called orthogonal
projection-based SGA (OPSGA), which converts finding GSVs to finding maximal
OP-based SVs. This approach turns out to be the same as using an automatic target
generation process (ATGP) to find its targets. Accordingly, what we developed,
RHBP-ATGP, in Chap. 15 is readily applied to developing RHBP-OPSGA, but it
requires significant modifications.
simplexes, in this chapter we use growing simplex volume analysis (GSVA) instead
of SGA to reflect the growing nature of (18.1). Interestingly, finding such maximal
OP, hj+1, can actually be done using ATGP, as shown by OPSGA in Chap. 11. This
fact provides us with a feasible means of developing RHBP-OPSGA to be
implemented as RHBP-OPSGA, where OPGSA is used to replace SGA.
j T
First, let mj j¼1 be the jth endmembers and mjl ¼ mj1l ; mj2l ; . . . ; mjll an l-
dimensional vector formed by the first l bands of the jth endmember elj , and let m
T
ðlÞ ¼ m1l ; m2l ; . . . ; mjl be a j-dimensional vector obtained by the lth band from
each of the j endmembers. Now let
h i
Ulj ¼ m1l m2l mjl
2 3
m11l
m21l
mðlj1Þ1 mj1l
6 7
6 ml l
⋱ mðlj1Þ2 mj2l 7
6 m22 7 " #
6 12
7 Uðl1Þj ð18:2Þ
6 7
¼6 ⋮ ⋱ ⋱ ⋱ ⋮ 7¼ :
6 7 m T
ð l Þ
6 ml l
mðlj1Þðl1Þ 7
mjlðl1Þ 7
6 1ðl1Þ m2ðl1Þ
4 5
m1ll m2ll mðlj1Þl mjll
h i1 h i
1
Uðl1Þj
T
Ulj Ulj ¼ mðlÞ
UðTl1Þj
mT ðlÞ ð18:3Þ
1
¼ UðTl1Þj Uðl1Þj þ mðlÞmT ðlÞ ;
2 3
l
m11 l
m21 mðlj1Þ1 mj1l
6 7
6 ml m22l
⋮ mðlj1Þ2 mj2l 7
6 12 7
where Uðl1Þj ¼ 6 7 is an ðl 1Þ p
6 ⋮ ⋮ ⋱ ⋮ ⋮ 7
4 5
m1ðl1Þ m2ðl1Þ mðj1Þðl1Þ mjðl1Þ
l l l l
P⊥
Ulj ¼ Ill Ulj Ulj
#
0 1
1
"
T
#B Uðl1Þj Uðl1Þj UðTl1Þj ρ C
l ðl1Þ
Uðl1Þj B C
¼ Ill B h iC
B C
mT ðlÞ @ 1 ρ ρ T UT ρ ρ T mðlÞ A
l ðl1Þ lðl1Þ ðl1Þj l ðl1Þ lðl1Þ
1 þ mT ðlÞρ
l ðl1Þ
2 3
P⊥
Uðl1Þj Uðl1Þj ρ
l ðl1Þ
6 7
¼6
4 1
T 7
5
Uðl1Þj UðTl1Þj Uðl1Þj mðlÞ 1 mT ðlÞρ
l ðl1Þ
2 3
Uðl1Þj ρ ρ T UT Uðl1Þj ρ ρ T mð l Þ
l ðl1Þ lðl1Þ ðl1Þj l ðl1Þ lðl1Þ
1 6 7
þ 4 5
1 þ mT ðlÞρ mT ðlÞρ ρ T UðTl1Þj mT ðlÞρ ρ T mðlÞ
l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ
ð18:5Þ
T
⊥
P rl 2 ¼ r T P⊥ rl ¼ r T P⊥ rl1 þ r l 1 mT ðlÞρ
Ulp l Ulp l1 Uðl1Þp rl1 2r l Uðl1Þp ρ l ðl1Þ l ðl1Þ
rl
1
þ T
rl1 Uðl1Þp ρ þ ρ T mðlÞr l ρ T UðTl1Þp rl1 þ ρ T mðlÞr l ;
1 þ mT ðlÞρ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ
l ðl1Þ
ð18:6Þ
1
ðlÞ
2. At the lth iteration, find mj by maximizing rT P⊥
Ulj r via (18.6) over all data
N
sample vectors, frl ðiÞgi¼1 . This can be done in the following steps:
ðlÞ
(a) Initial conditions: Let maxlj ¼ ðrl ðiÞÞT P⊥ Ulj rl ðiÞ and mj rl ðiÞ. Set
i ¼ 2.
(b) At the ith iteration, calculate ðrl ðiÞÞT P⊥
Ulj rl ðiÞ.
(c) If ð r l ð i Þ Þ T P⊥ j
Ulj rl ðiÞ > maxl , then maxlj ¼ ðrl ðiÞÞT P⊥
Ulj rl ðiÞ and
ðlÞ
mj rl ðiÞ. Otherwise continue.
(d) If i < N, then i i þ 1, and go step 1.b. Otherwise, the algorithm is
ðlÞ
terminated and mj is already found.
(e) Let l l þ 1, and use (18.3) and (18.5) to update U#lj and P⊥ Ulj via
previously calculated U#ðl1Þj and P⊥
Uðl1Þj .
l=j
l l 1
Recursive
Band
Process ml rl (1)
Progressive process
T T
ri (l ) P ri (l ) ml P ml
Ul U l No
Yes
ml rl (i )
No
l =L
Yes
j j 1
Yes No
Stop j> p
j T
First, let mj j¼1 be the jth endmembers and mjl ¼ mj1l ; mj2l ; . . . ; mjll be an l-
dimensional vector formed by the first l bands of the jth endmember elj , and let m
T
ðlÞ ¼ m1l ; m2l ; . . . ; mjl be a j-dimensional vector obtained by the lth band from
each of the j endmembers. Now let
536 18 Recursive Hyperspectral Band Processing of Growing Simplex Volume Analysis
h i
Ulj ¼ m1l m2l mjl
2 3
m11l
m21l
mðlj1Þ1 mj1l
6 7
6 ml mj2l 7
⋱ 7 2 3
l
6 m22 mðlj1Þ2
6 12
7 U ð18:10Þ
6 7 4 ðl1Þj 5
¼6 ⋮ ⋱ ⋱ ⋱ ⋮ 7¼ :
6 7
6 ml l
mðlj1Þðl1Þ mjlðl1Þ 7 mT ðlÞ
6 1ðl1Þ m2ðl1Þ 7
4 5
m1ll m2ll mðlj1Þl mjll
2 3
l
m11 l
m21 mðlj1Þ1 mj1l
6 7
6 ml l
⋮ mðlj1Þ2 mj2l 7
6 m22 7
¼6 7 is an ðl 1Þ p
12
where Uðl1Þj 6 7
6 ⋮ ⋮ ⋱ ⋮ ⋮ 7
4 5
m1l ðl1Þ m2l ðl1Þ mðlj1Þðl1Þ l
mjðl1Þ
matrix and
#
P⊥
U
¼ Ill Ulj Ulj
lj
0 1
1
T T
ρ
" #B U U
ðl1Þj ðl1Þj U ðl1Þj lðl1Þ C
Uðl1Þj B C
¼ Ill B
C
B 1 T T T C
m ðlÞ
T @ ρ ρ U ρ ρ mðlÞ A
l ðl1Þ lðl1Þ ðl1Þj l ðl1Þ lðl1Þ
1 þ m ðlÞρ
T
l ðl1Þ
2 3
PU⊥
Uðl1Þj ρ
ðl1Þj l ðl1Þ
6 7
¼64 1
T 7
5
1 mT ðlÞρ
T
Uðl1Þj Uðl1Þj Uðl1Þj mðlÞ
l ðl1Þ
2 3
Uðl1Þj ρ ρ T Uðl1Þj Uðl1Þj ρ ρ T
T
mð l Þ
1 6 l ð l1 Þ l ðl1 Þ l ðl1 Þ l ðl1 Þ 7
þ 6 7
1 þ mT ðlÞρ 4 T T T T
5
l ðl1Þ m ðlÞρ ρ Uðl1Þj m ðlÞρ
T
ρ mðlÞ
l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ
ð18:12Þ
18.3 Recursive Hyperspectral Band Processing of Geometric Simplex Growing Algorithm 537
T
⊥
P rl 2 ¼ r T P⊥ rl ¼ r T P⊥
l U l1 U rl1 2r l Uðl1Þp ρ rl1 þ r l 1 mT ðlÞρ rl
U lp lp ðl1Þp l ðl1Þ l ðl1Þ
1
Uðl1Þp ρ þ ρ T ρ T Uðl1Þp rl1 þ ρ T
T
þ T
rl1 mðlÞr l mðlÞr l ;
1 þ mT ðlÞρ l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ
l ðl1Þ
ð18:13Þ
1
where ρ
T
¼ Uðl1Þj Uðl1Þj mðlÞ is a j-dimensional vector.
l ðl1Þ
i ¼ 2.
(b) At the ith iteration, calculate ðrl ðiÞÞT P⊥
Ulj rl ðiÞ.
(c) If ðrl ðiÞÞT P⊥
U
rl ðiÞ > maxlj , then maxlj ¼ ðrl ðiÞÞT P⊥
U
rl ðiÞ and
lj lj
ðlÞ
mj rl ðiÞ. Otherwise, continue.
(d) If i < N, then i i þ 1, and go step 1(b). Otherwise, the algorithm is
ðlÞ
terminated and mj is already found.
#
3. Let l l þ 1, and use (18.3) and (18.5) to update Ulj and P⊥
U
via
lj
#
previously calculated Uðl1Þj and P⊥
Uðl1Þj
.
538 18 Recursive Hyperspectral Band Processing of Growing Simplex Volume Analysis
l=j>1
l l 1
Band
Progressive
process ml rl (1)
Successive process
T T
ri (l ) P ri (l ) ml P ml
Ul Ul No
Yes
ml rl (i )
No
l =L
Yes
j j 1
Yes No
Stop j> p
The image data to be studied are from the HYperspectral Digital Imagery Collec-
tion Experiment (HYDICE) image scene shown in Fig. 18.3a (and Fig. 1.10a),
which has a size of 64 64 pixel vectors with 15 panels in the scene and the ground
truth map in Fig. 18.3b (Fig. 1.10b).
Note that panel pixel p212, marked yellow in Fig. 18.3b, is of particular interest.
Based on the ground truth, this panel pixel is not a pure panel pixel and is marked
yellow as a boundary panel pixel. However, in our extensive and comprehensive
experiments, this yellow panel pixel is always extracted as the one with the most
spectrally distinct signature compared to the R panel pixels in row 2. This indicates
that a signature of spectral purity is not equivalent to a signature of spectral
distinction. In fact, in many cases panel pixel p212 instead of panel pixel p221 is
the one extracted by EFAs to represent the panel signature in row 2. Also, because
of such ambiguity, the panel signature representing panel pixels in the second row is
either p221 or p212, which is always difficult to determine using EFAs. This implies
that the ground truth of the R panel pixels in the second row in Fig. 18.3b may not
be as pure as was thought.
Since GSGA has been shown to be the best among all GSGA variants (Chap. 12),
the RHBP-GSVA used for the following experiments is implemented by RHBP-
GSGA. Figure 18.4 illustrates a comparative study between GSGA using full bands
and RHBP-GSGA. Figure 18.4a shows the results of applying GSGA to the
HYDICE scene with the number of endmembers, nE, estimated for this scene as
nE ¼ 18 in Chang et al. (2010a, 2011b). Figure 18.4b shows the spatial locations of
endmembers found by RHBP-GSGA using the same value of 18 for nE, where the
magenta circle denotes the endmembers extracted in the previous bands, the cyan
upper triangle highlights the endmembers identified in the current band, and the
number next to the triangle indicates the orders of the extracted endmembers. As
also shown in the figure, RHBP-GSGA produced the same results as GSGA at the
last iteration, in which case RHBP-GSGA can be considered a slow motion of a
a b
p11, p12, p13
Fig. 18.3 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial locations
of the 15 panels
540 18 Recursive Hyperspectral Band Processing of Growing Simplex Volume Analysis
Fig. 18.4 18 endmembers found by GSGA and RHBP-GSGA, with nE ¼ 18. (a) GSGA. (b)
RHBP-GSGA
one-shot operation GSGA. Most interestingly, some moving targets can be unveiled
during the process, where the pixels in the magenta circle represent the additional
endmembers detected by RHBP-GSGA. Furthermore, a total of eight ground truth
pixels (seven R panel pixels and one Y panel pixel) can be discovered by RHBP-
GSGA, compared with only five ground truth pixels (four R panel pixels and one Y
panel pixel) picked up by GSGA.
Let nl denote the number of the first l bands being used for data processing.
Figure 18.5 further shows the spatial locations of endmembers found by PHBP-
GSGA using different numbers of bands processed, nl, with l starting from
18 bands.
In Fig. 18.5, the red cross indicates the spatial locations of the R panel pixels in
the HYDICE scene, the yellow cross shows the spatial locations of the Y panel
pixels in the scene, the magenta circle denotes the endmembers extracted in the
previous bands, the cyan upper triangle highlights the endmembers identified by the
current band, and the numbers next to the triangle indicate the orders of the
extracted endmembers. p212 and p521 were the first two panel pixels extracted,
with 18 bands collected and received, followed by p311 with nl ¼ 41, p312 with
nl ¼ 49, p11 with nl ¼ 50, p411 with nl ¼ 60, p221 with nl ¼ 67, and p412, the last panel
pixel identified, with nl ¼ 149. No more panel pixels were extracted after nl ¼ 149.
Figure 18.6 shows progressive magnitude changes in RHBP-GSGA in detecting
the 18 endmembers identified by RHBP-GSGA, where the x-, y-, and z-axes denote
respectively extracted endmembers, the number of the first bands, nl, being used for
data processing, and the magnitudes calculated by RHBP-GSGA. The magenta
arrow indicates that a particular panel pixel is being identified for the first time as an
endmember. For example, p212 was identified for the first time as an endmember
with nl ¼ 18, which was the fourth endmember identified by RHBP-GSGA.
18.4 Real Image Experiments 541
e(18)
4
= p212
e9(18) = p521
PBP-OPSGA Calculated Height
(49)
e68 = p312
4 e(41) = p311
x 10 60
5 0
4 e(67) = p e(50) = p11
94 221 71 25
3 e(60) = p
89 411
2 50
1 75 s
nd
0 Ba
d
100 se
E1 E2 es
E3 E4 e(149) oc
Pr
= p412
E5 E6 115
125 f
E7 E8 ro
E9 E10 be
Endm
embe
E11 E12 150 Num
rs E13 E14
E15 E16
E17 E18 170
Table 18.1 summarizes the minimal number of bands, nl, required to identify
signatures, orders of extracted signatures, and orders of signatures extracted by
particular bands for all eight panel pixels identified by RHBP-GSGA in the
HYDICE scene.
542 18 Recursive Hyperspectral Band Processing of Growing Simplex Volume Analysis
Table 18.1 Summary of minimal number of bands processed, nl, to identify signatures and orders
of extracted signatures by particular bands
Minimal nl
Signatures found by identifying Order of signatures Order of signatures
RHBP-GSVA signatures to be found found in particular bands
ð50Þ 50 71 16
p11 ¼ t71
ð18Þ 18 4 4
p221 ¼ t4
ð67Þ 67 94 11
p212 ¼ t94
ð41Þ 41 60 9
p311 ¼ t60
ð49Þ 49 68 6
p312 ¼ t68
ð60Þ 60 89 18
p411 ¼ t89
ð149Þ 149 115 15
p412 ¼ t115
ð18Þ 18 9 9
p521 ¼ t9
18.5 Conclusions
This chapter presented a new approach, RHBP, to finding endmembers that allows
users to find endmembers band by band progressively. RHBP extends a well-known
SGA into RHBP-GSVA, which has two derived versions, RHBP-OPSGA and
RHBP-GSGA, corresponding to OPSGA and GSGA in Chaps. 11 and 12, respec-
tively. Since endmembers are generally considered insignificant targets because of
their rare appearance in the data, they may very likely be missed and overwhelmed
by strong targets in many cases when full bands are used for one-shot operations.
The RHBP-OPSGA/RHBP-GSGA developed in this chapter provides advantages
for finding endmembers during its progressive interband processing. In fact, as
shown in experiments, RHBP-GSGA can find more endmembers than GSGA
during its progressive endmember finding process.
Chapter 19
Recursive Hyperspectral Band Processing
of Iterative Pixel Purity Index
Abstract The pixel purity index (PPI) is a very popular endmember finding
algorithm due to its availability in ENvironment for Visualizing Images software.
It is discussed in Chang (Hyperspectral data processing: algorithm design and
analysis. Hoboken, 2013) in great detail, where various versions of PPI are devel-
oped from an algorithmic implementation point of view. Most recently, PPI was
further expanded into the iterative PPI (IPPI) by Chang and Wu (IEEE J Sel Top
Appl Earth Obs Remote Sens 8(6):2676–2695, 2015) with details also discussed in
Chang (Real time progressive hyperspectral image processing: endmember finding
and anomaly detection. New York, 2016). All such developments of PPI and IPPI
are derived according to a data acquisition format, band-interleaved-by-pixel/sam-
ple (BIP/BIS), where data sample vectors are collected with full band information
by a hyperspectral imaging sensor. This chapter takes a different approach to
designing IPPI from another data acquisition point of view, called the band-
sequential (BSQ) format, which allows users to process IPPI bandwise without
waiting for full bands of data information to be collected. BIP/BIS and BSQ are the
two types of data acquisition formats generally used to acquire hyperspectral
imagery (Remote sensing: models and methods for image processing, New York,
1997). Basically, IPPI is suitable for both data acquisition formats. Interestingly,
designing an IPPI using the BSQ format has not received attention to date. To make
it work, IPPI must be capable of calculating and updating PPI counts of data sample
vectors band by band. Intuitively, it seems that could be easily done. Unfortunately,
several major challenging issues arise. Unlike many other hyperspectral imaging
algorithms that only deal with sample vectors and bands by visualizing their
progressive changes in algorithmic performance in a three-dimensional (3D) plot,
IPPI has an extra parameter, called skewers, which are randomly generated vectors,
that needs to be specified. By including skewers as a third parameter in addition to
sample vectors and bands, PPI counts will require a four-dimensional plot to
visualize progressive changes in PPI counts in a 3D parameter space specified by
sample vectors, bands, and skewers. In this chapter we introduce a new concept for
implementing IPPI band by band in a progressive manner called progressive
hyperspectral band processing of IPPI (PHBP-IPPI) to address these issues by
developing bandwise calculation of PPI counts of sample vectors sample by sample
and bandwise calculation of PPI counts of sample vectors skewer by skewer.
To further facilitate the use of PHBP-IPPI in practical applications, recursive
equations are also derived for PHBP-IPPI to be used in the development of
19.1 Introduction
One of the most popular endmember extraction algorithms is the pixel purity index
(PPI), developed by Boardman (1994). It searches for a set of vertices of a convex
hull in a given data set, which are supposed to be pure signatures present in the data.
It has been widely used because it is available in the ENvironment for Visualizing
Images (ENVI) software system originally developed by Analytical Imaging and
Geophysics (AIG) (Research Systems, Inc. 2001) and has found many applications
in various areas. Owing to its propriety nature and limited published results, its
detailed implementation has never been made public. Therefore, most of the people
who use the PPI to find endmembers either use ENVI software or implement their
own versions of the PPI based on what is available in the literature. This chapter
presents our experience with PPI and investigates several issues arising in the
practical implementation of the PPI.
To better take advantage of the PPI, several approaches have been developed to
extend the PPI, such as the fast iterative PPI (FIPPI) (Chang and Plaza 2006),
random PPI (RPPI) (Chang et al. 2010c), and iterative PPI (IPPI) (Chang and Wu
2015). However, to the author’s best knowledge, all the PPI-variant developments
reported in the literature are based on using the full set of spectral bands. This
chapter presents a rather different extension of IPPI, called progressive
hyperspectral band processing (PHBP) of IPPI, from a data acquisition point of
view, band-sequential (BSQ) format (Schowengerdt 1997), where bands are
acquired band by band instead of the commonly used data acquisition format,
band-interleaved-sample/pixel (BIS/BIP) or band-interleaved-line (BIL), which
acquires data samples using full band information sample by sample or line by
line, respectively. As a result, PHBP-IPPI can be performed in real time as a new
band comes in without waiting for all bands being completely collected.
It is important to realize the differences between BIS/BIP and BSQ; the former
can be considered a sequential process, whereas the latter is a progressive process.
In general, BIS/BIP acquires data sample vectors sample by sample sequentially
using all spectral bands with no need to revisit data sample vectors. By contrast,
under the BSQ format, data samples are acquired band by band progressively in the
sense that data sample vectors must be revisited L times, where L is the total number
19.1 Introduction 545
of spectral bands. In other words, if we interpret BSQ acquiring data samples by one
particular band as a state process, then BSQ can be considered an L-stage progres-
sive process. In the context of the preceding interpretations, the entire PPI, along
with its extensions reported in the literature according to the BIS/BIP format, can be
considered sequential hyperspectral imaging techniques, while PHBP-IPPI using
the BSQ format can be considered a progressive hyperspectral imaging technique
where the spectral resolution is progressively and gradually increased and improved
by adding more bands. Consequently, PHBP-IPPI takes advantage of its progres-
sive nature to capture details in profiling spectral variations as more bands are
added for data processing.
A key to implementing PHBP-IPPI dealing with the issue of PPI count calcula-
tion, which varies with the number of bands, p, as well as with the total number of
skewers, K, to be used by IPPI. Technically speaking, a skewer is a randomly
generated unit vector pointing to a particular direction with which data samples can
be aligned. How to effectively determine an appropriate value for K remains
unresolved and is a very challenging issue (Chang and Plaza 2006). When it
comes to PHBP-IPPI, a similar issue also arises when determining how many
bands are sufficient for IPPI to find a desired set of endmembers. In addition, a
new challenging issue that is not encountered in IPPI using full band information is
how to process skewers band by band. In other words, developing RHBP-IPPI is
much more difficult than it looks because the PPI count calculation involves
skewers that vary with bands every time a new band is fed in. To address this
issue, two approaches are developed in RHBP-IPPI. One is to fix a set of skewers
for all bands currently being considered, and skewers are only updated by the most
recent band information provided by an incoming band. Another approach is to
vary skewer sets with bands by randomly generating a completely new set of
skewers as a band comes in. More specifically, as a new band comes in, the skewer
sets used by PHBP-IPPI must be independently generated, and in this way, the
skewer sets change band by band. By varying skewer sets, the profile changes in the
PPI counts of data sample vectors captured by PHBP-IPPI provide very valuable
information for data analysis. This is indeed a great benefit resulting from PHBP-
IPPI. In particular, when skewer sets also vary band by band, PHBP-IPPI offers
additional valuable information about the impact of varying skewer sets on changes
in PPI counts. Such advantages allow image analysts to keep track of the signifi-
cance of data sample vectors. Another approach is to help find crucial and signif-
icant bands for data processing according to the progressive changes in PPI counts.
Intuitively, PHBP-IPPI can be considered a band-to-band slow-motion version of
IPPI that can dictate slow varying changes in spectral variation that cannot be
provided by IPPI using full band data processing as a one-shot operation.
To facilitate the use of PHBP-IPPI, recursive equations are also derived to make
PHBP-IPPI a recursive process, called recursive band processing of IPPI (RHBP-
IPPI). With such recursive structures RHBP-IPPI can be implemented more effi-
ciently and effectively in practical applications, specifically in hardware design. It
is believed that RHBP-IPPI may open up many future applications such as satellite
data communication and transmission.
546 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
19.2 PPI
e1
maximal NPPI (e1)=1 maximal
projection projection
skewer3
skewer2
x
NPPI (x)=1
skewer1
e3
NPPI (e3)=3
e2
NPPI(e2)=2 minimal
projection
Fig. 19.1 Illustration of PPI with three endmembers, e1, e2, and e3
19.3 IPPI
Although the MATLAB version of PPI described in Sect. 19.2 remedies some of
drawbacks inherited in PPI, such inherent disadvantages make PPI difficult to apply
directly to real-time or causal processing because PPI requires generating all
K skewers prior to the computation of PPI counts for data sample vectors. Also,
according to its original design, PPI is also not iterative. This section presents two
progressive versions of IPPI, referred to as progressive iterative PPI (P-IPPI) and
causal iterative PPI (C-IPPI), to resolve the aforementioned issues. While the idea
of developing IPPI can be traced back to Wu and Chang (2012), the proposed
P-IPPI and C-IPPI are actually derived from successive N-FNDR (SC N-FINDR)
and sequential N-FINDR (SQ N-FINDR) (Wu et al. 2008), respectively.
19.3 IPPI 549
19.3.1 P-IPPI
The idea of P-IPPI was inspired by the work of Wu et al. (2008) and Xiong et al.
(2011), where two iterative processes are designed, one for data sample vectors
implemented in an inner loop and the other for skewers in an outer loop. For each
inner loop the iterative process is carried out for each incoming data sample vector.
In the context of the work of Wu et al. (2008) and Xiong et al. (2011), it is
considered a single-pass real-time processing of all data sample vectors for a
given skewer in the inner loop, while the number of skewers, K, determines how
many passes are needed for P-IPPI to complete its task in the outer loop. The most
important advantage of P-IPPI is that there is no need to fix the value of K, as is
required by PPI. Thus, technically speaking, P-IPPI can be run indefinitely until its
performance is no longer improved. Another advantage is that P-IPPI can updates
PPI counts of all data sample vectors in real time skewer by skewer once a new
skewer is generated. More specifically, the PPI count for each data sample vector
can be easily updated without recalculating its PPI count.
What follows is a step-by-step detailed implementation of P-IPPI, with index
i used to run data sample vectors in the inner loop for each given generated skewer,
while index k is used to run skewers in the outer loop.
P-IPPI
1. Initialization:
N
Assume that fri gi¼1 are data sample vectors inputted from the data set according
to 1, 2, . . ., N and K is the number of skewers. Let N PPI ðri Þ ¼ 0 and k ¼ 1.
Outer loop: iteratively producing a new skewer
2. Generate a random unit vector skewerk. Set maxðkÞ ¼ fg, max value ¼
r1T skewerk , minðkÞ ¼ fg, and min value ¼ r1T skewerk .
Inner loop: iteratively processing incoming data sample vectors for each skewer
specified by the outer loop
3. For i 2, check whether
If so, maxðkÞ ¼ fig and max valueðkÞ ¼ riT skewerk . Otherwise, continue.
4. Repeat step 3 to find min(k) and continue.
5. Let i i þ 1, and check whether i ¼ N. If not, go to step 3. Otherwise, continue.
6. N PPI ðri Þ N PPI ðri Þ þ 1 for i 2 maxðkÞ or i 2 minðkÞ, and continue.
7. If there is a prescribed number of skewers, K, then the algorithm is terminated
when k reaches K. Otherwise, let k k þ 1, and go to step 2.
550 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
19.3.2 C-IPPI
K
Inner loop: iteratively processing all skewers fskewerk gk¼1 for each inputted
data sample vector ri with i 2.
3. For k 2, check whether
19.3.3 FIPPI
(i) Perform P-IPPI via (19.2) and (19.3) for all the data samples for each skewer,
ðkÞ ðkÞ
skewerj by projecting each data sample vector ri onto all skewerj in
ðkÞ ðkÞ
{skewerj } to find NPPI(ri ) defined by (19.1), where the superscript (k) is
used to indicate that the PPI count of ri is obtained using the skewer set
ðkÞ
{skewerj }.
Or alternatively:
N
(ii) Perform C-IPPI via (19.4) and (19.5) for each of the sample vectors fri gi¼1
ðkÞ ðkÞ
by projecting each data sample vector ri onto all skewerj in {skewerj } to
ðkÞ
find NPPI(ri ) defined by (19.1), where the superscript (k) is used to indicate
ðkÞ
that the PPI count of ri is obtained using the kth skewer set, {skewerj }.
From either (i) or (ii) a new skewer set can be generated by finding a joint set
given by
n o n o n o
ðkþ1Þ ðkÞ ðkÞ
skewerj ¼ ri [ skewerj : ð19:7Þ
N PPI ðri Þ>0
ðk Þ
Two issues arise in using skewers when it comes to IPPI, that is, determination of
K skewers and randomness resulting from skewers.
Recall that when P-IPPI is performed, it can be implemented indefinitely as long
as the index k used to dictate the number of skewers is continuously increased one
skewer at a time. The only issue that remains for P-IPPI is when P-IPPI should be
terminated, that is, when k reaches a specifically assigned value. On the other hand,
C-IPPI, like PPI, assumes that the number of skewers, K, is fixed. It then carries out
IPPI in a causal manner for each of the data samples starting from the first data
sample, r1, continuing the second data sample, r2, and so forth, until it reaches the
last data samples, rN, at which point the process is terminated. As noted in the
previous section, both P-IPPI and C-IPPI implement two loops indexed by i and k in
reverse order. More specifically, let i and k be two indices used to keep track of data
samples to be processed and the number of skewers to be processed, respectively.
Then P-IPPI implements two loops, with an outer loop indexed by k to iterate
skewers one at a time and an inner loop indexed by i to iterate all data samples,
N
fri gi¼1 , one sample vector at a time for a given skewerk. By contrast, C-IPPI also
552 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
implements two loops with an outer loop indexed by i to iterate all data samples,
N
fri gi¼1 , one sample vector at a time and an inner loop indexed by k to iterate all the
K
K skewers, fskewerk gk¼1 , at the ith iteration for the ith data sample, ri, where the
number of skewers is fixed at K during the entire k-indexed iterative process
running k from 1 to K. Both P-IPPI and C-IPPI make use of the same index i to
iterate all data samples as well as the same index k to iterate skewers in two
different ways, where P-IPPI increases the index k one at a time indefinitely in
the outer loop, as opposed to C-IPPI, which uses the index k to run through all
K
skewers in a given fixed set of K skewers, fskewerk gk¼1 , in the inner loop. Two
main issues arise in either case: (1) skewers are randomly generated and (2) deter-
mining how many skewers is enough. In what follows, we will show that two
previously developed random PPI (R-PPI) in Chang et al. (2010c) and FIPPI in
Chang and Plaza (2006) can be used to resolve these two issues.
According to P-IPPI, the number of skewers, K, is a varying parameter that can
be used as a variable to allow P-IPPI to run indefinitely while progressively
increasing K by one. This is a tremendous advantage of P-IPPI since P-IPPI
calculates PPI counts for all data sample vectors for a given skewer, skewerk,
skewer by skewer. As a result, the PPI count of each data sample vector can be
updated as K is increased and does not have to be reimplemented repeatedly,
compared to the original PPI, which must be reimplemented to recalculate the
PPI counts of all data sample vectors. To be more precise, P-IPPI takes advantage of
previous PPI counts obtained from a smaller value of K and then updates its PPI
counts for a larger value of K until the value of K is either preset or meets a certain
stopping criterion. Interestingly, a stopping rule similar to the stopping rule used by
FIPPI can also be derived for P-IPPI as follows. Specifically, P-IPPI is terminated if
k
N PPI ðri Þ ¼ N kþ1
PPI ðri Þ, with N PPI ðri Þ > 0 for 1 i N;
k
ð19:8Þ
where NkPPI (ri) is the PPI count of the ith data sample vector ri using k skewers to
generate its PPI count and the index k used to specify the outer loop of P-IPPI is
increased by one.
Now we can further extend P-IPPI by expanding n one skewer
o specified by (19.8)
ðkÞ
to a kth skewer set denoted by skewer setðkÞ ¼ skewerj . Then the single step of
(19.8) can be replaced by a new middle iterative loop indexed by j, which is
described as follows:
Loop indexed by j n o
ðkÞ ðk Þ
For each skewerj in skewer setðkÞ ¼ skewerj
ðk Þ
skewerj N
perform IPPI to find N PPI ðri Þ for fri gi¼1 : ð19:9Þ
Note that there are three scenarios in which P-IPPI and C-IPPI can be
implemented jointly as generalizations of IPPI.
19.3 IPPI 553
A natural extension to P-IPPI and C-IPPI is to have them implemented jointly. Two
versions can be derived. One is to replace IPPI implemented in P-IPPI with C-IPPI
in each pass progressively. The resulting IPPI is called progressive causal-IPPI
(PC-IPPI). The other is to replace IPPI implemented in C-IPPI with P-IPPI for
skewer sets sample by sample. The resulting IPPI is called causal P-IPPI (CP-IPPI)
and can implement IPPI for all skewers as a real-time process rather than one
skewer at a time progressively.
Table 19.1 Various versions of IPPI using two iterative loops indexed by two parameters, i,k, and
three iterative loops indexed by three parameters, i,j,k
Corresponding IPPI
Outer loop Middle loop ( j) Inner loop algorithms
kth skewer Data samples C-IPPI (k/i)
Data samples kth skewer P-IPPI (i/k)
kth skewer_set Data samples Skewers in kth VC-IPPI (k/i/j)
skewer_set
kth skewer_set Skewers in kth Data samples GP-IPPI (k/j/i)
skewer_set
kth skewer_set Data samples Skewers in kth FIPPI (VC-IPPI) (k/i/j)
skewer_set
kth skewer_set Skewers in kth Data samples FIPPI (GP-IPPI) (k/j/i)
skewer_set
One major advantage of VC-IPPI and GP-IPPI over PPI is that they require no
prior knowledge of K. This is a significant benefit in hardware design, such as Field
Programmable Gate Array, because several recent efforts to implement PPI in
hardware design require a specific value of K for its implementation and they
cannot vary with the parameter K as its value changes (Bernabe et al. 2011; Bernabe
et al. 2013). This feature paves the way to the feasibility of implementing IPPI in
hardware chip design.
The VC-IPPI and GP-IPPI developed in the previous section are progressive
algorithms that use full band information. In what follows, we derive recursive
hyperspectral band processing (RHBP) versions of implementing IPPI progres-
sively and recursively band by band.
N
Assume that fri ðlÞgi¼1 is a data set made up of all data samples with the
first l bands, where, for each 1 i N; ri(l ) can be represented by
T T
ri ðlÞ ¼ r i1 ; . . . ; r iðl1Þ ; r il ¼ riT ðl 1Þ, r il . Also suppose that there is a given
K T
set of K skewers denoted by fsk ðlÞgk¼1 ; with sk ðlÞ ¼ sk1 ; . . . ; skðl1Þ ; skl ¼
T T
sk ðl 1Þ, skl . Then rTi (l )sk(l ) can be updated by riT ðl 1Þsk ðl 1Þ progressively
band by band and calculated recursively by riT ðlÞsk ðlÞ ¼ riT ðl 1Þsk ðl 1Þ þ r il skl
band by band.
19.4 Recursive Hyperspectral Band Processing of IPPI 555
(ii) For 1 i N
A. If riT ðlÞ^s k ðlÞ > minl1 ðkÞ and riT ðlÞ^s k ðlÞ < maxl1 ðkÞ, then
ðkÞ
N PPI ðri ðlÞÞ ¼ 0. Otherwise, continue.
ðk Þ
B. If riT ðlÞ^s k ðlÞ < minl1 ðkÞ, then N PPI ðri ðlÞÞ ¼ 1. Otherwise, continue.
ðkÞ
C. If ri ðlÞriT ðlÞ^s k ðlÞ > maxl1 ðkÞ, then N PPI ðri ðlÞÞ ¼ 1. Otherwise, continue.
D. Let i i þ 1, and go to step 1(a)(i).
(b) Let k k þ 1 and go to step 1(a).
XK ðk Þ
2. Calculate N PPI ðri ðlÞÞ ¼ N
k¼1 PPI i
r ðlÞ.
3. Let l l þ 1 and go to step 1.
The previously described RHBP-IPPI is neither a causal nor a real time process
because step 1(a) must find maxl(k) and minl(k) via (19.10) and (19.11) for all the
N
data samples fri ðlÞgi¼1 . In the following sections we can make RHBP-IPPI a
progressive algorithm using a fixed skewer set to be implemented as RHBP-P-
IPPI, where for each skewer RHBP-P-IPPI can be implemented as a real-time
process and a causal algorithm (RHBP-C-IPPI) using a fixed skewer set, which
can also be implemented as a real-time process.
K
For the first band, l ¼ 1, we randomly generate a set of K skewers, fskewerk ð1Þgk¼1 ,
where skewerk ð1Þ ¼ ðsk1 Þ is actually a scalar skl. These K scalar skewers will then
be used to update skewers across all bands. In other words, when the band number is
K
increased to l, the K-skewer set used for l is fskewerk ðlÞgk¼1 , where each
T
skewerk(l ) is actually made up of skewerk ðlÞ ¼ skewerk ðl 1Þ, skl . More
T
• Let i i þ 1.
End (Inner Loop)
• Let k k þ 1.
End (Middle Loop)
XK ðkÞ
Calculate N PPI ðri ðlÞÞ ¼ k¼1
N PPI ri ðlÞ.
• Let l l þ 1.
End (Outer Loop)
If we swap the inner and middle loops in P-IPPI, the result is C-IPPI.
ðkÞ
• N PPI ðr1 ðlÞÞ ¼ 1 for all 1 k K.
• i¼2
• l¼2 n o
• maxl1 ði; kÞ ¼ max1ji rjT ðl 1Þ^s k ðl 1Þ
n o
• minl1 ði; kÞ ¼ min1ji rjT ðl 1Þ^s k ðl 1Þ
K ð1Þ
○ Normalize each of the skewers in fskewerk ð1Þgk¼1 to a unit vector,
K ð1Þ
f^s k ð1Þgk¼1 .
• For l from 2 to L
○ Randomly generate a set of K(l ) skewers denoted by
K ðlÞ
fskewerk ðlÞ ¼ sk1 gk¼1 .
K ð1Þ
○ Normalize each of the skewers in fskewerk ð1Þgk¼1 to a unit vector,
K ð1Þ
f^s k ð1Þgk¼1 .
Middle Loop for each K(l ) with l from 2 to L
Inner Loop from i ¼ 1 to N
Initial conditions:
• N ðPPI
kÞ
ðri ðlÞÞ ¼ 0
• Find maxl1(k) by (19.10) and minl1(k) by (19.11).
• If riT ðlÞ^s k ðlÞ > minl1 ðkÞ and riT ðlÞ^s k ðlÞ < maxl1 ðkÞ, then
ðkÞ
N PPI ðri ðlÞÞ ¼ 0. Otherwise, continue.
ðkÞ
• If riT ðlÞ^s k ðlÞ < minl1 ðkÞ, then N PPI ðri ðlÞÞ ¼ 1. Otherwise, continue.
ðkÞ
• If riT ðlÞ^s k ðlÞ > maxl1 ðkÞ, then N PPI ðri ðlÞÞ ¼ 1. Otherwise, continue.
• Let i i þ 1.
End (Inner Loop)
• Let k k þ 1.
End (Middle Loop)
XK ðkÞ
Calculate N PPI ðri ðlÞÞ ¼ N
k¼1 PPI i
r ðlÞ.
Let l l þ 1.
End (Outer Loop)
Similarly, we can swap the inner and middle loops to derive an RHBP version of
C-IPPI as follows.
• For l from 2 to L
○ Randomly generate a set of K(l ) skewers denoted by
K ðlÞ
fskewerk ðlÞ ¼ sk1 gk¼1 .
K ð1Þ
○ Normalize each of the skewers in fskewerk ð1Þgk¼1 to a unit vector,
K ð1Þ
f^s k ð1Þgk¼1 .
Middle Loop from i ¼ 1 to N
Inner Loop for each K(l ) with l from 2 to L
Initial conditions:
• Randomly generate a set of K(l ) skewers denoted by
n T oKðlÞ
skewerk ðlÞ ¼ sk1 ; . . . ; skðl1Þ ; skl .
k¼1
K ðlÞ
• Normalize each of the skewers in fskewerk ðlÞgk¼1 to a unit vector,
K ðlÞ
f^s k gk¼1 .
• Calculate r1T ðlÞ^s k ðlÞ ¼ r1T ðl 1Þ^s k ðl 1Þ þ r 1l skl
• maxl ð1; kÞ ¼ r1T ðlÞ^s k ðlÞ for all 1 k K.
ðkÞ
• N PPI ðr1 ðlÞÞ ¼ 1 for all 1 k K.
• i¼2 n o
• maxl ði; kÞ ¼ max1ji rjT ðlÞ^s k ðlÞ
n o
• minl ði; kÞ ¼ min1ji rjT ðlÞ^s k ðlÞ
ðkÞ
• If minl ði; kÞ < riþ1T
ðlÞ^s k ðlÞ < maxl ði; kÞ, then N PPI ðriþ1 ðlÞÞ ¼ 0. Other-
wise, continue.
ðk Þ ðkÞ
T
• If riþ1 ðlÞ^s k ðlÞ > maxl ði; kÞ, then N PPI ðri ðlÞÞ ¼ 0 and N PPI ðriþ1 ðlÞÞ ¼ 1.
Otherwise, continue.
ðk Þ ðkÞ
T
• riþ1 ðlÞ^s k ðlÞ < minl ði; kÞ, then N PPI ðri ðlÞÞ ¼ 0 and N PPI ðriþ1 ðlÞÞ ¼ 1.
Otherwise, continue.
• Let i i þ 1.
End (Inner Loop)
• Let k k þ 1.
End (Middle Loop)
XK ðkÞ
Calculate N PPI ðri ðlÞÞ ¼ k¼1
N PPI ri ðlÞ.
Let l l þ 1.
End (Outer Loop)
560 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
Note that there are at least two significant differences between RHBP-IPPI using
a fixed skewer set and varying skewer sets. First, the former uses the same fixed
skewer set for all bands, compared to the latter, which uses skewer sets that vary
with the bands. Second, the size of the used fixed skewer set, K, for the former must
be determined a priori beforehand, while the latter can adapt skewer sets with size K
(l ) to various bands. Even in the case of K(l ) fixed at K, the skewer sets are also
different since they are randomly generated every time a band is fed in.
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
500
PPI Count Value
400
300
0
25
50
Nu
m 75
be
ro
fP 100
ro
ce
ss 125 M(1, 1)
ed
Ba 150 K(1, 1)
nd C(1, 1)
s 175 B(1, 1) es
A(1, 1) 5 Signatur
Fig. 19.3 3D plots of PPI counts versus nl for the five different signatures
19.5.1 TI Experiments
The image scene in Fig. 19.2 was processed by RHBP-IPPI. Figure 19.3 shows the
plots of PPI count versus nl defined as the first l bands used to process the five
mineral signatures A, B, C, K, and M in the TI scene, where the x-axis denotes
various values of nl, the y-axis indicates different mineral signatures, and the z-axis
dictates the PPI count value. In this figure, 500 skewers are generated and utilized in
the experiments. Since all of the 20 pure pixels for each signature in the first two
columns can be extracted simultaneously, only the top left corner pixels in the first
column for each signature are selected and marked A(1, 1), B(1, 1), C(1, 1), K(1, 1),
562 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
Fig. 19.4 Endmember candidates found by RHBP-IPPI using different values of nl: (a) nl ¼ 1,
(b) nl ¼ 3, (c) nl ¼ 10, (d) nl ¼ 27, (e) nl ¼ 189 (full bands)
and M(1, 1) for demonstration. As shown in Fig. 19.3, the magenta arrow represents
a first-time band, whereby a particular signature is identified as an endmember
candidate. For instance, the signature B(1, 1) was first identified as an endmember
candidate when nl ¼ 27, that is, the first 27 bands collected and processed.
Figure 19.4 shows the spatial locations of endmember candidates identified by
RHBP-IPPI using various values of nl; Fig. 19.4a starts with only 1 band then
progressively increases to nl ¼ 3 in Fig. 19.4b, nl ¼ 10 in Fig. 19.4c, nl ¼ 27 in
Fig. 19.4d, and finally reaches the full bands, that is, nl ¼ 189, in Fig. 19.4e, each of
which shows endmember candidates found by RHBP-IPPI in the particular transi-
tion bands specified by Fig. 19.3. In Fig. 19.4, a pixel marked by a red cross
indicates a ground truth pixel in the TI scene, a pixel marked by cyan upper triangle
highlights an endmember candidate extracted in the current band, and those
endmember candidates identified in the previous bands are marked by green circles.
Table 19.2 summarizes the five mineral signatures A, B, C, K, and M among the
endmember candidates identified by RHBP-IPPI when nl varies, with nl ¼ 1, nl ¼ 3,
nl ¼ 10, nl ¼ 27, to full bands, nl ¼ 189. As demonstrated in the table, EðnlsÞ in the
second column was identified using ns with the first l bands collected and processed,
nl where ns is defined as the number of skewers.
19.5 Synthetic Image Experiments 563
Table 19.2 Summary of mineral signatures identified by RHBP-IPPI versus nl for TI scene
nl Signatures identified by RHBP-IPPI
1 ð1Þ ð1Þ
Cð1; 1Þ 2 E500 , Kð1; 1Þ 2 E500
3 ð3Þ
Að1; 1Þ 2 E500
10 ð10Þ
Mð1; 1Þ 2 E500
27 ð27Þ
Bð1; 1Þ 2 E500
189 ð189Þ ð189Þ ð189Þ ð189Þ ð189Þ
Að1; 1Þ 2 E500 , Bð1; 1Þ 2 E500 , Cð1; 1Þ 2 E500 , Kð1; 1Þ 2 E500 , Mð1; 1Þ 2 E500
a b
6
M(1, 1)
4
A(1, 1)
3
0
0 25 50 75 100 125 150 175
Number of Processed Bands
Fig. 19.5 Comparison of IPPI and RHBP-IPPI: (a) IPPI-identified signatures; (b) orders of
signatures identified by RHBP-IPPI
Furthermore, Fig. 19.5a shows endmember candidates identified by PPI for the
TI scene using 500 skewers with all bands received and processed, but the same
results can be achieved by RHBP-IPPI using nl ¼ 27 bands. Moreover, Fig. 19.5b
plots the order of all mineral signatures A, B, C, K and M identified as endmember
candidates versus different values of nl. Clearly, we see that with only 27 bands
available, all 5 signatures were extracted.
As mentioned, 500 skewers were generated and used by RHBP-IPPI to process
the TI scene. However, how to select an appropriate number of skewers, ns, remains
an unsolved problem in the literature. To address this issue, a progressive-skewer
version of RHBP-IPPI is further proposed, called RHBP-Progressive-Skewer-IPPI
(RHBP-PS-IPPI). Figure 19.6 plots PPI counts versus various values of nl as the
number of skewers, ns, increases from 50 to 500 with step size ¼ 50 for the five
mineral signatures, where the x-axis denotes the value of nl, the y-axis shows the
number of skewers, ns, generated for the RHBP-PS-IPPI process, and the z-axis
dictates the value of the PPI count.
564 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
As shown in Fig. 19.6, the height of each bar represents the value of the PPI
count for a particular combination of nl and ns generated for the process. The
magenta upward pointing arrow indicates that a specific material was being iden-
tified for the first time as an endmember candidate in that particular band by RHBP-
PS-IPPI. For instance, signature B(1,1) is identified for the first time as an
endmember candidate by the 27th band with nl ¼ 27 available in RHBP-IPPI
150
PPI Count Value
100
50
0
0
3
25
50
Nu
m
75
be
ro
fP
100
ro
ce
ss
125 500
ed
450
400
Ba
350
nd
150 300
s
250
rs
175 150
200
of S kewe
100 Nu mber
50
0
15
PPI Count Value
10
0
0
25
27
50
Nu
m
be
75
ro
fP
100
ro
ce
ss
500
ed
125
450
Ba
400
nd
350
150
s
300
250
200 rs
ewe
175 150
er of Sk
50
100 Numb
0
Fig. 19.6 PPI counts versus nl and the number of skewers for each of five mineral signatures:
(a) A(1,1), (b) B(1,1), (c) C(1,1), (d) K(1,1), (e) M(1,1) processed by RHBP-PS-IPPI
19.5 Synthetic Image Experiments 565
600
PPI Count Value
400
200
0
0
1
25
50
Nu
m
be
75
ro
fP
100
ro
ce
ss
500
ed
125
450
Ba
400
nd
350
150 300
s
250
200 wers
175 150 Ske
er of
100 Numb
50
0
600
PPI Count Value
400
200
0
0
1
25
50
Nu
m
be
75
ro
fP
ro
100
ce
ss
ed
125 500
450
Ba
400
nd
350
s
150 300
250
200 ers
175 150 r of Skew
100 N umbe
50
0
150
PPI Count Value
100
50
0
0
10
25
50
Nu
m
be
75
ro
fP
ro
100
ce
ss
ed
125 500
Ba
450
400
nd
350
s
150 300
250
rs
200 kewe
175 150 er of S
100 Numb
50
0
using 500 skewers. However, in the RHBP-PS-IPPI process, it is further shown that
B(1,1) could be identified as an endmember candidate using only 300 generated
skewers. This means that the smaller the number of skewers, the lower the com-
putational complexity.
To further address the causal issue that only visited pixels can be used for data
processing, RHBP-C-IPPI was also developed for this purpose. Figure 19.7 plots
progressive changes in the PPI count varying np, defined as the number of first
pixels being processed, which increases from ns ¼ 2000 to 40,000 with a step
size of 2000 for the five mineral signatures, where the x-axis denotes nl, the y-
axis shows np included in the process, and the z-axis dictates the value of the PPI
count. As shown in Fig. 19.7, the height of each bar represents the value of the PPI
count for a particular combination of nl and ns. The magenta upward pointing arrow
indicates that a specific material was identified for the first time as an endmember
candidate in that particular band by RHBP-C-IPPI. For instance, signature B(1,1) is
identified for the first time as an endmember candidate by RHBP-C-IPPI using
nl ¼ 27,500 skewers and np ¼ 14,000 pixels collected.
Moreover, Fig. 19.8 also shows an RHBP-C-IPPI process, with np ranging
from ns ¼ 4000 to 40,000 and with a step size of 4000 in nl ¼ 1. As shown in the
figure, a red cross indicates a ground truth panel pixel location of the TI scene, a
cyan upper triangle highlights a ground truth pixel being identified as an
endmember candidate. The white area in the scene indicates that the pixels in
this area are not yet available for data processing. The beauty of the RHBP-C-
IPPI process is that it can be used to find weak targets in a scene. For instance,
subpixel material A in the third column can be identified as an endmember
candidate by RHBP-C-IPPI only using np ¼ 8000–24,000 pixels available but
vanishing subsequently.
19.5 Synthetic Image Experiments 567
a
PPI Count Value
150
100
50
0
3
25
Nu
50
m
be
ro
75
fP
ro
ce
100
ss
ed
Ba
125
nd
40000
38000
s
36000
34000
32000
150 30000
28000
26000
24000
22000
20000
18000 ix els
16000 ssed P
f Proce
175 14000
12000 o
Number
10000
8000
6000
4000
2000
0
b
PPI Count Value
15
10
25
27
50
Nu
m
be
75
ro
fP
ro
ce
100
ss
ed
Ba
125
40000
nd
38000
36000
s
34000
32000
150 30000
28000
26000
24000
22000
20000
18000 els
175 14000
16000
cess ed Pix
12000
of Pro
Number
10000
8000
6000
4000
2000
0
Fig. 19.7 Plots of PPI count versus nl and np for each of five mineral signatures: (a) A(1,1), (b) B
(1,1), (c) C(1,1), (d) K(1,1), (e) M(1,1) processed by RHBP-C-IPPI
568 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
c
PPI Count Value
600
400
200
0
1
25
50
Nu
m
75
be
ro
fP
100
ro
ce
ss
ed
125
40000
Ba
38000
36000
nd
34000
32000
30000
s
150 28000
26000
24000
22000
20000
18000 els
16000 ed Pix
175 12000
14000
of P rocess
Number
10000
8000
6000
4000
2000
0
d
PPI Count Value
600
400
200
0
1
25
50
Nu
m
be
ro
75
fP
ro
ce
100
ss
ed
Ba
125
40000
nd
38000
36000
s
34000
32000
150 30000
28000
26000
24000
22000
20000
18000
16000 els
175 14000 ed Pix
12000
rocess
6000
8000
10000
Numb er of P
4000
2000
0
e
PPI Count Value
150
100
50
0
10
25
Nu
50
m
be
ro
75
fP
ro
ce
100
ss
ed
Ba
125
nd
40000
38000
36000
s
34000
32000
150 30000
28000
26000
24000
22000
20000
18000
16000 ix els
175 14000 ssed P
12000
f Proce
mber o
10000
8000
4000
6000 Nu
2000
0
19.5.2 TE Experiments
Figure 19.9 shows plots of the PPI count versus nl for the five mineral signatures A,
B, C, K, and M in a TE scene, where the x-axis denotes the number of first l bands
processed, nl, the y-axis indicates different signatures, and the z-axis dictates the
PPI count. In this figure, 500 skewers were generated and used in the experiments.
Since all of the 20 pure pixels for each signature in the first two columns can be
extracted simultaneously, only the pixels in the top left corner in the first column for
each signature are selected and marked A(1,1), B(1,1), C(1,1), K(1,1), and M(1,1)
for demonstration purposes. As shown in Fig. 19.9, the magenta arrow shows the
first band in which a particular signature is identified for the first time as an
endmember candidate. For instance, the signature B(1,1) is identified for the first
time as an endmember candidate by RHBP-IPPI using nl ¼ 21.
Figure 19.10 shows the spatial locations of the endmember candidates identified
by RHBP-IPPI using different values of nl. Figure 19.10a starts with only one band,
that is, nl ¼ 1, then progressively increases to nl ¼ 2 in Fig. 19.10b, nl ¼ 9 in
Fig. 19.10c, nl ¼ 21 in Fig. 19.10d, and nl ¼ 134 in Fig. 19.10e, and it finally reaches
the full bands, nl ¼ 189, in Fig. 19.10f, each of which shows the endmember
candidates found by RHBP-IPPI in the particular transition bands specified by
Fig. 19.9. In Fig. 19.10, a red cross indicates a ground truth pixel in the TE
scene, a cyan upper triangle highlights an endmember candidate found in the
current band, and those endmember candidates identified in the previous bands
are marked by green circles.
Table 19.3 summarizes the five mineral signatures A, B, C, K, and M among the
endmember candidates identified by RHBP-IPPI when nl has varying values,
1 band, 2 bands, 9 bands, 21 bands, 134 bands, and the full 189 189 bands. As
570 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
Fig. 19.8 Example of RHBP-C-IPPI process using different np pixels in band nl ¼ 1: (a)
np ¼ 4000, (b) np ¼ 8000, (c) np ¼ 12,000, (d) np ¼ 16,000, (e) np ¼ 20,000, (f) np ¼ 24,000, (g)
np ¼ 28,000, (h) np ¼ 32,000, (i) np ¼ 36,000, (j) np ¼ 40,000 (full size of image pixels)
19.5 Synthetic Image Experiments 571
K(1, 1) Œ E (1)
500
600
500
PPI Count Value
400
300
100 (2)
A(1, 1) Œ E 500
B(1, 1) Œ E (21)
500
0
0
25
Nu 50
m
be 75
ro C(1, 1) Œ E (134)
fP 100
500
ro
ce
ss 125
ed M(1, 1)
Ba K(1, 1)
nd 150
s C(1, 1)
175 B(1, 1) es
A(1, 1) 5 Signatur
Fig. 19.9 3D plots of PPI count versus nl for the five different signatures
Fig. 19.10 Endmember candidates found by RHBP-IPPI using different values of nl: (a) nl ¼ 1,
(b) nl ¼ 2, (c) nl ¼ 9, (d) nl ¼ 21, (e) nl ¼ 134 (f) nl ¼ 189 (full bands)
572 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
a b
6
Order of Signatures Identified by RBP-PPI
C(1, 1)
5
B(1, 1)
4
M(1, 1)
3
A(1, 1)
2
K(1, 1)
1
0
0 25 50 75 100 125 150 175
Number of Processed Bands
Fig. 19.11 Comparison of IPPI and RHBP-IPPI: (a) PPI-identified signatures; (b) orders of
signatures identified by RHBP-IPPI
demonstrated in the table, EðnlsÞ in the second column was identified using ns with the
first l bands collected and processed, nl.
Furthermore, Fig. 19.11a shows the endmember candidates identified by PPI for
the TE scene using 500 skewers with all the bands received and processed where
five mineral signatures and three subpixel could be extracted as endmember
candidates. However, with only 134 bands collected and processed, that is,
nl ¼ 134, all five signatures could be extracted by RHBP-IPPI using the same
number of skewers, 500. Additionally, four subpixel could be extracted as
endmember candidates by RHBP-IPPI from the TE scene, in which case one
more additional subpixel could be found, compared to PPI (Fig. 19.11a). Moreover,
Fig. 19.11b plots the order of each mineral signature A, B, C, K, and M identified as
endmember candidates versus the number of bands processed, nl. It is clear to see
that with only 134 bands available, nl ¼ 134, all five signatures could be extracted.
As noted earlier, 500 skewers were generated and used in the RHBP-IPPI
process for the TE scene. However, selecting an appropriate number of skewers
19.5 Synthetic Image Experiments 573
150
PPI Count Value
100
50
0
0
2
25
50
Nu
m
be
75
ro
fP
ro
100
ce
ss
ed
125 500
450
Ba
400
nd
350
150
s
300
250
200 ers
175 150 er o f Skew
100 Numb
50
0
20
PPI Count Value
15
10
5
0
0
21
25
50
Nu
m
be
75
ro
fP
ro
100
ce
ss
ed
125 500
Ba
450
nd
400
350
s
150 300
250
ers
175 150
200
r o f Skew
e
100 Numb
50
0
Fig. 19.12 Plots of PPI count versus number of bands, nl and ns, processed by RHBP-PS-IPPI for
each of five mineral signatures: (a) A(1,1), (b) B(1,1), (c) C(1,1), (d) K(1,1), (e) M(1,1)
574 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
6
PPI Count Value
0
0
25
50
Nu
m
75
be
ro
fP
100
ro
ce
ss
125 500
ed
134 450
400
Ba
350
nd
150 300
s
250
200 wers
175 150 er o f Ske
100 Numb
50
0
600
PPI Count Value
400
200
0
0
1
25
50
Nu
m
be
75
ro
fP
ro
100
ce
ss
ed
125 500
Ba
450
nd
400
350
s
150 300
250
200 wers
175 150 Ske
100 Num ber of
50
0
150
PPI Count Value
100
50
0
0
9
25
50
Nu
m
be
75
ro
fP
ro
100
ce
ss
ed
125 500
Ba
450
nd
400
350
s
150 300
250
200 rs
175 150 of Skewe
100 Nu mber
50
0
As shown in the figures, the height of each bar represents the value of the PPI
count for a particular combination of the number of processed bands, nl, and the
number of skewers generated for the process. The magenta upward pointing arrow
indicates that one specific material was identified for the first time as an endmember
candidate in a particular band by RHBP-IPPI. For instance, the signature B(1,1) was
identified as an endmember candidate by RHBP-IPPI for the first time using
21 bands, that is, nl ¼ 21, with 500 skewers. Interestingly, as a matter of fact,
RHBP-PS-IPPI only needs 300 skewers to identify all five mineral signatures as
endmember candidates. Thus, the smaller the number of skewers, the better the
computational complexity. Based on our extensive experiments, since the spectrum
of material C is very similar to the background, it is difficult to identify it. This is
why it is the last material to be identified using nl ¼ 134, which also requires the
largest number of skewers.
To further address the causal issue that only visited sample vectors can be
included in data processing, a causal version of RHBP-IPPI, called RHBP-C-
IPPI, was developed. Figure 19.13 plots the progressive changes in the PPI count,
varying nl as the number of processed pixels defined as np increases from 2000 to
40,000 with a step size of 2000 for the five mineral signatures in the TE scene,
where the x-axis denotes the value of nl, the y-axis shows the number of pixels
included in the process, np, and the z-axis dictates the value of the PPI count.
As shown in Fig. 19.13, the height of each bar represents the value of the PPI
count for a particular combination of nl and np. The magenta upward pointing arrow
indicates that a specific material was identified for the first time as an endmember
candidate in the particular band by RHBP-IPPI. For instance, the signature B(1,1) is
identified as an endmember candidate by RHBP-IPPI for the first time with
576 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
a
PPI Count Value
150
100
50
0
2
25
Nu
50
m
be
ro
75
fP
ro
ce
ss
100
ed
Ba
125
nd
40000
s
38000
36000
34000
32000
150 30000
28000
26000
24000
22000
20000
18000
els
175 12000
14000
16000
cess ed Pix
10000 of Pro
4000
6000
8000
Number
2000
0
b
PPI Count Value
20
10
21
25
50
Nu
m
be
75
ro
fP
ro
100
ce
ss
ed
Ba
125
40000
nd
38000
36000
34000
s
32000
150 30000
28000
26000
24000
22000
20000
18000 els
175 14000
16000
cess ed Pix
12000 of Pro
Number
10000
8000
6000
4000
2000
0
Fig. 19.13 Plots of PPI count versus nl and ns for each of five mineral signatures: (a) A(1,1),
(b) B(1,1), (c) C(1,1), (d) K(1,1), (e) M(1,1) processed by RHBP-C-IPPI
19.5 Synthetic Image Experiments 577
c
PPI Count Value
25
50
Nu
m
75
be
ro
fP
100
ro
ce
ss
ed
125
40000
Ba
134 38000
36000
nd
34000
32000
30000
s
150 28000
26000
24000
22000
20000
18000 Pixels
14000
16000
cessed
175 12000 of Pro
6000
8000
10000
Number
4000
2000
0
d
PPI Count Value
600
400
200
0
1
25
Nu
50
m
be
ro
fP
75
ro
ce
ss
100
ed
Ba
nd
125
s
40000
38000
36000
34000
32000
150 30000
28000
26000
24000
22000
20000
18000 els
16000 ed Pix
Process
175 14000
12000
10000 of
4000
6000
8000 Number
2000
0
e
PPI Count Value
150
100
50
0
9
25
Nu
m
50
be
ro
fP
75
ro
ce
ss
ed
100
Ba
nd
125
s
40000
38000
36000
34000
32000
150 30000
28000
26000
24000
22000
20000
18000 els
16000 ed Pix
175 14000
rocess
er of P
12000
10000
6000
8000 Numb
4000
2000
0
500 skewers using 21 bands available, nl ¼ 21. In fact, RHBP-C-IPPI can identify
the five mineral signatures as endmember candidates using only the first
14,000 pixels collected, np ¼ 14,000.
Moreover, Fig. 19.14 shows an example of implementing RHBP-C-IPPI using
np ranging from 4000 to 40,000 with a step size of 4000 in band 1, that is, nl ¼ 1. As
shown in the figure, a red cross indicates a ground truth pixel location in the TI
scene, a cyan upper triangle highlights a ground truth pixel identified as an
endmember candidate. The white area in the scene indicates that the pixels in this
area are not yet available for data processing. The beauty of RHBP-C-IPPI is its
ability to capture weak targets in a scene. For instance, subpixel material A in the
third column could be identified as an endmember candidate with np from 8000 to
24,000 pixels but vanishing afterwards.
Fig. 19.14 Example of RHBP-C-IPPI process using different np pixels in band nl ¼ 1. (a)
np ¼ 4000, (b) np ¼ 8000, (c) np ¼ 12,000, (d) np ¼ 16,000, (e) np ¼ 20,000, (f) np ¼ 24,000, (g)
np ¼ 28,000, (h) np ¼ 32,000, (i) np ¼ 36,000, (j) np ¼ 40,000 (full size of image pixels)
580 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
a b
c
7000
P1
6000 P2
P3
P4
5000 P5
Radiance
4000
3000
2000
1000
0
0 20 40 60 80 100 120 140 160 180
Band
Fig. 19.15 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial
locations of the 15 panels; (c) spectra of p1, p2, p3, p4, and p5
center pixels in row i and are shown in Fig. 19.15c (Fig. 1.11). According to the
HYDICE image scene in Fig. 19.13, there are 19 R panel pixels arranged in a 5 3
matrix, where there are three panel pixels in row 1 and four panel pixels in each of
rows 2–5.
Note that the only pure panel pixels present in the data scene in Fig. 19.15 are
those in column 1 (one panel pixel in row 1 and two panel pixels in each of the other
four rows) and column 2 (one panel pixel in each of the five rows) corresponding to
the five panel signatures p1, p2, p3, p4, and p5 in Fig. 19.15c.
Figure 19.16a, b shows the plots of the PPI count versus nl for the different 19 R
panel pixels with 1000 skewers being used by RHBP-IPPI, where the x-axis
specifies nl, the y-axis is the coordinates of 19 R-panel pixels according in row-
by-row order, that is, p11, p12, p13, p211, p221, p22, p23, p311, p312, p32, p33, p411, p412,
p42, p43, p511, p521, p52, and p53, and the z-axis is the corresponding PPI count value.
19.6 Real Image Experiments 581
70
60
50
PPI Count Value
40
p511∈ E1000
(9) p52∈ E(22)
1000
p521∈ E(21)
1000
30
p42∈ E(24)
1000
20 p412∈E(34)
p221∈ E(5) 1000
1000 p312 ∈E(52)
p22∈ E1000
(9) 1000
10 (36)
p411∈ E1000
0 p32∈ E(52)
1000
0 p211∈ E(28)
1000 p311∈E(50)
1000
20
40
Nu p53
60 p52
mb (101)
p11∈ E1000 p511
p521
er p43
of 80 p42
Pro p412
ces 100 p33
p411
sed p32
Ba p312
120 p311
nd p23 ixels
s p22 anel P
140
p211
p221 19R P
p13
160 p12
170 p11
70
60
PPI Count Value
50
40
p511∈ E(9) p ∈ E(22)
1000 p ∈E(21) 52 1000
521 1000
30
p42∈ E(24)
1000
20 p412∈E(34)
p221∈E(5) 1000
1000 p312∈E1000
(52)
p22∈ E(9)
1000
10 p411∈ E(36)
1000
0 p32∈ E1000
(52)
0 p211∈E(28)
1000 p311 ∈E(50)
1000
20
40
p52
Nu 60 p521
mb p11 ∈ E(101)
1000 p511
er 80 p42
of p412
Pro p411 d)
ces 100 p32 lude
sed l Exc
-pixe
p312
Ba 120 p311
ub
nd p22
els (S
s 140 p221
l Pix
p211
Pane
ure R
160 p12
170
p11
14 P
Fig. 19.16 3D plots of PPI counts versus nl for different R panel pixels: (a) PPI count changes for
the 19 R panel pixels; (b) PPI count changes for the 14 pure R panel pixels
Figure 19.16a plots the variation in the PPI count for all of 19 R panel pixels, while
Fig. 19.16b only plots progressive changes in the PPI count for the 14 pure R panel
pixels and excludes the 5 subpixels to make the 3D graphs clearer. In these two
graphs, the height of a line indicates the value of the PPI count. The magenta arrows
in the graph show the transition bands, which provide information about the PPI
count of a particular panel pixel increasing from zero to a nonzero value and can be
used to identify endmember candidate being extracted for the first time in the
current band. For instance, the PPI count value of panel pixel p11 jumps from 0 to
1 and is identified as an endmember candidate for the first time when nl ¼ 101 bands
582 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
were received and processed. As we can see in the figure, the PPI counts of the
panel pixels in the first column are greater than the PPI counts of the one-pixel-size
panel pixels in the second column. Meanwhile, the PPI counts of the subpixel
panels in the third column remain 0 during the entire process. In other words,
subpixel panels could not be extracted as endmember candidates by RHBP-IPPI. It
is also clear that the PPI counts of both p511 and p521 are greater than those of p52,
and the PPI count of p53 is 0 during the entire process and it cannot be identified as
an endmember candidate. Also, the PPI counts of the panel pixels tended to
stabilize as more bands were added to the process. Without such progressively
changing values in the PPI counts provided by RHBP-IPPI we would not be able to
show such valuable information, which was compromised by subsequent bands.
Furthermore, Fig. 19.17 highlights the panel pixels with PPI counts greater than
0 in their transition bands, where a red cross indicates the spatial location of an R
panel pixel, a cyan upper triangle shows a panel pixel with its PPI count greater than
0 extracted in the current band, and a yellow circle marks a panel pixel found in
previous bands. As we can see, after band 101, that is, nl ¼ 101, 13 out of 19 panel
pixels were identified as endmember candidates.
Figure 19.18 conducts a comparative study between IPPI and RHBP-IPPI
applied to the HYDICE scene. Figure 19.18a shows the ground truth pixels iden-
tified as endmember candidates by IPPI, where a red cross indicates the spatial
location of an R panel pixel in the HYDICE scene, and a cyan upper triangle shows
a ground truth pixel identified as an endmember candidate. A total of 7 R panel
pixels were identified as endmember candidates. However, according to
Figs. 19.16a and 19.17, a total of 13 R panel pixels were identified by RHBP-
IPPI as endmember candidates, p11, p211, p221, p22, p311, p312, p32, p411, p412, p42,
p511, p521, and p52. Figure 19.18b further shows the order of these 13 panel pixels
found by RHBP-IPPI, where the x-axis denotes nl, the number of bands being
processed, and the y-axis represents the order of panel pixels identified as
endmember candidates. For instance, p221 represents the first panel pixel identified
by RHBP-IPPI with only the first five bands, nl ¼ 5, being processed, and p22 and
p521 are the second and third panel pixels identified after nl ¼ 9. After nl ¼ 101
bands were processed, the last panel pixel, p11, was extracted.
As noted in the preceding discussions, RHBP-IPPI can identify a set of
endmember candidates band by band progressively. Figure 19.18 plots the number
of distinct endmember candidates extracted by RHBP-IPPI as the number of
processed bands, nl, gradually increases using 1000 skewers, where RHBP-IPPI
identified a total of 373 endmember candidates for the next step in the process. As
more bands are included in the process, the number of distinct endmember candi-
dates extracted by RHBP-IPPI increases. After 106 bands are processed, nl ¼ 106,
the number of distinct endmember candidates stopped varying and achieved the
same results as when full bands were processed.
Furthermore, Fig. 19.19 show a 3D plot to demonstrate how frequently a
particular pixel was picked up by RHBP-IPPI as an endmember candidate, where
the x-axis and y-axis are used to specify the columns and rows of the spatial
locations of the extracted pixels. Note that the z-axis in Fig. 19.19 indicates the
19.6 Real Image Experiments 583
Fig. 19.17 Panel pixels with PPI counts greater than 0 after processing of nl bands: (a) nl ¼ 5, (b)
nl ¼ 9, (c) nl ¼ 21, (d) nl ¼ 22, (e) nl ¼ 24, (f) nl ¼ 28, (g) nl ¼ 34, (h) nl ¼ 36, (i) nl ¼ 50, (j) nl ¼ 52,
(k) nl ¼ 101, (l) nl ¼ 169
584 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
a b
15
p311
10
p411
p412
p211
p42
p52
5
p521
p22, p511
p221
0
0 25 50 75 100 125 150
Number of Processed Bands
Fig. 19.18 Comparative study of IPPI and RHBP-IPPI: (a) IPPI-identified signatures; (b) orders
of signatures identified by RHBP-IPPI
Number of Distinct Endmember Candidate Identified by RBP-PPI
400
350
300
250
200
150
100
50
0
0 25 50 75 100 125 150
Number of Processed Bands
Fig. 19.19 Plot of number of distinct endmember candidates versus number of processed bands,
nl, for HYDICE data
p11
200
Frequency
150 p221
100 p22
p211
50 p312
p311
0
p
0 p41232
8 p411 p42
p521
16
p52
24 p511
64
Ro 32
w 56
In 48
de 40
x 40
48 32
ex
24 Ind
mn
56 16 Colu
8
64 0
In the HYDICE experiments, 1000 skewers were generated and used by RHBP-
IPPI. However, as noted, selecting an appropriate number of skewers remains an
unresolved problem. In this case, a progressive-skewer version of RHBP-IPPI,
RHBP-PS-IPPI, was further developed to address this issue regarding how many
skewers need to be generated for RHBP-IPPI. Figure 19.20 plots progressive
changes in the PPI count versus the number of processed bands, nl, as the number
of skewers increases from 100 to 1000 with a step size of 100 for R panel pixels p11,
p211, p311, p411, and p511, where the x-axis denotes the number of bands collected
and processed, the y-axis shows the number of skewers generated for RHBP-PS-
IPPI, and the z-axis dictates the value of the PPI count.
As shown in these figures, the height of each bar represents the value of the PPI
count for a particular combination of number of processed bands, nl, and number of
skewers, ns, generated for data processing. The magenta upward pointing arrow
indicates that a specific material signature was identified as an endmember candi-
date for the first time in the band by RHBP-PS-IPPI. For instance, R panel pixel p211
is identified for the first time by RHBP-PS-IPPI using 100 skewers as an
endmember candidate when nl ¼ 28, (i.e., the first 28 bands) were processed. As a
matter of fact, RHBP-PS-IPPI could identify all panel pixels in the first column as
endmember candidates using only 100 skewers generated. Table 19.4 summarizes
the number of bands processed, nl, specifying a particular pixel that was extracted
for the first time as an endmember candidate and the minimum number of skewers
required to extract all the R panel pixels in the first column identified as endmember
candidates.
As seen from the table, to extract all the R panel pixels in the first column, a
minimum of 700 skewers were required. A smaller set of skewers generated in the
process could have reduced the computational complexity. To address the causal
586 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
Table 19.4 Summary of all R panel pixels in first column extracted by RHBP-PS-IPPI
Number of bands processed required
R panel pixels for a pixel to be first extracted Minimum number of skewers
in first column as an endmember candidate required to extract a particular pixel
p11 101 500
p211 28 100
p221 5 500
p311 50 700
p312 52 100
p411 36 700
p412 34 400
p511 9 600
p521 21 700
issue in the data collection process, RHBP-C-IPPI is further proposed. Figure 19.22
plots the changes in the PPI count with different values of nl as the number of
processed pixels, np, increases from 256 to 4096 with a step size of 256 for the 5 R
panel pixels in the first column in the HYDICE scene, where the x-axis denotes the
number of bands collected and processed, the y-axis is the number of pixels
included in the process, np, and the z-axis represents the value of the PPI count.
As shown in Fig. 19.22, the height of each bar represents the value of the PPI count
for a particular combination of the number of processed bands, nl and the pixels
included in the process, np. The magenta upward pointing arrow indicates that one
material is beginning to be identified as an endmember candidate in the band that it
first identified as an endmember candidate in RHBP-C-IPPI. For instance, R panel
pixel p411 is first identified as an endmember candidate with 36 bands available, that is,
nl ¼ 36, by RHBP-IPPI using ns ¼ 1000 skewers and 3072 pixels, that is, np ¼ 3072.
Moreover, Fig. 19.22 shows an RHBP-C-IPPI process with the number of
processed pixels ranging from 512 to 4096 with a step size of 512 in band 19. As
shown in the figure, a red cross indicates the ground truth location of an R panel
pixel in the HYDICE scene, and a cyan upper triangle highlights the ground truth
pixel identified as an endmember candidate. The white area in the scene indicates
that the pixels in this area are not yet available. The beauty of the RHBP-C-IPPI
process is that it extracts weak targets in the scene. For instance, R panel pixel p22 in
the second column can be identified as an endmember candidate with
1536–3072 pixels available, and it vanishes afterwards.
A graphical user interface (GUI) (screenshot shown in Fig. 19.24) was developed
using MATLAB’s GUIDE to aid in algorithm performance analysis. The three
images displayed at the top of the window show a color image of the scene, a
19.7 Graphical User Interface 587
a
PPI Count Value
5
4
3
2
1
0
0
25
Nu 50
m
be
ro 75
fP
ro
ce 1000
ss 100 900
ed 101
800
Ba 700
nd 125 600
s 500
400 wers
150 Ske
300 er of
200 Numb
170 100
0
b
PPI Count Value
5
4
3
2
1
0
0
25
28
50
Nu
m
be
ro 75
fP
ro 1000
ce 100
ss 900
ed 800
Ba 700
nd 125 600
s 500
400 ers
150 300 Skew
200 Num ber of
170 100
0
Fig. 19.21 Plots of PPI count versus number of processed bands, nl, and different number of
skewers: (a) p11, (b) p211, (c) p311, (d) p411, (e) p511
588 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
20
PPI Count Value
15
10
5
0
0
25
50
Nu
m
be
ro 75
fP
ro
ce 100 1000
ss 900
ed 800
Ba 700
nd 125 600
s 500
400 ers
150 300 Skew
200 Num ber of
170 100
0
d
PPI Count Value
5
4
3
2
1
0
0
25
36
Nu 50
m
be
ro 75
fP
ro
ce 1000
ss 100 900
ed 800
Ba 700
nd 125 600
s 500
400
150 wers
300
er of Ske
100
200 Numb
170
0
15
PPI Count Value
10
0
0
9
25
Nu 50
m
be
ro 75
fP
ro
ce 1000
ss 100 900
ed 800
Ba 700
nd 125 600
s 500
400
150 300 Sk ewers
er of
200 Numb
170 100
0
grayscale image of the current band being processed, and the result after the current
band is processed. In the bottom window are shown the variations in the IPPI counts
as more bands are collected for the different signatures. Once the data are loaded,
the user can start the process by clicking the start button. At each iteration, each
band is received and processed. Upon completion of processing, the current band
and the resulting image are updated to allow the user to observe the results. A final
note on Fig. 19.24 is worthwhile. The plots shown in the bottom window represent a
real-time progressive version of RHBP-IPPI. Unfortunately, this nice feature can
only be demonstrated in a real-time process.
19.8 Conclusions
a
PPI Count Value
200
100
0
0
25
50
Nu
m
be
75
ro
fP
ro
ce
100
ss
101
4096
ed
3840
3584
Ba
125 3328
3072
nd
2816
s
2560
2304
2048
150 1792
1536 xels
1280 essed Pi
1024 r of Proc
512
768 Numbe
170 256
0
b
PPI Count Value
40
20
0
0
25
28
50
Nu
m
be
ro
75
fP
ro
ce
100
ss
ed
4096
3840
Ba
3584
3328
nd
125 3072
s
2816
2560
2304
2048
150 1792 ls
1536 se d Pixe
1024
1280 of Proces
768 Number
170 512
256
0
Fig. 19.22 PPI count variations versus number of bands and number of pixels, nl, processed by
RHBP-C-IPPI for each R panel pixel: (a) p11, (b) p211, (c) p311, (d) p411, (e) p511
19.8 Conclusions 591
c
PPI Count Value
40
20
0
0
25
Nu
50
m
be
ro
fP
75
ro
ce
ss
100
ed
Ba
4096
3840
nd
3584
125 3328
s
3072
2816
2560
2304
2048
150 1792
1280
1536
es se d Pixels
1024 of Proc
512
768 Number
170 256
0
d
PPI Count Value
40
20
0
0
25
36
Nu
50
m
be
ro
fP
75
ro
ce
ss
100
ed
Ba
4096
3840
nd
3584
125 3328
s
3072
2816
2560
2304
2048
150 1792
1536 Pixels
1280 essed
768
1024
Numbe r of Proc
170 512
256
0
e
PPI Count Value
20
10
0
0
9
25
Nu
50
m
be
rof
75
P
ro
ce
ss
100
ed
Ba
4096
3840
nd
3584
125 3328
s
3072
2816
2560
2304
2048
150 1792
1536 ls
1280 sed Pixe
1024 of Proces
170 512
768 Number
256
0
Fig. 19.23 An example of RHBP-C-IPPI process using different np pixels in band nl ¼ 1: (a)
np ¼ 4000, (b) np ¼ 8000, (c) np ¼ 12,000, (d) np ¼ 16,000, (e) np ¼ 20,000, (f) np ¼ 24,000, (g)
np ¼ 28,000, (h) np ¼ 32,000, (i) np ¼ 36,000, (j) np ¼ 40,000 (full size of image pixels)
594 19 Recursive Hyperspectral Band Processing of Iterative Pixel Purity Index
data samples from band to band. In addition, the potential of RHBP-IPPI in finding
significant bands in changes in PPI counts was also demonstrated by experiments.
To further study RHBP-IPPI, we can extend RHBP-IPPI to real-time implementa-
tion for future hardware design.
Chapter 20
Recursive Band Processing of Fast
Iterative Pixel Purity Index
Abstract As noted in Chap. 19, the performance of the pixel purity index (PPI) is
largely determined by the number of skewers, K, to be used to calculate PPI counts
for data sample vectors. Recently, two approaches were investigated. One is the
iterative PPI (IPPI), developed by Chang and Wu [IEEE J Sel Top Appl Earth Obs
Remote Sens 8(6):2676–2695, 2015], where Chap. 19 develops a progressive
version of the IPPI, called progressive hyperspectral band processing of IPPI
(PHBP-IPPI), and a recursive version of IPPI, called recursive hyperspectral band
processing of IPPI (RHBP-IPPI), both of which vary with bands rather than skewer
sets according to the band-sequential (BSQ) acquisition format so that bands can be
collected band by band while data processing is taking place. Another approach is
the fast iterative PPI (FIPPI) developed by Chang and Plaza [IEEE Geosci Remote
Sens Lett 3(1):63–67, 2006] to address two major issues arising in PPI: the use of
skewers whose number must be determined a priori and inconsistent final outcomes
resulting from the use of skewers, which cannot be reproduced. FIPPI is an iterative
algorithm that iterates each process until it reaches a final set of endmembers. Most
importantly, it is an unsupervised algorithm, as opposed to the PPI, which requires
human intervention to manually select a final set of endmembers. As shown in
Chang and Plaza [IEEE Geosci Remote Sens Lett 3(1):63–67, 2006] and Chang and
Wu [IEEE J Sel Top Appl Earth Obs Remote Sens 8(6):2676–2695, 2015] both the
FIPPI and an IPPI produce very similar results, but FIPPI converges very rapidly
with significant savings in computation. Following a similar treatment derived for
RHBP-IPPI in Chap. 19, this chapter extends FIPPI to an RHBP version for FIPPI,
called recursive hyperspectral band processing of FIPPI (RHBP-FIPPI) in such a
way that data analysts can observe the progressive profiles of interband changes
among bands produced by RHBP-FIPPI. The idea of implementing RHBP-FIPPI is
to use two loops specified by skewers and bands implemented in RHBP-IPPI in
Chap. 19 to process FIPPI. Depending on which one is implemented in the outer
loop, two different versions of RHBP-FIPPI can be designed. When the outer loop
is iterated band by band, it is called RHBP-FIPPI. When the outer loop is iterated by
growing skewers, it is called recursive skewer processing of FIPPI (RSP-FIPPI).
Interestingly, both versions provide different insights into the design of FIPPI but
produce close results. Finally, note that since RHBP is implemented bandwise band
by band, it does not require dimensionality reduction like the original FIPPI.
20.1 Introduction
The pixel purity index (PPI) has been widely used for finding endmembers (Chang
2013, 2016). However, its utility has been demonstrated to be more than just for
endmember finding; for example, it can be used to find seed training samples for
multispectral image classification (Chen et al. 2013), to find feasible sets for
potential endmember candidates (Xiong et al. 2011), and so forth. One of the
major issues associated with PPI is its sensitivity to input parameters, namely
K (number of skewers) and t (cutoff threshold value) used in PPI. Another impor-
tant issue is the random procedure employed by PPI to generate so-called skewer
vectors (or skewers), which are then used to compute endmember candidates. A
third issue is its computational complexity. A fourth issue is the requirement of
human intervention to manually select a final set of endmembers by visual inspec-
tion. Most importantly, PPI is not an iterative process and does not guarantee its
convergence in finite runs, despite the fact that it may converge asymptotically, as is
claimed. To address these issues, a fast iterative algorithm for implementing PPI,
referred to as fast iterative pixel purity index (FIPPI), was developed by Chang and
Plaza (2006) and discussed in Sect. 19.3.3. It has several significant advantages
over PPI. First, it takes advantage of a recently developed concept, virtual dimen-
sionality (VD), developed by Chang (2003a) and Chang and Du (2004), to estimate
the number of endmembers, p, required to be generated. Using VD allows us to
replace the two parameters k and t used in PPI so that the algorithm’s sensitivity to
these parameters can be resolved. Second, FIPPI makes use of the automatic target
generation process (ATGP) discussed in Sect. 4.4.2.3 to generate an appropriate set
of initial endmembers that can speed up the algorithm considerably. Third, because
ATGP-generated targets used as initial endmembers are specific, there is no issue of
randomness resulting from skewers. Fourth, FIPPI is an iterative algorithm that
converges very rapidly with tremendous savings in computational time. Most
importantly, PPI requires a visualization tool to manually select a final set of
endmembers. Such a problem is avoided by FIPPI because FIPPI is automatically
terminated and the final set of FIPPI-generated endmembers is consistent. This is
considered one of the most significant advantages of FIPPI over PPI. Additionally,
FIPPI only requires a few iterations to generate its final set of endmembers
compared with PPI, which requires the parameter k to be a very high number, for
example, 104 or greater. Most significantly, FIPPI substantially reduces the com-
putational complexity via an iterative process.
Since a recursive hyperspectral band processing version of PPI, called recursive
hyperspectral band processing of PPI (RHBP-PPI), is developed in Chap. 19, it
would be highly desirable to adopt a similar treatment to derive a recursive version
of FIPPI that could also process FIPPI band by band recursively according to band
availability without waiting for full bands to be completely acquired. This chapter
explores this idea and further develops a recursive hyperspectral band processing of
FIPPI (RHBP-FIPPI) as a counterpart to RHBP-PPI using two iterative processes to
perform FIPPI with bands in the outer loop and skewers in the inner loop. If these
two loops are swapped by growing skewers in the outer loop and processing bands
20.2 FIPPI 597
in the inner loop, the resulting RHBP-FIPPI is called recursive skewer processing of
FIPPI (RSP-FIPPI). Interestingly, both versions provide different insights into the
design of FIPPI but produce similar results.
20.2 FIPPI
One of the major drawbacks of PPI is its computational complexity. For instance,
the algorithm took more than 50 min to project every data sample vector of a
350 350 pixel subset of the 224-band Cuprite AVIRIS image scene into 104
skewers on a PC with an AMD Athlon 2.6 GHz processor and 512 MB RAM. To
reduce computational complexity, most users of the ENvironment for Visualizing
Images (ENVI) software preprocess data by Minimum Noise Fraction transforma-
tion so that the original data’s dimensionality can be reduced to ease computation.
Additionally, PPI is not iterative and can only guarantee that it will produce optimal
results asymptotically. Thus, it is recommended that the algorithm be implemented
using as many skewers as possible in order to obtain optimal results. Also,
according to our experiments, PPI is also very sensitive to the initial values of
parameters K and especially t. Furthermore, ENVI’s PPI makes use of a built-in
algorithm to randomly generate a large set of so-called skewers, and users of the
ENVI’s PPI do not have a choice of their initial sets of skewers. Finally, PPI
requires a supervised procedure to manually select a final set of endmembers,
which largely depends on human interpretation. This section addresses these issues
by developing FIPPI. Interestingly, by virtue of IPPI, as discussed in Chap. 19,
FIPPI can be reinterpreted as a variant of growing skewer set P-IPPI (Sect. 19.3.3).
As noted, one of the major disadvantages resulting from implementing PPI is the
inconsistency in the final set of extracted endmembers caused by randomly generated
skewers. One way to mitigate this issue is to develop a random version of P-PPI, called
random P-IPPI (RP-IPPI), similar to RC-IPPI described in Chang (2016). Another way
to resolve this issue is to specify a set of specially generated initial endmembers by a
certain algorithm, an approach that is in direction opposition to RP-IPPI. One such
algorithm is FIPPI, which is described in detail as follows where an initial set of
endmembers is generated by a specific algorithm, ATGP, in Sect. 4.4.2.3.
FIPPI
Initial conditions: Find the VD using the Harsanyi–Farrand–Chang (HFC)
method, and let it be p (see Chap. 4).
Outer loop from l ¼ p to L
Inner loop
n op
ð0Þ
1. Initial condition: Let skewerl, j be an initial set of p skewers generated
j¼1
by selecting those pixels that correspond to target pixels generated by ATGP.
598 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
n op
ð 0Þ
2. Normalize skewerl, j into a unit vector.
j¼1
ðkÞ
3. Iterative rule: At iteration k, for each skewerl; j , all sample vectors are
ðkÞ
projected onto this particular skewerl; j to find those that are at its extreme
ðkÞ
positions to form an extrema set, denoted by Sextrema(skewerl; j ). Find the
ðkÞ
sample vectors that produce the largest NPPI(rl; j ) and let them be denoted by
ðkÞ
{rl; j }.
4. Form the joint set
n o n o
ðkþ1Þ ðk Þ ðkÞ
skewerl, j ¼ arg maxrðkÞ N PPI rl, j [ skewerl, j : ð20:1Þ
l, j
n o n o
ðkþ1Þ ðkÞ
5. If skewerl, j ¼ skewerl, j , then no new endmembers are added to the
skewer set. In this case, the algorithm is terminated, breaking the inner loop.
Otherwise, let k k þ 1, and go to step 2.
End Inner loop
End Outer loop
Note that in the original
n design of
o FIPPI developed by Chang and Plaza (2006),
ðkþ1Þ
the new skewer set, skewerl, j , is augmented by including all data sample
ðkÞ ðk Þ
vectors rl; j with N PPI rl, j > 0, that is,
n o n o n o
ðkþ1Þ ðkÞ ðkÞ
skewerl, j ¼ rl, j ðk Þ
[ skewerl, j ; ð20:2Þ
N PPI rl, j >0
ðkÞ
instead of (20.1), which only includes all data sample vectors rl; j with
ðkÞ
max NPPI(rl; j ). The reason for using (20.1) is so that the skewer sets do not grow
too fast so we can see the progressive profiles of finding endmembers from growing
skewer sets.
Figure 20.1 depicts a flowchart showing a detailed, step-by-step implementation
of FIPPI.
Note that when the algorithm is terminated
in step 4, the vectors corresponding
ðkþ1Þ ðkþ1Þ
to the pixels rj with maxN PPI rj are the desired endmembers, denoted by
n o
ðkþ1Þ
ej . Additionally, in step 1 of initialization, ATGP in Sect. 4.4.2.3 is used to
n op
ð0Þ
generate the initial set of p initial skewers, skewerj . ATGP can be replaced
j¼1
by any method, including the one used in PPI that randomly generates so-called
20.2 FIPPI 599
Pre-processing:
Find nVD = p
(0) p
Use ATGP to produce an initial set of p skewers, skewer j .
j 1
skewer skewer
( k 1) (k ) No
skewer j skewer j ?
Yes
Algorithm stops
ð0Þ
skewers, {skewerj } as long as the initialization method provides a good estimate
of initial skewers. But, as will be shown experimentally, such randomly generated
initial skewers can only slow down the algorithm.
Several major benefits can be gained from FIPPI. First, it improves computa-
tional efficiency significantly compared to the original PPI since FIPPI is an
iterative algorithm, whereas PPI is not. Second, the parameters k and t are not
required. Thus, users need not input these values, avoiding human subjectivity.
Third, the number of endmembers that must be generated, p, can be estimated by
VD, unlike with PPI, which must be carried out on a trial-and-error basis. Fourth,
FIPPI implements a replacement rule to iteratively refine each iteration and is
terminated when a set of final endmembers, {ēj} is identified via an implemented
stopping rule. Fifth and most importantly, FIPPI is fully automated and requires no
human supervision, as does PPI.
600 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
n o n o
ðkþ1Þ ðk Þ
5. If skewerl, j ¼ skewerl, j , then no more new endmembers are added
to the skewer set. In this case, the algorithm is terminated and the data are
outputted. Go to outer loop. Otherwise, let k k þ 1, and go to step 2.
End Inner loop
Let l l þ 1.
End Outer loop
Figure 20.2 depicts a flowchart of a detailed, step-by-step implementation of
RHBP-FIPPI.
Pre-processing:
Find nVD = p
(0) p
Use ATGP to produce an initial set of p skewers, skewer j .
j 1
skewer skewer
No
skewer
Yes
l l 1
Yes No
Stop l>L
The following recursive algorithm swaps the inner loop and outer loop
implemented in RHBP-FIPPI by iterating the number of the first l bands being
used, denoted by nl in the inner loop and indexed by the parameter l recursively,
while iterating the number of skewers, denoted by nskewer, in the outer loop, indexed
by the parameter k progressively. Since the outer loop produces skewer sets
progressively, the resulting algorithm is called RSP-FIPPI and is described in
what follows.
n o n o
ðk;*Þ ðk Þ
4. If skewerl, j ¼ skewerl, j , then no more new endmembers are added to
the skewer set. In this case, the algorithm is terminated and the data are
outputted. Go to outer loop. Otherwise, let l l þ 1, and go to step 2.
• End Inner loop
Let k k þ 1.
End Outer loop
20.3 Recursive Hyperspectral Band Processing of FIPPI 603
N
Find N PPI (ri ) for all data sample vectors ri l l 1
i 1
skewer skewer
No
skewer skewer
Yes
(k ,*)
Output skewer j k k 1
Note that, unlike RHBP-FIPPI, which grows skewer sets recursively in the inner
loop for each fixed nl, RSP-FIPPI grows skewer sets recursively in the inner loop as
nl increases and in the meantime also uses previously generated skewer sets as its
new initial conditions in the outer loop to progressively produce new skewer sets
augmented from previously generated skewer sets. Thus, technically speaking, it
can run indefinitely as long as the outer loop is continuously executed. Furthermore,
since RSP-FIPPI also grows skewer sets recursively in the inner loop, it can also be
considered RHBP-FIPPI. Figure 20.3 depicts a flowchart of a detailed, step-by-step
implementation of RSP-FIPPI.
Interestingly, compared with RHBP-PPI derived in Chap. 19, RHBP-FIPPI and
RSP-FIPPI are much simpler and more easily implemented. A key difference
between RHBP-PPI and RHBP-FIPPI/RSP-FIPPI lies in their required prior knowl-
edge, where RHBP-PPI needs to know the value of nskewer, unlike RHBP-FIPPI/
RSP-FIPPI, which requires the knowledge of nVD ¼ p.
604 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
The synthetic images shown in Fig. 20.4, which is also described in Figs. 1.14 and
1.15 of Chap. 1, are used to substantiate and validate the utility of C-IPPI in
applications.
Among the 25 panels are five 4 4 pure pixel panels for each row in the first
column and five 2 2 pure pixel panels for each row in the second column, five
2 2mixed pixel panels for each row in the third column and five 1 1 subpixel
panels for each row in both the fourth and fifth columns, where the mixed and
subpanel pixels were simulated according to the legends in Fig. 20.4. Thus, a total
of 100 pure pixels (80 in the first column and 20 in the second column), referred to
as endmember pixels, were simulated in the data by the five endmembers, A, B, C,
K, and M. The area marked “BKG” in the upper right corner of Fig. 20.4a was
selected to find its sample mean, that is, the average of all pixel vectors within the
area “BKG”, denoted by b and plotted in Fig. 20.4b, was used to simulate the
background (BKG) for the image scene with a size of 200 200 pixels in Fig. 20.4.
The reason for this background selection was empirical since the selected “BKG”
area seemed more homogeneous than other regions. Nevertheless, other areas could
also have been selected for the same purpose. This b-simulated image background
was further corrupted by an additive noise to achieve a signal-to-noise ratio of 20:1,
which was defined in Harsanyi and Chang (1994) as a 50 % signature (i.e., reflec-
tance/radiance) divided by the standard deviation of the noise. Once the target
pixels and background are simulated, two types of target insertion can be designed
to simulate experiments for various applications.
There are two types of synthetic images used in experiments. The first type of
target insertion is target implantation (TI), which can be simulated by inserting
clean target panels into a clean image background plus additive Gaussian noise by
replacing their corresponding background pixels.
100%
50% signal + 50% any other four
A
50% signal + 50% background
B
A second type of target insertion is target embeddedness (TE), which can be also
simulated by embedding clean target panels into a clean image background plus
additive Gaussian noise by superimposing target pixels over the background pixels.
20.4.1 TI Experiments
Since there are five pure mineral and one background signatures, we assume that
nVD ¼ 6. However, the value of nVD can also be estimated by the target-specified
virtual dimensionality (TSVD) developed in Chap. 4. Figure 20.5a shows the
endmember candidates found for TI by FIPPI with nVD ¼ 6, where pixels
highlighted by red crosses are the spatial locations of the ground truth pixels and
the pixels highlighted by cyan upward pointing triangles are endmember candidates
found by FIPPI. Figure 20.5b shows the results of applying RHBP-FIPPI to TI; the
pixels highlighted by yellow circles are endmember candidates extracted by the
previous bands, and pixels highlighted by cyan upper triangles are the endmember
candidates extracted by the current band.
Now, let nl be the number of first l bands used by RHBP-FIPPI. Then all the
pixels marked by yellow circle in Fig. 20.5b were produced by RHBP-FIPPI using
nl with l < L and the pixels marked by cyan upper triangles were produced by
RHBP-FIPPI using nL, that is, the entire set of full bands, which are exactly the
same as those pixels marked by cyan upper triangles in Fig. 20.5a. In other words,
RHBP-FIPPI produced the pixels marked by yellow circles in Fig. 20.5b that cannot
be found in Fig. 20.5a until RHBP-FIPPI reached the last band, in which case
Fig. 20.5 Comparative study between FIPPI and RHBP-FIPPI applied to TI. (a) Endmember
candidates found by FIPPI with nVD ¼ 6, (b) Endmember candidates found by RHBP-FIPPI with
nVD ¼ 6
606 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
RHBP-FIPPI used the entire full set of bands to produce an identical set of FIPPI-
extracted pixels, marked by cyan upper triangles in Fig. 20.5a, b. Accordingly,
RHBP-FIPPI can be interpreted as a slow motion of a one-shot-operation FIPPI,
where target pixels circled by yellow represent the additional endmember candi-
dates extracted by RHBP-FIPPI were uncovered during the RHBP process but not
found by FIPPI. Furthermore, a comparison of panels a and b in Fig. 20.5 shows that
a total of 81 ground truth pixels can be extracted by RHBP-FIPPI, while only
43 ground truth pixels picked up by FIPPI.
Figure 20.6a also plots the number of endmember candidates identified in each
band, and Fig. 20.6b plots the total number of distinct endmember candidates as the
a b
Number of Distinct Endmember Candidates Found by PBP-FIPPI
225
40 200
175
30 150
125
20 100
75
10 50
25
0 0
0 25 50 75 100 125 150 175 0 25 50 75 100 125 150 175
lth Band Number of Processed Bands
Fig. 20.6 Endmember candidates extracted by RHBP-FIPPI: (a) number of endmember candi-
dates found by RHBP-FIPPI in each band; (b) number of distinct endmember candidates identified
by RHBP-FIPPI versus number of processed bands, nl; (c) spatial locations of RHBP-FIPPI-
extracted endmember candidates from TI
20.4 Synthetic Image Experiments 607
M(1, 1)
4
0
0 25 50 75 100 125 150 175
Number of Processed Bands
Fig. 20.8 Endmember candidates identified by RHBP-FIPPI with different numbers of bands
processed: (a) 6 bands, (b) 15 bands, (c) 28 bands, (d) 189 bands (full bands)
As shown in Fig. 20.10b, after the first outer iteration, 82 endmember candidates
were extracted by RSP-FIPPI, after which no additional pixels were extracted as
endmembers. This is why the plot in Fig. 20.10b is flat.
20.4.2 TE Experiments
Experiments Similar to those conducted for TI were also performed for TE. Once
again we set nVD ¼ 6. Figure 20.11a, b shows endmember candidates extracted by
applying FIPPI and RHBP-FIPPI to TE, respectively, with nVD ¼ 6, where pixel
20.4 Synthetic Image Experiments 609
a b
Number of Distinct Endmembers Found by PSP-FIPPI
100 100
75 75
50 50
25 25
0 0
0 25 50 75 100 125 150 175 0 25 50 75 100 125 150 175
Number of Processed Bands Number of Processed Bands
Fig. 20.10 Plots of number of distinct endmember candidates extracted by RSP-FIPPI versus
number of bands processed, nl, in each iteration in outer loop. (a) First iteration. (b) Second
iteration
highlighted by red crosses are the spatial locations of the ground truth pixels and
pixels highlighted by cyan upper triangles are endmember candidates found by
FIPPI. Figure 20.11b shows the results of applying RHBP-FIPPI to TE; the pixels
highlighted by yellow circles are endmember candidates extracted by the previous
bands and pixels highlighted by cyan upper triangles are the endmember candidates
extracted by the current band. Furthermore, a total of 62 ground truth pixels were
extracted by RHBP-FIPPI compared with only 24 ground truth pixels extracted by
FIPPI.
610 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
Fig. 20.11 Comparative study between FIPPI and RHBP-FIPPI applied to TE. (a) Endmember
candidates found by FIPPI with nVD ¼ 6. (b) Endmember candidates found by RHBP-FIPPI with
nVD ¼ 6
a b
Number of Distinct Endmember Candidates Found by PBP-FIPPI
175
25
150
20
125
15 100
75
10
50
5
25
0 0
0 25 50 75 100 125 150 175 0 25 50 75 100 125 150 175
lth Band Number of Processed Bands
Fig. 20.12 Endmember candidates extracted by RHBP-FIPPI: (a) number of endmember candi-
dates found by RHBP-FIPPI in each band; (b) number of distinct endmember candidates identified
by RHBP-FIPPI versus number of processed bands, nl; (c) spatial locations of endmember
candidates found by RHBP-FIPPI from TE
Furthermore, RSP-FIPPI required only two iterations to terminate the outer loop.
Figure 20.16a plots the number of distinct endmember candidates extracted by
RSP-FIPPI versus the number of processed bands, nl, in each outer iteration. As
shown in Fig. 20.16b, after the first outer iteration, 94 endmember candidates were
extracted by RSP-FIPPI, after which no additional pixels were extracted as
endmembers. This is why the plot in Fig. 20.16b is flat.
612 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
B(1, 1)
4
M(1, 1)
3
0
0 25 50 75 100 125 150 175
Number of Processed Bands
Fig. 20.14 Endmember candidates identified by RHBP-FIPPI with different numbers of bands
processed, nl: (a) 6 bands, (b) 15 bands, (c) 28 bands, (d) 141 bands, (e) 189 bands (full bands)
20.5 Real Image Experiments 613
a b
100 100
Number of Distinct Endmembers Found by PSP-FIPPI
75 75
50 50
25 25
0 0
0 25 50 75 100 125 150 175 0 25 50 75 100 125 150 175
Number of Processed Bands Number of Processed Bands
Fig. 20.16 Plots of number of distinct endmember candidates extracted by RSP-FIPPI versus
number of bands processed, nl, in each iteration in outer loop. (a) First iteration. (b) Second
iteration
a b
c
7000
P1
6000 P2
P3
P4
5000 P5
Radiance
4000
3000
2000
1000
0
0 20 40 60 80 100 120 140 160 180
Band
Fig. 20.17 (a) HYDICE panel scene containing 15 panels; (b) ground truth map of spatial
locations of the 15 panels; (c) spectra of p1, p2, p3, p4, and p5
shown in Fig. 20.17c (Fig. 1.11). According to the HYDICE image scene in
Fig. 20.17b, there are 19 R panel pixels arranged in a 5 3 matrix, which has
three panel pixels in row 1 and four panel pixels in each of rows 2–5.
Note that the only pure panel pixels present in the data scene in Fig. 20.17 are
those in column 1 (one panel pixel in row 1 and two panel pixels in each of the other
four rows) and column 2 (one panel pixel in each of the five rows) corresponding to
five panel signatures, p1, p2, p3, p4, and p5, in Fig. 20.17c.
According to Chang (2003), when nVD is estimated using the HFC method to be
9, only three panel pixels in five rows can be found. However, if nVD is increased to
18, then five panel pixels in each row can be found (Chang et al. 2010a, 2011b).
Therefore, in the following sections we show the results of experiments for RHBP-
FIPPI with two cases, nVD ¼ 9 and 18.
20.5 Real Image Experiments 615
Fig. 20.18 Comparison of endmember candidates extracted by FIPPI and RHBP-FIPPI using
nVD ¼ 9: (a) FIPPI, (b) RHBP-FIPPI
Figure 20.18a, b shows endmember candidates extracted from the HYDICE scene
by FIPPI and RHBP-FIPPI using nVD ¼ 9. As shown in Fig. 20.18b, the R panel
pixels, the Y panel pixels, the endmember candidates extracted by previous bands,
and the endmember candidates extracted by the current band are highlighted by red
crosses, yellow crosses, magenta circles, and cyan upper triangles, respectively,
where RHBP-FIPPI produced the same results as FIPPI when it reached the last
band. In particular, many target pixels marked by magenta circles in Fig. 20.18b
were extracted by RHBP-FIPPI as additional endmember candidates during band-
by-band processing that had not been picked up by FIPPI. Specifically, compared to
the FIPPI, which only extracted three R panel pixels, p11, p312, and p521, RHBP-
FIPPI extracted two additional R panel pixels p221 and p311, plus 1 Y panel pixel,
p212, which had been missed by FIPPI.
Figure 20.19a further plots the number of endmember candidates extracted by
each band, and Fig. 20.19b shows the cumulative distinct endmember candidates
extracted by RHBP-FIPPI as the number of processed bands, nl, is increased. As
shown in Fig. 20.19b, a total of 57 distinct target pixels were extracted as
endmember candidates with their spatial locations shown Fig. 20.19c. When nl
94, no additional endmember candidates were extracted.
As noted in Fig. 20.18b, five R panel pixels and one Y panel pixel were extracted
as endmember candidates by RHBP-FIPPI. Figure 20.20 plots the order of these
extracted panel pixels versus the number of bands processed, nl. It clearly shows
that p521 was the first R panel pixel extracted using only nl ¼ 18 bands. After
nl 54, all six R panel pixels were eventually extracted by RHBP-FIPPI as
endmember candidates.
616 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
a b
15 60
50
12
40
9
30
6
20
3
10
0 0
0 25 50 75 100 125 150 0 25 50 75 100 125 150
lth Band Number of Processed Bands
Fig. 20.19 Endmember candidates extracted by RHBP-FIPPI: (a) number of endmember candi-
dates found by RHBP-FIPPI in each band; (b) number of distinct endmember candidates identified
by RHBP-FIPPI versus number of processed bands, nl; (c) spatial locations of RHBP-FIPPI-
extracted endmember candidates from HYDICE scene
Figure 20.21a, h further shows the spatial locations of the endmember candidates
identified by RHBP-FIPPI with different numbers of bands processed, nl, starting
from the first nine bands. In these figures, the spatial locations of the ground truth
pixels, the endmember candidates extracted by the previous bands, and the
endmember candidates extracted by the current band are highlighted by red
crosses, yellow circles, and cyan upper triangles, respectively. With nl ¼ 18
bands processed, the R panel pixel p521 was first extracted, followed by p221
20.5 Real Image Experiments 617
p311
4
p212
3
p221
2
p521
1
0
0 25 50 75 100 125 150
Number of Processed Bands
with nl ¼ 27 bands being processed. After nl 54 bands, all six R panel pixels
were eventually extracted by RHBP-FIPPI and no additional panel pixels were
extracted.
Experiments similar to those presented in Sect. 20.5.1 for nVD ¼ 9 were also
conducted for the HYDICE scene using nVD ¼ 18. Figure 20.22a, b shows
endmember candidates extracted from the HYDICE scene by FIPPI and RHBP-
FIPPI HYDICE using nVD ¼ 18, respectively, where FIPPI extracted four R panel
pixels, p11, p311, p411, and p521, plus one Y panel pixel, p212, versus six R panel
pixels, p11, p221, p311, p312, p411, and p521, plus one Y panel pixel, p212, extracted by
RHBP-FIPPI; the R panel pixels, Y panel pixels, endmember candidates extracted
by previous bands, and endmember candidates extracted by the current band are
highlighted by red crosses, yellow crosses, magenta circles, and cyan upper tri-
angles, respectively. The figure also shows that RHBP-FIPPI produced the same
results as FIPPI when it reached the last band. Figure 20.18a shows how FIPPI
extracted three R panel pixels, p11, p312, and p521, using nVD ¼ 9, whereas FIPPI
using nVD ¼ 18 extracted four R panel pixels, of which two R panel pixels, p11 and
p521, were identical, one R panel pixel, p312, was different from the R panel pixel
p311, and one more R panel pixel, p411, was also extracted. In particular, many target
618 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
Fig. 20.22 Comparison of endmember candidates extracted by FIPPI and RHBP-FIPPI using
nVD ¼ 18: (a) FIPPI, (b) RHBP-FIPPI
a b
25 110
20 90
80
70
15
60
50
10
40
30
5
20
10
0 0
0 25 50 75 100 125 150 0 25 50 75 100 125 150
lth Band Number of Processed Bands
Fig. 20.23 Results of RHBP-FIPPI: (a) number of skewers found by RHBP-FIPPI in each band;
(b) number of distinct skewers identified by RHBP-FIPPI versus number of processed bands; (c)
spatial locations of endmember candidates found by RHBP-FIPPI from HYDICE scene
Figure 20.26 plots the RSP-FIPPI results applied to the HYDICE scene using
nVD ¼ 9. As shown in the figure, the spatial locations of the ground truth pixels
and the endmember candidates extracted by the current band are highlighted by red
crosses and cyan upper triangles, respectively, where three R panel pixels were
extracted as endmember candidates by RSP-FIPPI.
20.5 Real Image Experiments 621
p312
4
p221
3
p212
2
p521
1
0
0 25 50 75 100 125 150
Number of Processed Bands
Furthermore, RSP-FIPPI required only two iterations to terminate the outer loop.
Figure 20.27a plots the number of distinct endmember candidates extracted by
RSP-FIPPI versus the number of processed bands, nl, in each outer iteration. As
shown in Fig. 20.27b, after the first outer iteration, 13 endmember candidates were
extracted by RSP-FIPPI, and then no additional pixels were extracted as
endmembers. This is why the plot in Fig. 20.27b is flat.
Figure 20.28 plots the RSP-FIPPI results applied to the HYDICE scene using
nVD ¼ 18. As shown in the figure, the spatial locations of the ground truth pixels
and the endmember candidates extracted by the current band are highlighted by red
crosses and cyan upper triangles, respectively, where four R panel pixels and one Y
panel pixel were extracted as endmember candidates by RSP-FIPPI.
Furthermore, RSP-FIPPI required only two iterations to terminate the outer loop.
Figure 20.29a plots the number of distinct endmember candidates extracted by
RSP-FIPPI versus the number of processed bands, nl, in each outer iteration. As
shown in Fig. 20.29b, after the first outer iteration, 21 endmember candidates were
extracted by RSP-FIPPI, and then no additional pixels were extracted as
endmembers. This is why the plot in Fig. 20.29b is flat.
622 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
Finally, Table 20.1 tabulates the number of panel pixels and the number of
endmember candidates extracted by FIPPI, RHBP-FIPPI, and RSP-FIPPI using
nVD ¼ 9 and 18.
20.5 Real Image Experiments 623
a b
15 15
Number of Distinct Skewers Found by PSP-FIPPI
12 12
9 9
6 6
3 3
0 0
0 25 50 75 100 125 150 0 25 50 75 100 125 150
Number of Processed Bands Number of Processed Bands
Fig. 20.27 Plots of number of distinct endmember candidates extracted by RSP-FIPPI versus
number of bands processed, nl, in each iteration in outer loop. (a) First iteration. (b) Second
iteration
624 20 Recursive Band Processing of Fast Iterative Pixel Purity Index
a b
25 25
Number of Distinct Skewers Found by PSP-FIPPI
20 20
15 15
10 10
5 5
0 0
0 25 50 75 100 125 150 0 25 50 75 100 125 150
Number of Processed Bands Number of Processed Bands
Fig. 20.29 Plots of number of distinct endmember candidates extracted by RSP-FIPPI versus
number of bands processed, nl, in each iteration in outer loop. (a) First iteration. (b) Second
iteration
Table 20.1 Number of panel pixels and number of endmember candidates extracted by FIPPI,
RHBP-FIPPI, and RSP-FIPPI using nVD ¼ 9 and 18
FIPPI RHBP-FIPPI RSP-FIPPI
Number Number of Number Number of Number Number of
of panel endmember of panel endmember of panel endmember
Algorithm pixels candidates pixels candidates pixels candidates
nVD extracted extracted extracted extracted extracted extracted
9 3 10 6 57 3 13
18 5 19 7 101 5 21
20.6 Conclusions 625
20.6 Conclusions
PPI has been around for more than two decades since being first proposed by
Boardman (1994). Despite several issues in its practical implementation, as pointed
out in the introduction and in Chang (2013, 2016), it continues to be one of the most
popular endmember finding algorithms thanks to its availability in the ENVI
software. However, the effectiveness of PPI is largely determined by the number
of skewers, K, used to calculate PPI counts for data sample vectors. To address this
issue, two approaches were investigated. One is IPPI, developed by Chang and Wu
(2015) [see Chap. 12 in Chang (2016)]. Chapter 19 presents several progressive
versions of IPPI, called recursive hyperspectral band processing of IPPI (PHBP-
IPPI), which vary with the bands rather than skewer sets. It is then further extended
to recursive hyperspectral band processing of IPPI (RHBP-IPPI). The other is the
algorithm developed by Chang and Plaza (2006), called FIPPI, which allows users
to process PPI progressively by growing skewer sets instead of fixing the number of
skewers at K as PPI does. As a result, potential endmember candidates can be
extracted according to varying skewer sets. Using an idea similar to that used in
Chap. 19 to derive RHBP-IPPI, this chapter extended FIPPI to RHBP-FIPPI, as well
as RSP-FIPPI.
Chapter 21
Conclusions
Abstract Writing a book has the great advantage over writing journal articles in
the sense that the former can set the tone and agenda for carrying out what the
author wants to deliver, as opposed to the latter, which must sacrifice certain things
to accommodate reviewers’ comments. This book comes from a long journey that
started with the author’s first book, Hyperspectral Imaging: Spectral Techniques for
Detection and Classification, published by Kluwer Academic Publishers in 2003
(Chang 2003), followed by a second book, Hyperspectral Data Processing: Algo-
rithm Design and Analysis, published by Wiley in 2013 (Chang 2013), and a third
book, Real Time Progressive Hyperspectral Image Processing: Endmember Find-
ing and Anomaly Detection, published by Springer in 2016 (Chang 2016). The
series finally concludes with this book, Real-Time Recursive Hyperspectral Sample
and Band Processing: Algorithm Architecture and Implementation, to complete a
cycle that accomplishes exactly what the author believes is important and derived
from research conducted in the Remote Sensing Signal and Image Processing
Laboratory (RSSIPL) at the University of Maryland, Baltimore County (UMBC).
The subjects covered in this book focus mainly on the design and development of
real-time recursive hyperspectral sample and band processing algorithms for dif-
ferent data acquisition formats, band-interleaved-by-sample (BIS) and band-inter-
leaved-by-pixel (BIP) (Fig. 21.1) and band-sequential (BSQ) (Fig. 21.2), so that
data can be processed progressively and recursively, at the same time that data
streams are being collected either sample by sample and line by line according to
BIS/BIP and BIL respectively or band by band according to BSQ. As a conse-
quence, the designed algorithms can be implemented recursively and progressively
in a causal and real-time manner and will have great potential in the future
hyperspectral imaging sensors operated in the space for data communication and
transmission.
21.1 Introduction
In the early days, when multispectral imagery was used for remote sensing image
processing, research was directed to land-cover/use applications where spatial
domain-based methods provided the necessary analysis tools, such as geographical
target detection technique, the automatic target generation process (ATGP) for
unsupervised target finding, the Reed–Xiaoli technique developed for an anomaly
detector along with its variants, such as the K-based anomaly detector (K-AD) and
R-based anomaly detector (R-AD), orthogonal subspace projection (OSP)-based
mixed pixel classification, fully constrained least-squares (FCLS) mixed pixel
quantification, and virtual dimensionality (VD) for estimating the number of spec-
trally distinct signatures, all of which are cutting edge techniques when they first
came out and have become well-known techniques used in hyperspectral imaging.
Ten years later, Chang (2013) continued his effort from Chang (2003) by including
new developments in hyperspectral data processing such as endmember extraction,
spectral information compression, and one-dimensional hyperspectral signal
processing plus applications to multispectral imaging and medical imaging in
magnetic resonance imaging (MRI). This book can be considered a sequel to
Chang (2003) and is also believed to be the first comprehensive book to cover
these topics extensively. In particular, endmember extraction, which at the time
attracted considerable interest in hyperspectraol data exploitation, has been covered
in great detail where several now well-known algorithms, such as SeQuential
N-FINDR (SQ N-FINDR), SuCcessive N-FINDR (SC N-FINDR), and simplex
growing algorithm (SGA), have been developed. Most recently, Chang (2016)
has changed gear to focus on real-time processing of hyperspectral imaging algo-
rithms, where most endmember extraction algorithms developed in Chang (2013)
have been rederived as endmember finding algorithms (EFAs) and real-time
processing algorithms sequentially or progressively in a causal manner when no
assumption is made about the presence of endmembers. Specifically, techniques
developed for anomaly detection in Chang (2003a), along with their versions using
local windows, are also retreated from a real-time processing perspective. The
current book can be considered a companion book of Chang (2016), where many
techniques presented in Chang (2016) are rederived as recursive hyperspectral
sample processing (RHSP) algorithms as well as recursive hyperspectral band
processing (RHBP) algorithms. Although theoretically the sequential and progres-
sive algorithms developed in Chang (2016) can be implemented in real time, it is
practically not realistic owing to the high computational complexity caused by
repeatedly calculating and updating causal signature matrices and causal covari-
ance/correlation band matrices. This is very similar to the same issue encountered
in statistical signal processing, where causal Wiener processing can be theoretically
implemented in real time, but technically it cannot because of its requirement to
process all previously visited data samples. To resolve this issue, Kalman filtering is
developed as a real-time causal process that can update results by recursive
equations via only the processed information obtained from previous data sample
vectors and innovations information obtained by new incoming data sample vectors
without reprocessing all data samples. This book takes this advantage and rederives
real-time processing algorithms developed in Chang (2016) from the viewpoint of
Kalman filtering as their recursive counterparts. Depending on the information used
for updating , two types of information processing are of interest, causal sample
processing (Parts II and III) and causal band processing (Parts IV and V).
630 21 Conclusions
Note that progressive and recursive processes are two completely different
concepts. Technically speaking, a progressive process is a process regarding how
an algorithm is implemented and carried out progressively. By contrast, a recursive
process is a process in which an algorithm takes advantage of the process’s
algorithmic recursive architecture to save data storage space and reduce computa-
tional complexity. With this interpretation, a progressive process can be considered
as a special type of data processing according to its implementation, while a
recursive process can be viewed as a process of utilizing its algorithmic structure
according to recurrence relations.
Finally, for those interested in hyperspectral imaging, Chang (2003), along with
the fundamentals in Chaps. 1–6 of Chang (2013) and Chaps. 1–5 of Chang (2016),
plus Chaps. 1–4 in this book, constitutes a very good collection of materials that
provide readers with a solid background in hyperspectral imaging algorithm design
and that will allow them to pursue advanced research in this fascinating area.
Recursive hyperspectral sample processing (RHSP) arises from the need for
processing data collected by BIS/BIP in real time (Fig. 21.1) when the same data
processing must be repeatedly implemented each time new data sample streams
come in by data acquisition.
Since data sample vectors that have already been visited and processed remain
the same during the entire data processing, they need not be reprocessed over and
over again and should be processed only when they are first visited. To address this
issue, we need to look into how algorithms are developed. In particular, two types
of sample information are used in algorithm design, sample spectral correlation-
based information and signature spectral correlation-based information.
Chap. 5 defines the causal sample correlation matrix (CSCRM) changing with r(n),
denoted by R(n), as
Xn
RðnÞ ¼ ð1=nÞ r rT;
i¼1 i i
ð21:1Þ
1
whose inverse can be calculated by R1 ðnÞ ¼ ððn 1Þ=nÞRðn 1Þ þ ð1=nÞrn rnT
through R1 ðn1 Þ and r(n) without recalculating the previous n–1 already visited data
1
sample vectors, fri gn1
i¼1 . As a result, R ðnÞ can be further shown by (5.12) to be
where R1 ðnÞ can be updated by the processed information term specified by
½ð1 1=nÞRðn 1Þ1 in the first term of (21.2), with the second term of (21.2) as
the innovations information term.
As we can see from (21.3), δnRT - CEM (rn) only needs to be updated recursively by
RT-CEM
(a) processed information specified by δn1 ðrn Þ and R1 ðn 1Þ; and (b) a new
T 1 1
incoming data sample vector, rn, and κðnÞ ¼ d R ðnÞd .
21.2.1.2 Real-Time AD
In analogy with (21.3), Chap. 6 also utilizes the same recursive equation of R1 ðnÞ
specified by (21.2) to derive the following recursive real-time version of anomaly
632 21 Conclusions
where δRT-CR-AD ðrn Þ only needs to be updated by (a) the processed information R1
ðn 1Þ and (b) new incoming data sample vectors, rn.
21.2.2.1 RHSP-ATGP
Assume that Up1 ¼ t1 tjþ1 tp1 is a matrix of size L ðp 1Þ made up of
p–1 previously found targets. Also, let tp be a new L-dimensional
target vector to be
added to form a new target matrix, Up ¼ Up1 tp ¼ t1 t2 tp1 tp . Then,
according to (7.4)–(7.7), P⊥Up can be updated by (a) the processed information
P⊥Up1 ; (b) new information tp; and (c) innovations information uep1 ¼ Up1 U#p1 tp ,
n h i o1
ep1 ¼ u
tpT u ep1
T
tp , and β ¼ tpT P⊥
Up1 tp via (7.8) as follows:
P⊥ ⊥
Up ¼ I Up Up ¼ PUp1 β u
#
ep1 u
ep1
T
2e
u p1 tpT þ tp tpT
T ð21:5Þ
¼ P⊥Up1 β u ep1 tp u ep1 tp :
21.2.2.2 RHSP-OSP
As for OSP, we assume that the signature matrix M in (4.17) is formed by p target
signatures, m1, m2, . . ., mp. Let d ¼ mp be the desired spectral signature of mp and
Up1 ¼ m1 m2 mp1 the undesired target spectral signature matrix made up of
m1 , m2 , . . . , mp1, which are the spectral signatures of the remaining p–1 undesired
targets. According to (4.20), the abundance fraction of the pth signature mp
detected by OSP is given by
T ⊥
^ OSP
α p ðrÞ ¼ d PUp1 r; ð21:6Þ
^ pRHSP-OSP ðrÞ ¼ dT P⊥
α Up r
T
¼ d T P⊥ eU eU
Up1 r βpp1 d mp m mp m ð21:7Þ
T
p
p1
p
p1
r;
n h i o1 h i
T h i
1
where β ¼ mpT P⊥
Up1 m p ¼ P ⊥
Up1 m p P ⊥
Up1 mp and
p p1
21.2.2.3 RHSP-LSMA
For LSMA we assume that the linear mixture model given by (9.1) is reexpressed as
r ¼ Mpþ1 α þ n: ð21:8Þ
1
1 n h i o1
and β ¼ T
mpþ1 I Mp MpT Mp MpT mpþ1 ¼ mpþ1
T
P⊥
Mp mpþ1 .
As noted in (21.5), (21.6), and (21.10), the major hurdle in finding these
solutions is inverting matrices, UpU#p in (21.5) and MTp Mp in (21.10). In many
practical applications, exact and precise knowledge about Mp or Up is lacking. In
this case, varying signatures in Mp and Up to find the best possible solutions is
necessary and must be obtained directly from data as data processing is ongoing. In
this case, we need repeatedly to invert such varying matrices when new data sample
vectors come in. As expected, the computing time will grow tremendously. Fortu-
nately, the recursive equations (21.5) (derived for ATGP), (21.7) (derived for OSP),
and (21.10) (derived for LSMA) not only avoid inverting matrix inverses but also
adapt to varying signature matrices Mp and Up by only updating new added
signatures without recalculating signatures already processed.
The close relationship between maximum likelihood estimation (MLE) and OSP
was first explored by Settle (1996) and Chang et al. (1998), where OSP and MLE
were shown to produce identical linear spectral unmixing results under an assump-
tion that the n in (21.8) is an additive Gaussian distribution. Under the same
Gaussian noise assumption it was also shown in Chang (2013) that α ^ LS ðrÞ in
(21.9) produced by LSMA is indeed identical to α ^ MLE
pþ1 ðrÞ given by
1
^ MLE
α pþ1 ð r Þ ¼ M T
pþ1 M pþ1
T
Mpþ1 r ¼ M#pþ1 r: ð21:12Þ
ð21:14Þ
or, alternatively,
n h io
RHSP-LS=OSP
tpþ1 ¼ arg maxmpþ1 mpþ1
T
P⊥
Mp m pþ1 : ð21:15Þ
On the other hand, as for finding determinant-based SV (DSV) in (2.1), the matrix
determinant must be calculated from the following equation:
" #
1 1 . . . 1
det
e1 e2 . . . ep
DSV e1 ; . . . ; ep ¼ ; ð21:16Þ
ðp 1Þ!
which involves significant computing time. It is even worse if (21.16) must also be
implemented repeatedly to find maximal DSV such as N-FINDR and SGA. This is
the main reason why many DSV-based EFAs suffer from the serious issue in
practical implementation. Chapters 11 and 12 resolve this issue by developing
two different approaches based on geometric SV (GSV).
One approach is recursive hyperspectral sample processing of OP-based SGA
(RHSP-OPSGA), which transforms a problem of finding the maximal DSV to one
of finding the maximal OP to the simplex formed by previously
D E found endmembers.
e
Let m1, . . ., mp be p endmembers and the hyperplane U p1 be generated by U e p1
¼ m e 3 m
e 2m e j ¼ mj m1 for all 2 j p and m
e p1 , with m e ¼ mp m1 .
⊥ p
According to (11.22)–(11.28), in Chap. 11 the maximal OP m
e p can be found by
⊥ ⊥
e p ¼
m P e
m
p ; ð21:17Þ
eU p1
where
⊥
e p ¼ arg maxer
m P
U e
r ð21:18Þ
ep1
636 21 Conclusions
and
T
P⊥
e ¼ P⊥
e β u ep u
ep1 m ep ;
ep1 m ð21:19Þ
Up U p1
1
e p1 U
e# m P⊥
ep1
with u ¼U p1 e p and β ¼ e pT
m e
ej
m . Using (21.18) we can
U p1
simply calculate the GSV of the p-vertex simplex formed by p endmembers,
m1, . . ., mp recursively by
⊥
e p GSV m1 ; . . . ; mp1 :
GSV m1 ; . . . ; mp ¼ ð1=ðp 1ÞÞm ð21:20Þ
mp1 mp1
with up1 ¼ ¼ , and the innovation information
ðmp1
T
mp1 Þ
1=2
jjmp1 jj
ep ¼ m
∇m ep m
e p1 ¼ mp mp1 ð21:23Þ
T
obtained from mp but not in mp1 . Specifically, Up1 Up1 can be simply updated by
T
Up2 and up1 up1 by
T
Xp1
Up1 Up1 ¼ i¼2
ui uiT ¼ Up2 þ up1 up1
T
: ð21:24Þ
ð21:26Þ
¼ hpGSGA GSV S m2GSGA ; . . . ; mp1
GSGA
;
Finally, it is worth noting that recursive processes arise from such a need to
avoid the exceedingly high computational complexity of repeatedly inverting
matrices, as in (21.5) and (21.10). Chapter 7 develops a recursive ATGP (RHSP-
ATGP) that makes use of recursive equations to eliminate the need for repeatedly
finding matrix inverses in (21.5). Chapter 8 extends the commonly used OSP to
RHSP-OSP, where the new signatures can be generated by a recursive equation to
form a new matrix, Up, to find OSP-detected abundance fractions by (21.7) without
actually inverting the matrix UpU#p in (21.5). Chapter 9 derives RHSP-LSMA for
LSMA to implement (21.11) without inverting the matrix MTp Mp in (21.10). With
an approach similar to RHSP-LSMA, Chap. 10 also derives RHSP-MLE, which
also implements (21.11) without inverting the matrix MTp Mp in (21.10). Finally,
Chaps. 11 and 12 redesign SGA as two new recursive versions, RHSP-OPSGA
(Chap. 11) and RHSP-GSGA (Chap. 12), both of which allow SGA to calculate
GSV recursively without involving the calculation of a determinant.
samples and significant bands. On the other hand, RHBP focuses on data to be
updated according to recursive equations specified by certain desired recurrence
relations. Thus, technically speaking, RHBP does not necessarily process data
progressively. Nevertheless, when RHBP is used in this book, it is generally
understood that RHBP is progressive in nature in its recursive process, in which
case RHBP processes data not only recursively to update data but also progressively
to produce progressive profiles of processed data.
In parallel to RHSP described in Parts II and III, we also develop its counterpart
according to the BSQ format (Fig. 21.2), called RHBP, presented in Parts IV and V,
where all algorithms designed and developed in Parts II and III also have counter-
parts in Parts IV and V, respectively.
Analogous to Parts II and III there are also two types of band information of
major interest, band spectral correlation-based information and band spectral
signature-based information. Hyperspectral imaging algorithms generally involve
using either band spectral correlations characterized by sample spectral statistics,
such as sample spectral covariance/correlation matrix used by CEM and AD target
detection algorithms, or spectral correlation characterized by signatures, such as
ATGP, OSP, or EFAs. Thus, Part IV focuses on how to implement RHBP on
sample spectral statistics-based covariance/correlation matrices causally in a pro-
gressive and recursive manner, while Part V is devoted to implementing RHBP on
signature spectral statistics-based correlation matrices causally in a progressive and
recursive manner.
N
Assume that fri gi¼1 is a set of all data sample vectors in the lth band. Let Xl
N
¼ ½r1 r2 rN1 rN be the data matrix formed by fri gi¼1 , with ri ¼ ðr i1 ; r i2 ; . . . ; r iL ÞT,
and let r(l ) be the current data streams acquired from the lth band, denoted by
rðlÞ ¼ ðr 1 ; r 2 ; . . . ; r l ÞT . Now we define a data matrix
2 3
r 11 r ðN1Þ1 r N1
6 7
6 ⋮ ⋱ ⋮ ⋮ 7
Xl ¼ ½r1 r2 rN1 rN ¼ 6
6r
7 ð21:29Þ
4 1ðl1Þ r ðN1Þðl1Þ r Nðl1Þ 7
5
r 1l r ðN1Þl r Nl
Xl1
and Xl ¼ ; with
xT ðlÞ
640 21 Conclusions
2 3
r 11 r ðN1Þ1 r N1
6 7
6 ⋮ ⋱ ⋮ ⋮ 7
Xl1 ¼ 6
6r
7; ð21:30Þ
4 1ðl2Þ r ðN1Þðl2Þ r N ðl2Þ 7
5
r 1ðl1Þ r ðN1Þðl1Þ r N ðl1Þ
T 1
R1
ll ¼ ð1=N ÞXl Xl
2 h iT 3
R1
ð Þ ð Þ Xl1 xðlÞ Rð1 Þð Þ Xl1 xðlÞ Rð1 X x ð lÞ
6 R1 l1 l1
l1
l1
l1Þðl1Þ l1
7
6 ðl1Þðl1Þ þ ð1=N Þ 7
6 xT ðlÞP⊥ T xðlÞ xT ðlÞP⊥ T xðlÞ
7
6 Xl1 Xl1 7
¼6
6 h iT
7:
7
6 R1
1 7
6 ðl1Þðl1Þ Xl1 xðlÞ 7
4
⊥
N x ðlÞPX T xðlÞ
T 5
xT ðlÞP⊥ XT
xðlÞ l1
l1
ð21:32Þ
It is the recursive equation (21.32) that is a key element in making RHBP feasible
and practically useful in many real-world applications, not only for real-time
processing but also hardware design.
21.3.1.1 RHBP-CEM
ð21:33Þ
1
where κl ¼ dT ðlÞR1 ll dðlÞ is a scalar, ν ¼ R1
ðl1Þðl1Þ Xl1 xðlÞ , and
l ðl1Þ
n h i o1
β ¼ xT ðlÞ P⊥ T xð l Þ .
l ðl1Þ X l1
21.3.1.2 RHBP-AD
Since, like CEM, AD uses a sample spectral correlation matrix, we can take
advantage of (21.32) to derive RHBP-AD similarly to how RHBP-CEM is derived
in (21.33) as follows:
2
δRHBP-R-AD ðrðlÞÞ ¼ δRHBP-R-AD ðrðl 1ÞÞ þ ð1=N Þβ Nr l η ;
l ðl1Þ l ðl1Þ
ð21:34Þ
n h i o1
where η ¼ rT ðl 1ÞR1
ðl1Þðl1Þ Xl1 xðlÞ and β ¼ xT ð l Þ P⊥ T x ð lÞ .
l ðl1Þ l ðl1Þ X l1
ð21:35Þ
where
T
Φll ¼ R1ll μðlÞ μ ðlÞRll
1
" #!
μð l 1 Þμ T
ð l 1 Þ μ ð l 1 Þμl
¼ R1
ll R1
ll ; ð21:36Þ
μl μT ðl 1Þ μ2l
642 21 Conclusions
with
" #" #
Aðl1Þðl1Þ b μðl 1Þ
R1
ll μðlÞ ¼
bT d μl
" #
Aðl1Þðl1Þ μðl 1Þ þ bμl
¼ ð21:37Þ
bT ðl 1Þμðl 1Þ þ dμl
and
" #T " #" #
μðl 1Þ Aðl1Þðl1Þ b μðl 1Þ
ρ ¼ μT ðlÞR1
ll μðlÞ ¼
l ðl1Þ μl μl ð21:38Þ
bT d
¼ μT ðl 1ÞAðl1Þðl1Þ μðl 1Þ þ 2μT ðl 1Þbμl þ dμ2l ;
n h i o1
β ¼ xT ð l Þ P⊥ T x ð lÞ ; ð21:39Þ
l ðl1Þ X l1
d ¼ Nβl|(l1), and
1
Aðl1Þðl1Þ ¼ RðTl1Þðl1Þ
T
þ ð1=N Þβ R1 1
ðl1Þðl1Þ Xl1 xðlÞ Rðl1Þðl1Þ Xl1 xðlÞ ;
l ðl1Þ
ð21:40Þ
Φll
K1 ¼ R1
ll þ : ð21:41Þ
ll
1 ρ
l ðl1Þ
As we can see from (21.35) to (21.41), calculating the inverse of the causal sample
covariance matrix Kll is not as trivial as we thought.
Part V in this section deals with another spectral correlation matrix that has been
used by many hyperspectral imaging algorithms. It is the causal signature correla-
tion matrix that must vary with new signatures added to the signature matrix to
update signature knowledge when a new band comes in. In what follows, we
describe four applications where the algorithms used employ a signature spectral
correlation matrix to update information progressively and recursively.
21.3 Recursive Hyperspectral Band Processing 643
21.3.2.1 RHBP-ATGP
P⊥
Ulj ¼ Ill Ulj Ulj
#
2 3
P⊥
Uðl1Þj Uðl1Þj ρ
l ðl1Þ
6 7
¼6
4
1 T 7
5
Uðl1Þj Uðl1Þj Uðl1Þj
T
mðlÞ 1 mT ðlÞρ
l ðl1Þ
2 3
Uðl1Þj ρ ρ T UT Uðl1Þj ρ ρ T mðlÞ
l ðl1Þ lðl1Þ ðl1Þj l ðl1Þ lðl1Þ
1 6 7
þ 4 5:
1 þ mT ðlÞρ m ðlÞρ
T
ρ T UT
m ðlÞρ
T
ρ T
mðlÞ
l ðl1Þ l ðl1Þ lðl1Þ ðl1Þj l ðl1Þ lðl1Þ
ð21:42Þ
21.3.2.2 RHBP-OSP
T ⊥
^ OSP
α p ðrl Þ ¼ dl PUlp rl ; ð21:44Þ
where
1
P⊥
Ulp ¼ I Ulp Ulp ¼ I Ulp Ulp Ulp
# T T
Ulp ð21:45Þ
and
h i1 h i 1
Uðl1Þp
T
Ulp Ulp ¼ UðTl1Þp mðlÞ
m T ðl Þ ð21:46Þ
1
¼ UðTl1Þp Uðl1Þp þ mðlÞmT ðlÞ ;
where
2 3
m11 m21 mðp1Þ1 mp1
6 7
6 m12 m22 ⋮ mðp1Þ2 mp2 7
6 7
Uðl1Þp ¼ 6 is an ðl 1Þ p
6 ⋮ ⋮ ⋱ ⋮ ⋮ 7 7
4 5
m1ðl1Þ m2ðl1Þ mðp1Þðl1Þ m
p ðl1Þ
T
matrix and mðlÞ ¼ m1l ; m2l ; . . . ; mpl is a p-dimensional vector. Finally, we can
^ OSP
derive an RHBP version of α ^ OSP
dl ðrl Þ in (21.44) in terms of α dl1 ðrl1 Þ as follows:
" #
^ RHBP
α -OSP ðr Þ ¼ d T P⊥ r ¼ d T
d l P⊥
rl1
dl l l Ulp l l1 Ulp
rl
T
^ RHBP
¼α -OSP ðr Þ d U
ðl1Þp lðl1Þ rl1 dl1 Uðl1Þp ρlðl1Þ r l
ρ T
dl1 l1 l
1
þ d l 1 mT ðlÞρ rl þ
l ðl1Þ 1 þ mT ðlÞρ
l ðl1Þ
T
dl1 Uðl1Þp ρ þ dl mT ðlÞρ ρ T UðTl1Þp rl1 þ ρ T mðlÞr l :
l ðl1Þ l ðl1Þ l ðl1Þ l ðl1Þ
ð21:47Þ
21.3.2.3 RHBP-LSMA
As noted in Chang (2007c), OSP can also be used to derive LSOSP, which yields
exactly the same results derived from the abundance-unconstrained LS (UCLS)
method. However, there are several subtle differences between OSP and the UCLS
method. One is that OSP is a detection technique designed based on signal-to-noise
ratio (SNR), while UCLS is an estimation method based on LSE. A second
difference is that OSP detects the abundance fraction of one signature at a time,
which is a scalar value, whereas UCLS estimates an abundance vector as a whole. A
third difference is more crucial because UCLS does not make use of OSP to
perform estimation. As a result, the P⊥ Ulp in (21.42) and (21.45) used to derive
21.3 Recursive Hyperspectral Band Processing 645
0 1
v v T
l ðl1Þ lðl1Þ
^ RHBP-LS ðlÞ ¼ @Ipp
α Pl1 Aα
^ LS ðl 1Þ
1 þ ρ
l ðl1Þ
0 1
v v T
l ðl1Þ lðl1Þ
þ @P1
l1
Am
e l rl ; ð21:52Þ
1 þ ρ
l ðl1Þ
646 21 Conclusions
where
T 1
v ¼ Ml1 Ml1 m e l ¼ P1 e l;
l1 m ð21:53Þ
l ðl1Þ
T 1
ρ e lT Ml1
¼m Ml1 m e lT P1
el ¼ m e lT v
el ¼ m
l1 m : ð21:54Þ
l ðl1Þ l ðl1Þ
21.3.2.4 RHBP-SGA
One of the great potential benefits of RHBP is that it addresses the data dimension-
ality issue, particularly in the application of endmember finding. Since the number
of endmembers to be found in data is assumed to be much smaller than the data
dimensionality, an EFA generally requires DR as a preprocessing method prior to
data processing. Despite the fact that many EFAs have been reported in the
literature, SV-based algorithms have received the most interest, specifically the
N-finder algorithm (N-FINDR) developed by Winter (1999a, b). Unfortunately,
N-FINDR requires excessive computing time and has three practical problems in its
implementation: determining the number of endmembers that need to be found,
finding all endmembers simultaneously as a result of an exhaustive search, and
calculating SV by finding the determinant of an ill-rank matrix. The SGA devel-
oped by Chang et al. (2006) actually solves the first two problems by growing
simplexes to find endmembers one after another instead of finding all endmembers
together at the same time, as N-FINDR was originally designed for. Thus, for SGA
to be effective, the number of endmembers must be estimated reliably. This can be
generally done by VD, developed by Chang (2003) and Chang and Du (2004).
However, the value of VD generally varies with different algorithms (Chap. 5 in
Chang 2013). In addition, Chap. 4 in this book extends VD to TSVD, which can
vary with real targets. Consequently, many possible values can result from various
VD methods. The development of SGA arose precisely out of the need which is to
allow SGA to adapt its VD values as it grows simplexes. However, SGA also runs
into the same problem as N-FINDR, which calculates SVs by a matrix determinant,
an issue already discussed in Chap. 2 in connection with the use of full bands to
calculate SVs, which requires DR. To avoid using a matrix determinant to calculate
SVs, two approaches, OP-based SGA (OPSGA) in Chap. 11 and Geometric SGA
(GSGA) in Chap. 12, were developed for this purpose. It seems that OPSGA and
GSGA have resolved all three of the aforementioned issues encountered in
N-FINDR. However, we can take another look at this issue. If we assume that
each endmember can only be accommodated by a unique well-characterized spec-
tral band, then the number of endmembers determines the number of bands required
to find endmembers. In other words, this implies that there is no need to use full
bands to find endmembers as all the current EFAs do, including OPSGA and
GSGA. The key issue is finding a spectral band that can be used as a fingerprint
of a particular endmember. An RHBP version of SGA, recursive hyperspectral band
21.4 Future Work on Hyperspectral Band Processing 647
SGA represents one important category of EFAs that use SVs with full abundance
constraints as a criterion to find endmembers. There is another important category
of EFAs that use OP without imposing any abundance constraint as a criterion to
find endmembers. Many algorithms that belong to this category can actually be
derived from the well-known pixel purity index (PPI) developed by Boardman
(1994). For example, there are two general approaches to extending PPI. One is to
redesign PPI as an iterative PPI (IPPI), as developed by Chang and Wu (2015),
where two types of IPPI are further derived as causal IPPI (C-IPPI) and progressive
IPPI (P-IPPI) (Chang 2016) (see also Chap. 19). The other is to reformulate PPI as a
sequential algorithm, such as ATGP or vertex component analysis (VCA), as shown
in Chang et al. (2013). Interestingly, by taking advantage of the concept of VD in
Chap. 4 and ATGP, Chang and Plaza (2006) further derived a fast sequential
algorithm, called fast IPPI (FIPPI) to implement PPI. In fact, using IPPI, we can
interpret FIPPI as a special version of P-IPPI with its initial condition particularly
specified by ATGP-generated target sample vectors. Most recently it was shown in
Chang (2016) and Chang et al. (2016) that ATGP, VCA, and SGA are essentially
the same algorithm. Since ATGP and SGA can be extended to their RHBP versions,
RHBP-ATGP (Chap. 15) and RHBP-SGA (Chap. 18), it would be interesting to
also extend IPPI to RHBP-IPPI in a similar manner. While this may seem straight-
forward, it indeed is easier said than done because IPPI involves one more param-
eter—a set of randomly generated vectors referred to as skewers. Chapters 19 and
20 address this issue.
No single book, including this one, can cover all the topics in one area. The subjects
presented here were selected to represent a small set of research conducted in the
area of hyperspectral imaging in the Remote Sensing Signal and Image Processing
Laboratory (RSSIPL) at the University of Maryland, Baltimore County (UMBC).
As a result, many interesting topics were unfortunately left out and could not be
included in the book. In what follows, we describe some research currently being
undertaken at the RSSIPL.
648 21 Conclusions
Multispectral imaging (MSI) has been widely used in many applications, for
example, climate change detection, environmental mentoring, forest and natural
resource management, geological surveying, agriculture land cover/use, and GIS, to
name just a few. However, owing to its low spectral/spatial resolution, MSI also has
limited applications in other areas, such as subpixel target detection, mixed pixel
classification, and quantification. To expand MSI’s capability, hyperspectral imag-
ing (HSI) has recently emerged as a new remote sensing technique that is developed
to address MSI’s drawbacks. In the early stages in the development of HSI
techniques, many MSI techniques were extended for this purpose, most notably,
maximum likelihood classification and principal components analysis. However, as
pointed out in Chang (2013), significant differences exist between MSI and HSI.
A natural extension of MSI to HSI is generally not effective since the two methods
have completely different design philosophies. Specifically, MSI is considered a
spatial domain-based analysis (referred to as literal analysis) technique, as opposed
to HSI, which is considered a spectral domain-based analysis technique (referred to
as nonliteral analysis) (Chang 2003). For example, owing to the large number of
spectral bands used by HSI sensors, dealing with the enormous amount of
hyperspectral data volumes becomes an issue, which is not an issue for MSI. In
this case, data reduction or data compression is generally used. This in turn gives
rise to new challenges, for example, how many bands or dimensions must be
determined and which bands or dimensions need be selected; neither of these issues
comes up in MSI. In addition, the very high spectral resolution provided by HSI
sensors can further reveal many subtle material substances that cannot be detected
by MSI sensors, such as endmembers, subpxiel targets, and mixed pixels. None of
the aforementioned issues can be solved by spatial domain-based techniques.
Chang (2003) is believed to be the first to apply statistical signal processing theory
to derive various signal processing techniques for addressing the issues in effective
data dimensionality by VD, subpixel target detection, and mixed pixel classifica-
tion/quantification. It was further expanded comprehensively by Chang (2013) to
address issues in endmember extraction, data compression, and reduction, as well
as real-time processing capabilities in Chang (2016). With such a rich theory
developed for HSI, can we take advantage of well-established spectral domain-
based HSI algorithms and resolve issues that cannot be resolved by MSI, specifi-
cally, target detection. To do that, a major issue is coping with the insufficient
number of bands that are only available in multispectral images. This chapter
considers that issue in multispectral target detection.
First, we need to come up with a means of expanding band dimensionality
(BD) from the original set of multispectral images. One idea is derived from a
random process that can generate statistical moments of any order by correlating
itself. Such band expansion process is referred to as a correlation band expansion
process (CBEP). Let {BCBEP
k } denote such correlated band images. The first attempt
21.4 Future Work on Hyperspectral Band Processing 649
at using CBEP was made in Ren and Chang (2000) and Chang (2013), where BD
was expanded via nonlinear correlations such as autocorrelation or cross correlation
among bands for multispectral classification. Another idea is to use band ratio
(BR) to create nonlinear band images. The use of BRs comes from the idea of
enhancing the spectral differences between bands as well as reducing the effects of
topography (Jensen 1996). It is performed by dividing one spectral band by another
to produce an image that can provide relative band intensities. Such band ratioed
images generally enhance the spectral differences between bands. Such a band
expansion process is referred to as a band ratioed expansion process (BREP), and
the resulting band set is denoted by {BBREP
k }. These two nonlinear generated band
sets are then
combined
to create
a nonlinear band expansion (NBE) set, denoted by
ΩNBE ¼ BkCBEP [ BkBREP . Adding ΩNBE to the original multispectral band set
ΩMS ¼ fBl g yields a new set of expanded multispectral band images, ΩEMS,
denoted by ΩEMS ¼ ΩNBE [ ΩMS , which should have sufficient band images for
HSI processing.
Since ΩNBE is not a new set of bands but rather created by an expanded ΩMS
through nonlinear functions, the issue of nonlinearity may come up with ΩNBE. To
further explore such nonlinearity, two processes are developed to generalize oper-
ators designed to perform data analysis. Two types of generalization are derived.
One is kernelization, which takes advantage of nonlinear kernels to design kernel-
based detectors that can solve linear nonseparability problems. The other is to
iterate detectors to filter out unwanted data to achieve better performance. For
each detector, there are two ways to perform detection depicted in Fig. 21.3: (a) the
detection band images are fed back to be included as part of NBE iteratively or
(b) the detection band images are only fed back to the detector iteratively. Such a
process is referred to as NBE iterative multispectral imaging (NBE-IMSI).
Many applications of interest can be found via NBE-IMSI, such as subpixel
detection, AD, and mixed pixel classification/quantification.
a
detector detection
MS data NBE
band image
b
detector detection
MS data NBE
band image
BS has been widely used for data reduction while retaining data integrality. It is
different from data DR in that DR requires the entire data to perform some sort of
transformation linearly or nonlinearly to reduce data volumes, compared to BS,
which completely discards unselected bands. However, in order to perform BS it
needs to know the prior knowledge about the number of bands to be selected, p. It is
then followed by solving an optimization problem to find a set of p optimal bands,
Ωp. Thus, two challenging issues arise in implementing BS. One is determining the
value of p. The other is that if the value of p changes, the set of Ωp also changes. In
other words, we cannot take advantage of previous Ωp to find a subsequent Ωq when
p < q since a different value of p produces a different set of Ωp. As a consequence,
when it comes to different applications that require different values of p, BS must
be reimplemented over and over again. Thus, BS is practically inapplicable. The
same issues also arise in DR as well. To address these dilemmas, the new concept of
a progressive process was recently introduced into DR and BS, referred to as a
progressive dimensionality reduction process (PDRP) in Chap. 20 in Chang (2013)
and progressive band dimensionality process (PBDP) in Chap. 21 in Chang (2013),
where DR and BS can be processed progressively back and forth in a backward or a
forward manner. With the interpretation of PDRP and PBDP, DR and BS were
further extended to progressive spectral dimensionality reduction in Chang and
Safavi (2011) and progressive band selection (PBS) in Chang 23 in Chang (2013).
Two new approaches are currently being investigated. One is band fusion (BF),
which fuses results obtained by different band subsets. It has great potential in
applications of satellite communication and transmission when more than one
ground station can be used to receive data from different band sets and process
data individually and separately. Then the obtained results can be further fused
together without reprocessing all the bands.
Another is called band tuning (BT), which allows users to communicate data
band by band back and forth without reimplementing BS. Using a radio as an
example, with a band considered a radio station, BT is similar to allowing a listener
to tune the radio to find a station of interest, as opposed to BS, which only allows
listeners to preselect particular radio stations by pushing preset buttons. Accord-
ingly, BT can be considered a contiguous band process, unlike BS, which can be
considered a discrete band process. Thus, BT is more appropriate for hyperspectral
data that are also acquired by contiguous spectral bands. Two types of BT can be
developed for the purpose of data communication. One is sequential BT (SBT),
which finds a particular band of interest that is crucial in a specific application. This
in turn leads to another new concept, band determination by tuning (BDT), which is
21.4 Future Work on Hyperspectral Band Processing 651
also different from BS in the sense that BDT finds bands of interest via tuning
without prior knowledge, compared to BS, which requires prior knowledge of p and
selects p optimal bands via optimization. Consequently, technically speaking, BF is
used for band determination, not for BS. Another type of BF is to implement BF in
conjunction with PSDP to further develop progressive BF (PBT), in which case
previously tuned bands are always included as part of the band set to be tuned later.
If such PBT is implemented via band prioritization and band decorrelation (Chang
et al. 1999), the resulting PBT turns out to be PBS (Chang and Liu 2014).
BT has several advantages over BS. First and foremost is BDT, which finds bands
of interest without prior knowledge via tuning, and in particular, without having to
know the value of p. The band to be found is determined by a specific application, not
criteria for BS. Second, BT is performed back and forth via tuning by a custom-
designed recursive equation. It is not band prioritization used by PBDP or BS used by
solving an optimization problem, both of which must be done prior to BS. Third, BF
has potential applications in satellite communication, where BF can be performed by
ground stations individually via data downlinks, and the results can be updated via
recursive equations by incoming bands without reprocessing already included bands.
Fourth, BT can be implemented as SBT so that a band can be ranked according to a
specific application in the same way that features are identified in Chang et al. (1999).
Finally, BT can also be implemented as PBS, where bands can be selected back and
forth via band prioritization in a progressive manner without reprocessing BS as the
number of bands to be selected, the value of p, changes.
Appendix A: Matrix Identities
1
where U# ¼ UT U U and
n h 1 i o 1 1
β ¼ dT I U UT U UT d ¼ dT P⊥
U d .
(d) Matrix-Inverse Identity
n h 1 i o1 T ⊥ 1
β ¼ dT I U UT U UT d ¼ d PU d ; ðA:4Þ
1
where Γ ¼ C BT A1 B .
A
AD Anomaly detection, Chap. 6
ALMM Adaptive linear mixing model, Chap. 9
ALSMA Adaptive linear spectral mixture analysis, Chap. 9
ANC Abundance nonnegativity constraint, Chap. 9
ARHBP Adaptive recursive hyperspectral band processing, Chap. 9
ASC Abundance sum-to-one constraint, Chap. 9
ATGP Automatic target generation process, Chap. 4
AVIRIS Airborne visible/infrared imaging spectrometer, Chap. 1
C
CBR-AD Causal band K-RXD, Chap. 14
CBR-AD Causal band R-RXD, Chap. 14
CBRCM Causal band correlation matrix, Chaps. 13 and 14
CEM Constrained energy minimization, Chap. 5
CK-AD Causal K-AD, Chap. 14
CLCRM Causal line correlation matrix, Chap. 5
CR-AD Causal R-AD, Chap. 6
CSCRM Causal sample correlation matrix, Chap. 5
CSCVM Causal sample covariance matrix, Chaps. 5 and 6
D
DSV Determinant-based simplex volume, Chap. 2
Dist-SGA Distance-based simple growing algorithm
DSGA Determinant-based SGA, Chaps. 2, 11, and 12
DR Dimensionality reduction
E
EFA Endmember finding algorithm
EIDA Endmember identification algorithm
F
FCLS Fully constrained least-squares method, Chap. 9
FPGA Field Programmable Gate Array
G
GSGA Geometric simplex growing algorithm, Chaps. 12 and 18
GSVA Growing simplex volume analysis, Chaps. 12 and 18
GSV Geometric simplex volume, Chap. 2
GSV-OP Geometric simplex volume by orthogonal projection, Chaps. 11 and 12
GSV-SH Geometric simplex volume by simplex height, Chap. 12
GSV-PD Geometric simplex volume by perpendicular distance, Chaps. 11 and 12
H
HFC Harsanyi–Farrand–Chang, Chap. 4
HOS High-order statistics
HSI Hyperspectral imaging
HYDICE HYperspectral Digital Imagery Collection Experiment, Chap. 1
I
IBSI Interband spectral information, Chap. 4
IPPI Iterative pure pixel index, Chap. 18
K
KF-OSP-GSGA Kalman filter–based orthogonal subspace projection geometric
simplex growing algorithm, Chap. 12
KF-OVP-GSGA Kalman filter–based orthorgonal vector projection geometric
simplex growing algorithm, Chap. 12
K-AD Anomaly detection using autocovariance matrix K, Chap. 5
L
LCMV Linearly constrained minimum variance, Chap. 5
LCVF Lunar Crater Volcanic Field, Chap. 1
LSE Least-squares error
LSU Linear spectral unmixing
LSMA Linear spectral mixture analysis, Chap. 9
M
MEAC Minimum estimated abundance covariance, Chap. 10
MLE Maximum likelihood estimation, Chap. 10
MVT Minimum volume transform
N
NCLS Nonnegativity constrained least-squares method, Chaps. 4 and 9
N-FINDR N-Finder algorithm
NPD Neyman–Pearson detection/detector
NWHFC Noise-whitened Harsanyi–Farrand–Chang, Chap. 4
Glossary 657
O
OP Orthogonal projection
OPSGA Orthogonal projection–based simple growing algorithm, Chap. 11
OSP Orthogonal subspace projection
P
P-AD Progressive anomaly detection, Chap. 6
PHBP Progressive hyperspectral band processing, Chaps. 14–20
P-CEM Progressive constrained energy minimization, Chap. 5
PKP Progressive skewer set processing, Chaps. 19 and 20
PPI Pixel purity index
PSP Progressive sample processing, Chap. 1
R
R-AD RXD using autocorrelation matrix R, Chap. 5
RHBP Recursive hyperspectral band processing, Chap. 1
RHBP-AD Recursive hyperspectral band processing of anomaly detection, Chap. 14
RHBP-ATGP Recursive hyperspectral band processing of the automatic target
generation process, Chap. 15
RHBP-CEM Recursive hyperspectral band processing of constrained energy
minimization, Chap. 13
RHBP-C-IPPI Recursive hyperspectral band processing of causal iterative pixel
purity index, Chap. 19
RHBP-FIPPI Recursive hyperspectral band processing of fast iterative pixel
purity index, Chap. 20
RHBP-GSGA Recursive hyperspectral band processing of geometric simplex
growing algorithm, Chap. 18
RHBP-GSVA Recursive hyperspectral band processing of growing simplex vol-
ume analysis, Chap. 18
RHBP-LSMA Recursive hyperspectral band processing of linear spectral mixture
analysis, Chap. 17
RHBP-OSP Recursive hyperspectral band processing of orthogonal subspace
projection, Chap. 16
RHBP-PS-IPPI Recursive hyperspectral band processing of progressive-skewer
iterative pixel purity index, Chap. 19
RHBP-SGA Recursive hyperspectral band processing of simplex growing algo-
rithm, Chap. 18
RHSP Recursive hyperspectral sample processing, Chap. 1
RHSP-ATGP Recursive hyperspectral sample processing of the automatic target
generation process, Chap. 7
RHSP-GSGA Recursive hyperspectral sample processing of geometric simplex
growing algorithm, Chap. 12
RHSP-LSMA Recursive hyperspectral sample processing of linear spectral mix-
ture analysis, Chap. 9
658 Glossary
Acito, N., M. Diani, and G. Corsini. 2009. A new algorithm for robust estimation of the signal
subspace in hyperspectral images in presence of rare signal components. IEEE Transactions on
Geoscience and Remote Sensing 47(11): 3844–3856.
———. 2010. Hyperspectral signal subspace identification in the presence of rare signal compo-
nents. IEEE Transactions on Geoscience and Remote Sensing 48(4): 1940–1954.
Adams, J.B., and M.O. Smith. 1986. Spectral mixture modeling: a new analysis of rock and soil
types at the Viking lander 1 suite. Journal of Geophysical Research 91(B8): 8098–8112.
Adams, J.B., M.O. Smith, and A.R. Gillepie. 1989. Simple models for complex natural surfaces: a
strategy for hyperspectral era of remote sensing. In Proceedings of IEEE international geo-
science and remote sensing symposium ‘89, 16–21.
———. 1993. Image spectroscopy: interpretation based on spectral mixture analysis. In Remote
geochemical analysis: elemental and mineralogical composition, ed. C.M. Pieters and
P.A. Englert, 145–166. Cambridge: Cambridge University Press.
Ambikapathi, A., T.H. Chan, C.-Y. Chi, and K. Keizer. 2013. Hyperspectral data geometry-based
estimation of number of endmembers using p-norm-based pure pixel identification algorithm.
IEEE Transactions on Geoscience and Remote Sensing 51(5): 2753–2769.
Anderson, T.W. 1984. An introduction to multivariate statistical analysis, 2nd ed. New York:
Wiley.
Ashton, E.A., and A. Schaum. 1998. Algorithms for the detection of sub-pixel targets in multi-
spectral imagery. Photogrammetric Engineering and Remote Sensing 64(7): 723–731.
Barjoski, P. 2009. Does virtual dimensionality work in hyperspectral images? In Algorithms and
technologies for multispectral, hyperspectral, ultraspectral imagery XV, SPIE, vol. 7334,
73341J-1–73341J-11, 16–19 April, Orlando, FL.
Basedow, R., P. Silverglate, W. Rappoport, R. Rockwell, D. Rosenberg, K. Shu, R. Whittlesey,
and E. Zalewski. 1992. The HYDICE instrument design. In Proceedings of international
symposium on spectral sensing research, vol. 1, 430–445.
Bates, C., and B. Curtis. 1996. A method for manual endmember selection and spectral unmixing.
Remote Sensing of Environment 55: 229–243.
Bates, C., A.P. Asner, and C.A. Wessman. 2000. Endmember bundles: a new approach to
incorporating endmember variability into spectral mixture analysis. IEEE Transactions on
Goescience and Remote Sensing 38: 1083–1094.
Bauman, J., E. Blasch, J. Jackson, and G. Sterling. 2005. Real-time ROC: a near-real-time
performance evaluation tool. In SPIE 05, vol. 5807, 380–390.
Bayliss, J., J.A. Gualtieri, and R.F. Cromp. 1997. Analyzing hyperspectral data with independent
component analysis. In Proceedings of SPIE, vol. 3240, 133–143.
Behrens, R.T., and L.L. Scharf. 1994. Signal processing applications of oblique projections
operators. IEEE Transactions on Signal Processing 42(6): 1413–1423.
Berman, M., H. Kiiveri, R. Lagerstrom, A. Ernst, R. Dunne, and J.F. Huntington. 2004. ICE: a
statistical approach to identifying endmembers in hyperspectral images. IEEE Transactions on
Geoscience and Remote Sensing 42(10): 2085–2095.
Bernabe, S., S. Lopez, A. Plaza, R. Sarmiento, and P.G. Rodriguze. 2011. FPGA design of an
automatic target generation process for hyperspectral image analysis. In 2011 I.E. 17th
international conference on parallel and distributed systems (ICPADS), 7–9 December,
Tainan, Taiwan, 1010–1015.
Bernabe, S., S. Lopez, A. Plaza, and R. Sarmiento. 2013. GPU Implementation of an automatic
target detection and classification algorithm for hyperspectral image analysis. IEEE Geosci-
ence and Remote Sensing Letters 10(2): 221–225.
Bioucas-Dias, J.M., and José Nascimento. 2005. Estimation of signal subspace on hyperspectral
data. In Proceedings of SPIE, vol. 5982, Bruges, Belgium, 191–198.
———. 2008. Hyperspectral subspace identification. IEEE Transactions on Geoscience and
Remote Sensing 46(8): 2435–2445.
Bioucas-Dias, J.M., A. Plaza, N. Dobigeon, M. Parente, Q. Du, P. Gader, and J. Chanussot. 2012.
Hyperspectral unmixing overview: geometrical, statistical, and sparse regression-based
approaches. IEEE Journal of Selected Topics in Applied Earth Observations and Remote
Sensing 5(2): 354–379.
Boardman, J.W. 1989. Inversion of imaging spectrometry data using singular value decomposi-
tion. In Proceedings of IEEE symposium on geoscience and remote sensing, 2069–2072.
———. 1990. Inversion of high spectral resolution data. Proceedings of SPIE 1298: 222–233.
———. 1993. Automated spectral unmixing of AVIRIS data using convex geometry concepts. In
Summaries, fourth JPL airborne geoscience workshop, JPL Publication 93-26, vol. 1, 11–14.
———. 1994. Geometric mixture analysis of imaging spectrometry data. In International geosci-
ence and remote sensing symposium, vol. 4, 2369–2371.
———. 1998. Leveraging the high dimensionality of AVIRIS data for improved sub-pixel target
unmixing and rejection of false positive: mixture tuned matched filtering. In Summaries of
seventh annual JPL earth science workshop, JPL Publication 98-4, vol. 1.
Boardman, J.W., F.A. Kruse, and R.O. Green. 1995. Mapping target signatures via partial
unmixing of AVIRIS data. In Summaries, fifth JPL airborne geoscience workshop, JPL
Publication 23-26.
Boker, L., S.R. Rotman, and D.G. Blumberg. 2008. Coping with mixtures of backgrounds in a
sliding window anomaly detection algorithm. In Proceedings of SPIE, vol. 7113, 711315-
1–711315-12.
Bowles, J.P., and D.B. Gilles. 2007. An optical real-tine adaptive spectral identification system. In
Hyperspectral data exploitation, ed. C.-I Chang, 77–106.
Bro, R., and S.D. Jong. 1997. A fast non-negativity-constrained least squares algorithm. Journal of
Chemometrics 11: 393–401.
Broadwater, J., and A. Banerjee. 2009. A Neyman-Pearson approach to estimating the number of
endmembers. In Proceedings of IEEE IGARSS, vol. 4, IV-693–IV-696.
Broadwater, J., J.R. Chellappa, A. Banerjee, and P. Burlina. 2007. Kernel fully constrained least
squares abundance estimates. In 2007 I.E. international geoscience and remote sensing
symposium (IGARSS), 23–28 July, 4041–4044.
Brumbley, C. 1998. Kalman Filtering and subspace projection approaches to multispectral and
hyperspectral image classification. Baltimore County, MD: Department of Computer Science
and Electrical Engineering, University of Maryland.
Brumbley, C., and C.-I Chang. 1999. An unsupervised vector quantization-based target signature
subspace projection approach to classification and detection in unknown background. Pattern
Recognition 32(7): 1161–1174.
Cawse, K., M. Sears, A. Robin, S.B. Damelin, K. Wessels, F. van den Bergh, and R. Mathieu.
2010. Using Random Matrix Theory to determine the number of endmembers in a
References 661
———. 2007c. Information-processed matched filters for hyperspectral target detection and
classification. In Hyperspectral data exploitation: theory and applications, ed. C.-I Chang,
47–74. New York: Wiley.
———. 2008a. Hyperspectral imaging: an emerging technique in remote sensing. In International
symposium spectral sensing research (ISSSR), June 23–27, Steven Institute of Technology, NJ.
———. 2008b. Unsupervised linear hyperspectral unmixing. In International symposium spectral
sensing research (ISSSR), June 23–27, Steven Institute of Technology, NJ.
———. 2008c. Three dimensional receiver operating characteristic (3D ROC) analysis for
hyperspectral signal detection and estimation. In International symposium spectral sensing
research (ISSSR), June 23–27, Steven Institute of Technology, NJ.
———. 2009. Virtual dimensionality for hyperspectral imagery. SPIE Newsroom, September 28.
doi:10.1117/2.1200909.1749.
———. 2010. Multiple-parameter receiver operating characteristic analysis for signal detection
and classification. IEEE Sensors Journal 10(3): 423–442 (Invited paper).
———. 2013. Hyperspectral data processing: algorithm design and analysis. Hoboken: Wiley.
———. 2016. Real time progressive hyperspectral image processing: endmember finding and
anomaly detection. New York: Springer.
Chang, C.-I, and C. Brumbley. 1997. An orthogonalization target signature space projection
approach to image classification in unknown background. In 31st Conference on information
sciences and systems, The Johns Hopkins University, 174–178.
———. 1999a. A Kalman filtering approach to multispectral image classification and detection of
changes in signature abundance. IEEE Transactions on Geoscience and Remote Sensing 37(1):
257–268.
———. 1999b. Linear unmixing Kalman filtering approach to signature abundance detection,
signature estimation and subpixel classification for remotely sensed images. IEEE Transac-
tions on Aerospace and Electronics Systems 37(1): 319–330.
Chang, C.-I, and S.-S. Chiang. 2002. Anomaly detection and classification for hyperspectral
imagery. IEEE Transactions on Geoscience and Remote Sensing 40(2): 1314–1325.
Chang, C.-I, and Q. Du. 1999a. A noise subspace projection approach to determination of intrinsic
dimensionality for hyperspectral imagery. In EOS/SPIE symposium on remote sensing, con-
ference on image and signal processing for remote sensing V, SPIE, vol. 3871, Florence, Italy,
September 20–24, 34–44.
———. 1999b. Interference and noise adjusted principal components analysis. IEEE Transactions
on Geoscience and Remote Sensing 37(5): 2387–2396.
———. 2004. Estimation of number of spectrally distinct signal sources in hyperspectral imagery.
IEEE Transactions on Geoscience and Remote Sensing 42(3): 608–619.
Chang, C.-I, and D. Heinz. 2000. Constrained subpixel detection for remotely sensed images.
IEEE Transactions on Geoscience and Remote Sensing 38(3): 1144–1159.
Chang, C.-I, and M. Hsueh. 2006. Characterization of anomaly detection for hyperspectral
imagery. Sensor Review 26(2): 137–146.
Chang, C.-I, and B. Ji. 2006a. Weighted least squares error approaches to abundance-constrained
linear spectral mixture analysis. IEEE Transactions on Geoscience and Remote Sensing 44(2):
378–388.
———. 2006b. Fisher’s linear spectral mixture analysis. IEEE Transactions on Geoscience and
Remote Sensing 44(8): 2292–2304.
Chang, C.-I, and Y. Li. 2016. Recursive band processing of automatic target generation process for
subpixel detection in hyperspectral imagery. IEEE Transactions on Geoscience and Remote
Sensing 54(9): 5081–5094.
Chang, C.-I, and K.-H. Liu. 2014. Progressive band selection for hyperspectral imagery. IEEE
Transactions on Geoscience and Remote Sensing 52(4): 2002–2017.
Chang, C.-I, and A. Plaza. 2006. Fast iterative algorithm for implementation of pixel purity index.
IEEE Geoscience and Remote Sensing Letters 3(1): 63–67.
References 663
Chang, C.-I, and H. Ren. 1999. Linearly constrained minimum variance beamforming for target
detection and classification in hyperspectral imagery. In IEEE 1999 international geoscience
and remote sensing symposium, Hamburg, Germany, 28 June–2 July, 1241–1243.
———. 2000. An experiment-based quantitative and comparative analysis of hyperspectral target
detection and image classification algorithms. IEEE Transactions on Geoscience and Remote
Sensing 38(2): 1044–1063.
Chang, C.-I, and H. Safavi. 2011. Progressive dimensionality reduction by transform for
hyperspectral image analysis. Pattern Recognition 44(10): 2760–2773. doi:10.1016/j.patcog.
2011.03.030.
Chang, C.-I, and S. Wang. 2006. Constrained band selection for hyperspectral imagery. IEEE
Transactions on Geoscience and Remote Sensing 44(6): 1575–1585.
Chang, C.-I, and J. Wang. 2008. Real-time implementation of field programmable gate arrays
(FPGA) design in hyperspectral imagery. US Patent, number 7,366,326, April 29.
Chang, C.-I, and Y. Wang. 2015. Anomaly detection using causal sliding windows. IEEE Journal
of Selected Topics in Applied Earth Observations and Remote Sensing 8(7): 3260–3270.
Chang, C.-I, and C.C. Wu. 2015. Design and development of iterative pixel purity index. IEEE
Journal of Selected Topics in Applied Earth Observations and Remote Sensing 8(6):
2676–2695.
Chang, C.-I, and W. Xiong. 2010. High order statistics Harsanyi-Farrand-Chang method for
estimation of virtual dimensionality. In SPIE, vol. 7810, San Diego, CA, August 2–5.
Chang, C.-I, Y. Cheng, M.L.G. Althouse, L. Zhang, and J. Wang. 1992. Multistage image coding:
a top-down gray-level triangle method. In Proceedings of international symposium on spectral
sensing research (ISSSR), Kauai, Hawaii, September 15–20, 497–511.
Chang, C.-J., C.-I Chang, and M.-L. Chang. 1993. Subband multistage predictive coding. In
Proceedings of international conference on signal processing ‘93/Beijing, Beijing, China,
October 26–30, 783–787.
Chang, C.-I, T.-L.E. Sun, and M.L.G. Althouse. 1998a. An unsupervised interference rejection
approach to target detection and classification for hyperspectral imagery. Optical Engineering
37(3): 735–743.
Chang, C.-I, X. Zhao, M.L.G. Althouse, and J.-J. Pan. 1998b. Least squares subspace projection
approach to mixed pixel classification in hyperspectral images. IEEE Transactions on Geosci-
ence and Remote Sensing 36(3): 898–912.
Chang, C.-I, Q. Du, T.S. Sun, and M.L.G. Althouse. 1999. A joint band prioritization and band
decorrelation approach to band selection for hyperspectral image classification. IEEE Trans-
actions on Geoscience and Remote Sensing 37(6): 2631–2641.
Chang, C.-I, J.-M. Liu, B.-C. Chieu, C.-M. Wang, C.S. Lo, P.-C. Chung, H. Ren, C.-W. Yang, and
D.-J. Ma. 2000. A generalized constrained energy minimization approach to subpixel target
detection for multispectral imagery. Optical Engineering 39(5): 1275–1281.
Chang, C.-I, H. Ren, and S.S. Chiang. 2001a. Real-time processing algorithms for target detection
and classification in hyperspectral imagery. IEEE Transactions on Geoscience and Remote
Sensing 39(4): 760–768.
Chang, C.-I, S.-S. Chiang, and I.W. Ginsberg. 2001b. Anomaly detection in hyperspectral imag-
ery. In SPIE conference on geo-spatial image and data exploitation II, Orlando, FL,
20–24 April, 43–50.
Chang, C.-I, Q. Du, S.-S. Chiang, D. Heinz, and I.W. Ginsberg. 2001c. Unsupervised subpixel
target detection in hyperspectral imagery. In SPIE conference on algorithms for multispectral,
hyperspectral and ultraspectral imagery VII, Orlando, FL, 20–24 April, 370–379.
Chang, C.-I, H. Ren, Q. Du, S.-S. Chiang, and A. Ifarragurri. 2001d. An ROC analysis for subpixel
detection. In IEEE international geoscience and remote sensing symposium, Sydney, Australia,
July 9–13.
Chang, C.-I, S.S. Chiang, J.A. Smith, and I.W. Ginsberg. 2002. Linear spectral random mixture
analysis for hyperspectral imagery. IEEE Transactions on Geoscience and Remote Sensing 40
(2): 375–392.
664 References
Chang, C.-I, Jing Wang, F. D’Amico, and J.O. Jensen. 2003. Multistage pulse code modulation for
progressive spectral signature coding. In Chemical and biological standoff detection—optical
technologies for industrial and environmental sensing symposia—Photonics West, 27–-
31 October, 252–261.
Chang, C.-I, H. Ren, C.-C. Chang, J.O. Jensen, and F. D’Amico. 2004a. Estimation of subpixel
target size for remotely sensed imagery. IEEE Transactions on Geoscience and Remote
Sensing 42(6): 1309–1320.
Chang, C.-I, W. Liu, and C.-C. Chang. 2004b. Discrimination and identification for subpixel
targets in hyperspectral imagery. In IEEE International conference on image processing,
Singapore, October.
Chang, C.-I, Jing Wang, C.-C. Chang, and C. Lin. 2006a. Progressive coding for hyperspectral
signature characterization. Optical Engineering 45(9): 097002-1–097002-15.
Chang, C.-I, C.C. Wu, W. Liu, and Y.C. Ouyang. 2006b. A growing method for simplex-based
endmember extraction algorithms. IEEE Transactions on Geoscience and Remote Sensing 44
(10): 2804–2819.
Chang, C.-I, Y. Du, J. Wang, S.-M. Guo, and P. Thouin. 2006c. A survey and comparative study of
entropic and relative entropic thresholding techniques. IEE Proceedings, Vision, Image and
Signal Processing 153(6): 837–850.
Chang, C.-I, H. Ren, C.-I Chang, and B. Rand. 2008. How to design synthetic images to validate
and evaluate hyperspectral imaging algorithms. In SPIE conference on algorithms and tech-
nologies for multispectral, hyperspectral, and ultraspectral imagery XIV, March 16–20,
Orlando, FL.
Chang, C.-I, S. Chakravarty, H. Chen, and Y.C. Ouyang. 2009. Spectral derivative feature coding
for hyperspectral signature. Pattern Recognition 42(3): 395–408.
Chang, C.-I, S. Chakravarty, and C.-S. Lo. 2010a. Spectral feature probabilistic coding for
hyperspectral signatures. IEEE Sensors Journal 10(3): 395–409.
Chang, C.-I, X. Jiao, Y. Du, and M.-L. Chang. 2010b. A review of unsupervised hyperspectral
target analysis. EURASIP Journal on Advanced in Signal Processing. 2010: Article ID 503752,
26 p. doi:10.1155/2010/503752.
Chang, C.-I, C.C. Wu, C.-S. Lo, and M.-L. Chang. 2010b. Real-time simplex growing algorithms
for hyperspectral endmember extraction. IEEE Transactions on Geoscience and Remote
Sensing 40(4): 1834–1850.
Chang, C.-I, C.-C. Wu, and H.M. Chen. 2010c. Random pixel purity index algorithm. IEEE
Geoscience and Remote Sensing Letters 7(2): 324–328.
Chang, C.-I, W. Xiong, W. Liu, C.C. Wu, and C.C.C. Chen. 2010d. Linear spectral mixture
analysis-based approaches to estimation of virtual dimensionality in hyperspectral imagery.
IEEE Transactions on Geoscience and Remote Sensing 48(11): 3960–3979.
Chang, C.-I, B. Ramakishna, J. Wang, and A. Plaza. 2010e. Exploitation-based hyperspectral
image compression. Journal of Applied Remote Sensing 4: 041760. doi:10.1117/1.3530429.
Chang, C.-I, C.-C. Wu, and C.-T. Tsai. 2011a. Random N-finder algorithm. IEEE Transactions on
Image Processing 20(3): 641–656.
Chang, C.-I, X. Jiao, Y. Du, and H.M. Chen. 2011b. Component-based unsupervised linear
spectral mixture analysis for hyperspectral imagery. IEEE Transactions on Geoscience and
Remote Sensing 49(11): 4123–4137.
Chang, C.-I, W. Xiong, H.M. Chen, and J.W. Chai. 2011c. Maximum orthogonal subspace
projection to estimating number of spectral signal sources for hyperspectral images. IEEE
Journal of Selected Topics in Signal Processing 5(3): 504–520.
Chang, C.-I, S. Wang, K.H. Liu, and C. Lin. 2011d. Progressive band dimensionality expansion
and reduction for hyperspectral imagery. IEEE Journal of Selected Topics in Applied Earth
Observations and Remote Sensing 4(3): 591–614.
Chang, C.-I, C.H. Wen, and C.C. Wu. 2013. Relationship exploration among PPI, ATGP and VCA
via theoretical analysis. International Journal of Computational Science and Engineering 8(4):
361–367.
References 665
Chang, C.-I, W. Xiong, and C.H. Wen. 2014a. A theory of high order statistics-based virtual
dimensionality for hyperspectral imagery. IEEE Transactions on Geoscience and Remote
Sensing 52(1): 188–208.
Chang, C.-I, S.Y. Chen, L. Zhao, and C.C. Wu. 2014b. Endmember-specified virtual dimension-
ality in hyperspectral imagery. In 2014 I.E. international geoscience and remote sensing
symposium (IGARSS), Quebec, Canada, July 13–18.
Chang, C.-I, R. Schultz, M. Hobbs, S.-Y. Chen, Y. Wang, and C. Liu. 2015a. Progressive band
processing of constrained energy minimization. IEEE Transactions on Geoscience and Remote
Sensing 53(3): 1626–1637.
Chang, C.-I, C.C. Wu, K.H. Liu, H.M. Chen, C.C.C. Chen, and C.H. Wen. 2015b. Progressive
band processing of linear spectral unmixing for hyperspectral imagery. IEEE Journal of
Selected Topics in Applied Earth Observations and Remote Sensing 8(7): 2583–2597.
Chang, C.-I, Y. Li, R. Schultz, M. Hobbs, and W.M. Liu. 2015c. Progressive band processing of
anomaly detection. IEEE Journal of Selected Topics in Applied Earth Observations and
Remote Sensing 8(7): 3558–3571.
Chang, C.-I, H.C. Li, M. Song, C. Liu, and L.F. Zhang. 2015d. Real-time constrained energy
minimization for subpixel detection. IEEE Journal of Selected Topics in Applied Earth
Observations and Remote Sensing 8(6): 2545–2559.
Chang, C.-I, H.C. Li and M. Song. 2017. Recursive geometric simplex growing analysis for
finding endmembers in hyperspectral imagery. IEEE Journal of Selected Topics in Applied
Earth Observation and Remote Sensing. To appear.
Chang, C.-I, L.-C. Lee, and D. Paylor. 2015e. Virtual dimensionality analysis for hyperspectral
imagery. In Satellite data compression, communication and processing XI (ST127), SPIE
International symposium on SPIE sensing technology + applications, Baltimore, MD,
20–24 April.
Chang, C.-I, Y. Li, and C.C. Wu. 2015f. Band detection in hyperspectral imagery by pixel purity
index. In 7th workshop on hyperspectral image and signal processing: evolution in remote
sensing, (WHISPERS), Tokyo, Japan, 2–5 June.
Chang, C.-I, C. Gao, and S.Y. Chen. 2015g. Recursive automatic target generation process. IEEE
Geoscience and Remote Sensing Letters 12(9): 1848–1852.
Chang, C.-I, W. Xiong, and S.Y. Chen. 2016a. Convex cone volume analysis for finding
endmembers in hyperspectral imagery. International Journal of Computational Science and
Engineering 12(2/3): 209–236.
Chang, C.-I, S.Y. Chen, H.C. Li, and C.-H. Wen. 2016b. A comparative analysis among ATGP,
VCA and SGA for finding endmembers in hyperspectral imagery. IEEE Journal of Selected
Topics in Applied Earth Observations and Remote Sensing (99):1–27.
Chang, C.-I, H.C. Li, C.C. Wu and M. Song. 2016c. Recursive geometric simplex growing
analysis for finding endmembers in hyperspectral imagery. IEEE Journal of Selected Topics
in Applied Earth Observations and Remote Sensing (99):1–13.
Chaudhry, F., C. Wu, W. Liu, C.-I Chang, and A. Plaza. 2006. Pixel purity index-based algorithms
for endmember extraction from hyperspectral imagery. In Recent advances in hyperspectral
signal and image processing, ed. C.-I Chang, 29–61. Trivandrum: Research Signpost, India.
Chen, S., C.F.N. Cowan, and P.M. Grant. 1991. Orthogonal least squares learning algorithm for
radial basis function networks. IEEE Transaction on Neural Networks 2: 302–309.
Chen, X., J. Chen, X.X. Jia, B. Somer, J. Wu, and P. Coppin. 2011. A quantitative analysis of
virtual endmembers’ increased impact on collinearity effect in spectral unmixing. IEEE Trans-
actions on Geoscience and Remote Sensing 49(8): 2945–2956.
Chen, S.-Y., D. Paylor, and C.-I Chang. 2013. Anomaly-specified virtual dimensionality. In SPIE
conference on satellite data compression, communication and processing IX (OP 405), San
Diego, CA, August 25–29.
Chen, S.-Y., Y. Wang, C.C. Wu, C. Liu, and C.-I Chang. 2014a. Real time causal processing of
anomaly detection in hyperspectral imagery. IEEE Transactions on Aerospace and Electronics
Systems 50(2): 1511–1534.
666 References
Chen, S.Y., D. Paylor, and C.-I Chang. 2014b. Anomaly discrimination in hyperspectral imagery.
In Satellite data compression, communication and processing X (ST146), SPIE international
symposium on SPIE sensing technology + applications, Baltimore, MD, 5–9 May.
Chen, S.Y., Y.C. Ouyang, and C.-I Chang. 2014c. Recursive unsupervised fully constrained least
squares methods. In 2014 I.E. international geoscience and remote sensing symposium
(IGARSS), Quebec Canada, July 13–18.
Chen, S.Y., Y.-C. Ouyang, C. Lin, H.-M. Chen, C. Gao, and C.-I Chang. (2015). Progressive
endmember finding by fully constrained least squares method. In 7th workshop on
hyperspectral image and signal processing: evolution in remote sensing (WHISPERS),
Tokyo, Japan, 2–5 June.
Cheng, Y. 1993. Multistage pulse code modulation (MPCM). MS thesis, Department of Electrical
Engineering, University of Maryland, Baltimore County, Baltimore, MD.
Chiang, S.-S., and C.-I Chang. 1999. Target subpixel detection for hyperspectral imagery using
projection pursuit. In EOS/SPIE symposium on remote sensing, conference on image and
signal processing for remote sensing V, SPIE, vol. 3871, Florence, Italy, September 20–24,
107–115.
———. 2001. Discrimination measures for target classification. In IEEE 2001 international
geoscience and remote sensing symposium, Sydney, Australia, July 9–13.
Chiang, S.-S., C.-I Chang, and I.W. Ginsberg. 2000. Unsupervised hyperspectral image analysis
using independent components analysis. In IEEE 2000 international geoscience and remote
sensing symposium, Hawaii, USA, July 24–28.
———. 2001. Unsupervised subpixel target detection for hyperspectral images using projection
pursuit. IEEE Transactions on Geoscience and Remote Sensing 39(7): 1380–1391.
Chowdhury, A., and M.S. Alam. 2007. Fast implementation of N-FINDR algorithm for
endmember determination in hyperspectral imagery. In Proceedings of SPIE, vol. 6565,
656526-1–656526-7.
Christophe, E., D. Leger, and C. Mailhes. 2005. Quality criteria benchmark for hyperspectral
imagery. IEEE Transactions on Geoscience and Remote Sensing 43(9): 2103–2114.
Conese, C., and F. Maselli. 1993. Selection of optimum bands from TM scenes through mutual
information analysis. ISPRS Journal of Photogrammetry and Remote Sensing 48(3): 2–11.
Cover, T., and J. Thomas. 1991. Elements of information theory. New York: Wiley.
Craig, M.D. 1994. Minimum-volume transforms for remotely sensed data. IEEE Transactions on
Geoscience and Remote Sensing 32(3): 542–552.
Dennison, P.E., and D.A. Roberts. 2003. Endmember selection for multiple endmember spectral
mixture analysis using endmember average RMSE. Remote Sensing of Environments 87:
123–135.
Dowler, A., and M. Andrews. 2011. On the convergence of N-FINDR and related algorithms: to
iterate or not to iterate? IEEE Geoscience and Remote Sensing Letters 8(1): 4–8.
Du, B., and L. Zhang. 2011. Random selection based anomaly detector for hyperspectral imagery.
IEEE Transactions on Geoscience and Remote Sensing 49(5): 1578–1589.
Du, Q. 2000. Topics in hyperspectral image analysis. Baltimore County, MD: Department of
Computer Science and Electrical Engineering, University of Maryland.
Du, Q. 2012. A new sequential algorithm for hyperspectral endmember extraction. IEEE Geosci-
ence and Remote Sensing Letters, vol. 9, no. 4, pp. 695–699, July 2012.
Du, Q., and C.-I Chang. 1998. Radial basis function neural networks approach to hyperspectral
image classification. In 1998 Conference on information science and systems, 721–726.
Princeton, NJ: Princeton University.
———. 1999. An interference rejection-based radial basis function neural network approach to
hyperspectral image classification. In International joint conference on neural network,
Washington DC, 2698–2703.
———. 2000. A hidden Markov model-based spectral measure for hyperspectral image analysis.
In SPIE Conference algorithms for multispectral, hyperspectral, and ultraspectral imagery VI,
Orlando, FL, 375–385.
References 667
———. 2001a. A linear constrained distance-based discriminant analysis for hyperspectral image
classification. Pattern Recognition 34(2): 2001.
———. 2001b. An interference subspace projection approach to subpixel target detection. In SPIE
conference on algorithms for multispectral, hyperspectral and ultraspectral imagery VII,
Orlando, FL, 20–24 April, 570–577.
———. 2004a. Linear mixture analysis-based compression for hyperspectral image analysis.
IEEE Transactions on Geoscience and Remote Sensing 42(4): 875–891.
———. 2004b. A signal-decomposed and interference-annihilated approach to hyperspectral
target detection. IEEE Transactions on Geoscience and Remote Sensing 42(4): 892–906.
———. 2007. Rethinking the effective assessment of biometric systems. http://newsroom.spie.
org/x17545.xml, 2007. doi:10.1117/2.1200711.0815.
———. 2008. 3D combination curves for accuracy and performance analysis of positive bio-
metrics identification. Optics and Lasers in Engineering 46(6): 477–490.
Du, Q., and J.E. Fowler. 2007. Hyperspectral image compression using JPEG2000 and principal
component analysis. IEEE Geoscience and Remote Sensing Letters 4: 201–205.
Du, Q. and R. Nevovei. 2005. Implementation of real-time constrained linear discriminant
analysis to remote sensing image classification in hyperspectral imagery. Pattern Recognition
38(4): 1–12.
Du, Q and R. Nekovei. 2009. Fast real-time onboard processing of hyperspectral imagery for
detection and classification. J. Real Time Image Processing 4(3):273–286.
Du, Q., C.-I Chang, D.C. Heinz, M.L.G. Althouse, and I.W. Ginsberg. 2000. Hyperspectral image
compression for target detection and classification. In IEEE 2000 international geoscience and
remote sensing symposium, Hawaii, USA, July 24–28.
Du, Q., H. Ren, and C.-I Chang. 2003. A comparative study for orthogonal subspace projection
and constrained energy minimization. IEEE Transactions on Geoscience and Remote Sensing
41(6): 1525–1529.
Du, Y., C.-I Chang, and P. Thouin. 2003b. An automatic system for text detection in single video
images. Journal of Electronic Imaging 12(3): 410–422.
Du, Q., W. Zhu, and J. Fowler. 2004a. Anomaly-based JPEG2000 compression of hyperspectral
imagery. IEEE Geoscience and Remote Sensing Letters 5(4): 696–700.
Du, Y., C.-I Chang, C.-C. Chang, F. D’Amico, and J.O. Jensen. 2004b. A new hyperspectral
measure for material discrimination and identification. Optical Engineering 43(8): 1777–1786.
Du, Z., M.K. Jeong, and S.G. Kong. 2007. Band selection of hyperspectral images for automatic
detection of poultry skin tumors. IEEE Transactions on Automatic Science and Engineering 4
(3): 332–339.
Du, Q., N. Raksuntorn, and N.H. Younan. 2008a. Variants of N-FINDR algorithm for endmember
extraction. In Proceedings of SPIE, vol. 7109, 71090G-1–71090G-8, 15–18 September.
Du, Q., N. Raksuntorn, N.H. Younan, and R.L. King. 2008b. Endmember extraction algorithms for
hyperspectral image analysis. Applied Optics 47(28): F77–F84.
Duda, R.O., and P.E. Hart. 1973. Pattern classification and scene analysis. New York: Wiley.
Eches, O., N. Dobigeon, and J.-Y. Tourneret. 2010. Estimating the number of endmembers in
hyperspectral images using the normal compositional model and a hierarchical Bayesian
algorithm. IEEE Journal of Selected Topics in Signal Processing 4(3): 582–591.
Epp, S.S. 1995. Discrete mathematics with applications, 2nd ed. Pacific Grove, CA: Brooks/Cole.
Fano, R.M. 1961. Transmission of information: a statistical theory of communication. New York:
Wiley.
Farrand, W., and J.C. Harsanyi. 1997. Mapping the distribution of mine tailing in the coeur
d’Alene river valley, Idaho, through the use of constrained energy minimization technique.
Remote Sensing of Environment 59: 64–76.
Filippi, A.M., and R. Archibald. 2009. Support vector machine-based endmember extraction.
IEEE Transactions on Geoscience and Remote Sensing 47(3): 771–791.
Fisher, K., and C.-I Chang. 2011. Progressive band selection for satellite hyperspectral data
compression and transmission. Journal of Applied Remote Sensing 4(1): 041770.
668 References
Frost III, O.L. 1972. An algorithm for linearly constrained adaptive array processing. Proceedings
of IEEE 60: 926–935.
Fukunaga, K. 1982. Intrinsic dimensionality extraction. In Classification, pattern recognition and
reduction of dimensionality, Handbook of Statistics, vol. 2, ed. P.R. Krishnaiah and
L.N. Kanal, 347–360. Amsterdam: North-Holland.
———. 1990. Statistical pattern recognition, 2nd ed. New York: Academic.
Gao, C., and C.-I Chang. 2014. Recursive automatic target generation process for unsupervised
hyperspectral target detection. In 2014 I.E. international geoscience and remote sensing
symposium (IGARSS), Quebec Canada, July 13–18.
Gao, C., S.Y. Chen, and C.-I Chang. 2014. Fisher’s ratio-based criterion for finding endmembers in
hyperspectral imagery. In Satellite data compression, communication and processing X
(ST146), SPIE international symposium on SPIE sensing technology + applications, Balti-
more, MD, 5–9 May.
Gao, C., Y. Li, and C.-I Chang. 2015a. Finding endmember classes in hyperspectral imagery. In
Satellite data compression, communication and processing XI (ST127), SPIE international
symposium on SPIE sensing technology + applications, Baltimore, MD, 20–24 April.
Gao, C., S.-Y. Chen, H.M. Chen, C.C. Wu, C.H. Wen, and C.-I Chang. 2015b. Fully abundance-
constrained endmember finding for hyperspectral images. In 7th Workshop on hyperspectral
image and signal processing: evolution in remote sensing (WHISPERS), Tokyo, Japan,
2–5 June.
Garzon, E.M., I. Garcia, and A. Plaza. 2012. Anomaly detection based on a parallel kernel RX
algorithm for multicore platforms. Journal of Applied Remote Sensing 6: 061503-1–061503-
10.
Gelb, A., ed. 1974. Applied optimal estimation. Cambridge: MIT Press.
Geng, X., K. Sun, L. Ji, and Y. Zhao. 2014. A fast volume-gradient-based band selection method
for hyperspectral image. IEEE Transactions on Geoscience and Remote Sensing 52(11):
7111–7119.
Gersho, A., and R.M. Gray. 1992. Vector quantization and signal compression. New York: Kluwer
Academics.
Gillespie, A.R., M.O. Smith, J.B. Adams, S.C. Willis, A.F. Fischer III, and D.E. Sabol. 1990.
Interpretation of residual images: spectral mixture analysis of AVIRIS images, Owens valley,
California. In Proceedings of 2nd AVIRIS workshop, 243–270.
Goetz, A.F.H., and J.W. Boardman. 1989. Quantitative determination of imaging spectrometer
specifications based on spectral mixing models. In Proceedings of IEEE international geosci-
ence and remote sensing symposium’89, 1036–1039.
Golub, G.H., and G.F. Van Loan. 1989. Matrix computations, 2nd ed. Baltimore: John Hopkins
University Press.
Gonzalez, R.C., and R.E. Woods. 2007. Digital Image Processing, 3rd ed. Upper Saddle River:
Prentice-Hall.
Green, A.A., M. Berman, P. Switzer, and M.D. Craig. 1988. A transformation for ordering
multispectral data in terms of image quality with implications for noise removal. IEEE
Transactions on Geoscience and Remote Sensing 26: 65–74.
Greg, I. 2010. An evaluation of three endmember extraction algorithms, ATGP, ICA-EEA and
VCA. In Hyperspectral image and signal processing: evolution in remote sensing (WHIS-
PERS), 1–4.
Gruniger, J., A.J. Ratkowski, and M.L. Hoje. 2004. The sequential maximum angle convex cone
(SMACC) endmember model. In Proceedings of SPIE, algorithms and technologies for
multispectral, hyperspectral, and ultraspectral imagery X, vol. 5425, 1–14.
Guilfoyle, K. 2003. Application of linear and nonlinear mixture models to hyperspectral imagery
analysis using radial basis function neural networks. Baltimore County, Baltimore, MD:
Department of Computer Science and electrical Engineering, University of Maryland.
References 669
Guilfoyle, K., M.L.G. Althouse, and C.-I Chang. 2001. A quantitative and comparative analysis of
linear and nonlinear spectral mixture models using radial basis function neural networks. IEEE
Transactions on Geoscience and Remote Sensing 39(10): 2314–2318.
———. 2002. Further results on linear and nonlinear mixture models for analyzing hyperspectral
imagery. In SPIE conference on algorithms and technologies for multispectral, hyperspectral
and ultraspectral imagery VIII, SPIE 4725, Orlando, FL, 1–5 April.
Harsanyi, J.C. 1993. Detection and classification of subpixel spectral signatures in hyperspectral
image sequences. Baltimore County, MD: Department of Electrical Engineering, University of
Maryland.
Harsanyi, J.C., and C.-I Chang. 1994. Hyperspectral image classification and dimensionality
reduction: an orthogonal subspace projection approach. IEEE Transactions on Geoscience
and Remote Sensing 32(4): 779–785.
Harsanyi, J.C., W. Farrand, and C.-I Chang. 1994a. Detection of subpixel spectral signatures in
hyperspectral image sequences. In Annual meeting, proceedings of American society of
photogrammetry and remote sensing, Reno, 236–247.
Harsanyi, J.C., W. Farrand, J. Hejl, and C.-I Chang. 1994b. Automatic identification of spectral
endmembers in hyperspectral image sequences. In International symposium on spectral sens-
ing research ‘94 (ISSSR), San Diego, July 10–15, 267–277.
Haskell, K.H., and R.J. Hanson. 1981. An algorithm for linear least squares problems with equality
and nonnegativity constraints generalized. Mathematical Programming 21: 98–118.
Heinz, D., and C.-I Chang. 2001. Fully constrained least squares linear mixture analysis for
material quantification in hyperspectral imagery. IEEE Transactions on Geoscience and
Remote Sensing 39(3): 529–545.
Heute, A.R. 1986. Separation of soil-plant spectral mixtures by factor analysis. Remote Sensing
Environment 19: 237–251.
Heylen, R., and P. Scheunders. 2013. Hyperspectral intrinsic dimensionality estimation with
nearest-neighbor distance ratios. IEEE Journal of Selected Topics in Applied Earth Observa-
tions and Remote Sensing 6(2): 570–579.
Hofmann, T., B. Scgikkopf, and A.J. Smola. 2008. Kernel methods in machine learning. Annals of
Statistics 36(3): 1171–1220.
Honeine, P. and C. Richard. 2012. Geometric unmixing of large hyperspectral images: a
barycentric coordinate approach. IEEE Trans. on Geoscience and Remote Sensing 50(4):
2185–2195.
Hsueh, M. 2004. Adaptive causal anomaly detection. M.S. thesis, Department of computer Science
and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, MD.
———. 2007. Reconfigurable computing for algorithms in hyperspectral image processing. Ph.D.
dissertation, Department of Computer Science and Electrical Engineering, University of
Maryland, Baltimore County, Baltimore, MD.
Hsueh, M., and C.-I Chang. 2004. Adaptive causal anomaly detection for hyperspectral imagery.
In IEEE international geoscience and remote sensing symposium, Alaska, September 20–24.
———. 2008. Field programmable gate arrays for pixel purity index using blocks of skewers for
endmember extraction in hyperspectral imagery. International Journal of High Performance
Computing Applications 22(4): 408–423.
Huang, R., and M. He. 2005. Band selection based feature weighting for classification of
hyperspectral data. IEEE Geoscience and Remote Sensing Letters 2(2): 156–159.
HYMSMO. 1998. HYperspectral MASINT support to military operations program.
Hyvarinen, A., and E. Oja. 1997. A fast fixed-point for independent component analysis. Neural
Computation 9(7): 1483–1492.
———. 2000. Independent component analysis: algorithms and applications. Neural Networks 13:
411–430.
Hyvarinen, A., J. Karhunen, and E. Oja. 2001. Independent component analysis. New York:
Wiley.
670 References
Ifarragaerri, A., and C.-I Chang. 1999. Hyperspectral image segmentation with convex cones.
IEEE Transactions on Geoscience and Remote Sensing 37(2): 756–770.
Ifarraguerri, A. 2000. Hyperspectral image analysis with convex cones and projection pursuit.
Baltimore County, MD: Department of Computer Science and Electrical Engineering, Uni-
versity of Maryland.
Ifarragaerri, A., and C.-I Chang. 2000. Multispectral and hyperspectral image analysis with
projection pursuit. IEEE Transactions on Geoscience and Remote Sensing 38(6): 2529–2538.
Jensen, J.R. 1996. Introductory digital image processing: A remote sensing perspective, 2nd
ed. Prentice-Hall, NJ: Upper Saddle River.
Jenson, S.K., and F.A. Waltz. 1979. Principal components analysis and canonical analysis in
remote sensing. In Proceedings American Society for photogrammetry 45th annual meeting,
337–348.
Ji, B. 2006. Constrained linear spectral mixture analysis. Ph.D. dissertation, Department of
Computer Science and Electrical Engineering, University of Maryland, Baltimore County,
Baltimore, MD.
Ji, B., and C.-I Chang. 2006. Principal components analysis-based endmember extraction algo-
rithms. In Recent advances in hyperspectral signal and image processing, ed. C.-I Chang,
63–91. Trivandrum, Kerala: Research Signpost, India.
Ji, B., C.-I Chang, J.O. Jensen, and J.L. Jensen. 2004. Unsupervised constrained linear Fisher’s
discriminant analysis for hyperspectral image classification. In 49th Annual meeting, SPIE
international symposium on optical science and technology, imaging spectrometry IX
(AM105), vol. 5546, 344–353. Denver, CO, August 2–4.
Jiao, X. 2010. Unsupervised hyperspectral target detection and classification. Ph.D. dissertation,
Department of Computer Science and Electrical Engineering, University of Maryland, Balti-
more County, Baltimore, MD.
Jiao, X., and C.-I Chang. 2008. Kernel-based constrained energy minimization (KCEM). In SPIE
conference on algorithms and technologies for multispectral, hyperspectral, and ultraspectral
imagery XIV, March 16–20, Orlando, FL.
Jimenez, L.O., and D.A. Landgrebe. 1999. Hyperspectral data analysis and supervised feature
reduction via projection pursuit. IEEE Transactions on Geoscience and Remote Sensing 37(6):
2653–2667.
Johnson, P., M. Smith, S. Taylor-George, and J. Adams. 1983. A semiempirical method for
analysis of the reflectance spectra of binary mineral mixtures. Journal of Geophysical
Research 88: 3557–3561.
Johnson, W.R., D.W. Wilson, W. Fink, M. Humayun, and G. Bearman. 2007. Snapshot
hyperspectral imaging in ophthalmology. Journal of Biomedical Optics 12(1): 014036-
1–014036-7.
Kailath, T. 1968. An innovations approach to least squares estimation Part I: linear filtering in
additive while noise. IEEE Transaction on Automatic Control 13(6): 646–655.
Kailath, T. 1980. Linear systems. Upper Saddle River: Prentice Hall.
Kanaev, A.V., E. Allman, and J. Murray-Krezan. 2009. Reduction of false alarms caused by
background boundaries in real time subspace RX anomaly detection. In Proceedings of SPIE,
vol. 7334, 733405.
Kelly, E.J. 1986. An adaptive detection algorithm. IEEE Transactions on Aerospace and Elec-
tronic Systems 22: 115–127.
Keshava, N., and J.F. Mustard. 2002. Spectral unmixing. IEEE Signal Processing Magazine 19(1):
44–57.
Khazai, S., S. Homayouni, A. Safari, and B. Mojaradi. 2011. Anomaly detection in hyperspectral
images based on an adaptive support vector method. IEEE Transactions on Geoscience and
Remote Sensing 8(4): 646–650.
Kraut, S., and L. Scharf. 1999. The CFAR adaptive subspace detector is a scale-invariant GLRT.
IEEE Transactions of Signal Processing 47(9): 2538–2541.
References 671
Kraut, S., L. Scharf, and L.T. McWhorter. 2001. Adaptive subspace detector. IEEE Transactions
on Signal Processing 49(12): 3005–3014.
Kullback, S. 1959/1968. Information theory and statistics. New York: Wiley.
Kuybeda, O., D. Malah, and M. Barzohar. 2007. Rank estimation and redundancy reduction of
high-dimensional noisy signals with preservation of rare vectors. IEEE Transactions on Signal
Processing 55(12): 5579–5592.
Kwan, C., B. Ayhan, G. Chen, J. Wang, B. Ji, and C.-I Chang. 2006. A novel approach for spectral
unmixing, classification, and concentration estimation of chemical and biological agents. IEEE
Transactions on Geoscience and Remote Sensing 44(2): 409–419.
Kwon, H., and N.M. Nasrabadi. 2005a. Kernel RX-algorithm: a nonlinear anomaly detector for
hyperspectral imagery. IEEE Transactions on Geoscience and Remote Sensing 43(2):
388–397.
———. 2005b. Kernel orthogonal subspace projection for hyperspectral signal classification.
IEEE Transactions on Geoscience and Remote Sensing 43(12): 2952–2962.
Kwon, H., S.Z. Der, and N.M. Nasrabadi. 2003. Adaptive anomaly detection using subspace
separation for hyperspectral imagery. Optical Engineering 42(11): 3342–3351.
Langrebe, D.A. 2003. Signal theory methods in multispectral remote sensing. Hoboken: Wiley.
Lawson, C.L., and R.J. Hanson. 1995. Solving least squares problems, CAM, vol. 15. Philadelphia,
PA: SIAM.
Leadbetter, M. 1987. Extremes and related properties of random sequences and processes.
New York: Springer.
Lee, L.C., and C.-I Chang. 2016. An information theoretical approach to multiple band selection
for hyperspectral imagery. In 2016 I.E. geoscience and remote sensing symposium, Beijing,
China, July 10–15.
Lee, D.D., and N.S. Seung. 1999. Learning the parts of objects by non-negative matrix factoriza-
tion. Science 401(21): 788–791.
Lee, J.B., A.S. Woodyatt, and M. Berman. 1990. Enhancement of high spectral resolution remote
sensing data by a noise-adjusted principal components transform. IEEE Transactions on
Geoscience and Remote Sensing 28(3): 295–304.
Lee, T.W., M.S. Lewicki, and M. Girolami. 1999. Bind source separation of more sources than
mixtures using overcomplete representations. IEEE Transactions on Signal Processing 6(4):
87–90.
Lee, L.-C., D. Paylor, and C.-I Chang. 2015. Anomaly discrimination and classification for
hyperspectral imagery. In 7th Workshop on hyperspectral image and signal processing:
evolution in remote sensing (WHISPERS), Tokyo, Japan, 2–5 June.
Li, H.C. 2016. Simplex Volume Analysis for Finding Endmembers in Hyperspectral Imagery,
Department of Computer Science and Electrical Engineering, University of Maryland, Balti-
more County, Baltimore, MD, May 2016.
Li, H.C., and C.-I Chang. 2015a. An orthogonal projection approach to simplex growing algorithm
for finding endmembers in hyperspectral imagery. In 7th Workshop on hyperspectral image
and signal processing: evolution in remote sensing (WHISPERS), Tokyo, Japan, 2–5 June.
———. 2015b. Linear spectral unmixing using least squares error, orthogonal projection and
simplex volume for hyperspectral Images. In 7th Workshop on hyperspectral image and signal
processing: evolution in remote sensing (WHISPERS), Tokyo, Japan, 2–5 June.
———. 2016a. Progressive band processing of fast iterative pixel purity index. In Remotely sensed
data compression, communications, and processing XII, Part of SPIE commercial + scientific
sensing and imaging, 17–21 April.
———. 2016b. Real time hyperspectral anomaly detection via band-interleaved by line. In
Remotely sensed data compression, communications, and processing XII, part of SPIE
commercial + scientific sensing and imaging, 17–21 April.
———. 2016c. Recursive orthogonal projection-based simplex growing algorithm. IEEE Trans-
actions on Geoscience and Remote Sensing 54(7): 3780–3793. doi:10.1109/TGRS.2016.
2527737.
672 References
Li, H.C., M. Song, and C.-I Chang. 2014a. Finding analytical solutions to abundance fully-
constrained linear spectral mixture analysis. In 2014 I.E. international geoscience and remote
sensing symposium (IGARSS), Quebec, Canada, July 13–18.
Li, Y., S.Y. Chen, C. Gao, and C.-I Chang. 2014b. Endmember variability resolved by pixel purity
index in hyperspectral imagery. In Satellite data compression, communication and processing
X (ST146), SPIE international symposium on SPIE sensing technology + applications, Balti-
more, MD, 5–9 May.
Li, H.C., M. Song, and C.-I Chang. 2015a. Simplex volume analysis for finding endmembers in
hyperspectral imagery. In Satellite data compression, communication and processing XI
(ST127), SPIE international symposium on SPIE sensing technology + applications, Balti-
more, MD, 20–24 April.
Li, Y., H.C. Li, C. Gao, and M. Song. 2015b. Progressive band processing of pixel purity index for
finding endmembers in hyperspectral imagery. In Satellite data compression, communication
and processing XI (ST127), SPIE international symposium on SPIE sensing technology
+ applications, Proceedings of SPIE, vol. 9501, 95010U-1–95010U-10, Baltimore, MD,
20–24 April.
Li, Y., C. Gao, and C.-I Chang. 2015c. Progressive band processing of automatic target generation
process. In 7th Workshop on hyperspectral image and signal processing: evolution in remote
sensing (WHISPERS), Tokyo, Japan, 2–5 June.
Li, H.C., C.-I Chang, and L. Wang. 2016. Constrained multiple band selection for hyperspectral
imagery. In 2016 I.E. geoscience and remote sensing symposium, Beijing, China, July 10–15.
Liu, W. 2005. Supervised and unsupervised classification for purdue indian pines test site.
M.S. thesis, Department of Computer Science and Electrical Engineering, University of
Maryland, Baltimore County, MD.
———. 2008. Unsupervised hyperspectral target recognition. Ph.D. dissertation, Department of
Computer Science and Electrical Engineering, University of Maryland, Baltimore County,
MD.
———. 2011. Progressive band prioritization and selection for linear spectral mixture analysis.
Baltimore County, Baltimore, MD: Department of Computer Science and Electrical Engineer-
ing, University of Maryland.
Liu, W., and C.-I Chang. 2004. A nested spatial window-based approach to target detection for
hyperspectral imagery. In IEEE international geoscience and remote sensing symposium,
Alaska, September 20–24.
———. 2008. Multiple window anomaly detection for hyperspectral imagery. In IEEE interna-
tional geoscience and remote sensing symposium, Boston, MA, July 6–11.
———. 2011. Dynamic band selection for hyperspectral imagery. In International geoscience and
remote sensing symposium, 24–29 July, Vancouver, Canada.
———. 2013. Multiple window anomaly detection for hyperspectral imagery. IEEE Journal of
Selected Topics in Applied Earth Observations and Remote Sensing 6(2): 664–658.
Liu, W., C.-I Chang, S. Wang, J. Jensen, J. Jensen, H. Hnapp, R. Daniel, and R. Yin. 2005. 3D
ROC analysis for detection software used in water monitoring. In OpticsEast, chemical and
biological standoff detection III (SA03), Boston, MA, October 23–26.
Liu, W., C.-C. Wu, and C.-I Chang. 2007. An orthogonal subspace projection-based estimation of
virtual dimensionality for hyperspectral data exploitation. In Algorithms and technologies for
multispectral, hyperspectral, and ultraspectral imagery XIII, SPIE defense and security sym-
posium, Orlando, FL, April 9–13.
Liu, K., E. Wong, Y. Du, C.C.C. Chen, and C.-I Chang. 2012. Kernel-based linear spectral mixture
analysis. IEEE Geoscience and Remote Sensing Letters 9(1): 129–133.
Lopez, S., P. Horstrand, G.M. Callico, J.F. Lopez, and R. Sarmiento. 2012a. A low-computational-
complexity algorithm for hyperspectral endmember extraction: modified vertex component
analysis. IEEE Geoscience and Remote Sensing Letters 9(3): 502–506.
References 673
———. 2012b. A novel architecture for hyperspectral endmember extraction by means of the
modified vertex component analysis (MVCA) algorithm. IEEE Journal of Selected Topics in
Applied Earth Observation and Remote Sensing 5(6): 1837–1848.
Luo, B., J. Chanussot, S. Douté, and L. Zhang. 2013. Empirical automatic estimation of the
number of endmembers in hyperspectral images. IEEE Geoscience and Remote Sensing Letters
10(1): 24–28.
Ma, W.-K., J.M. Bioucas-Dias, T.-H. Chan, N. Gillis, P. Gader, A.J. Plaza, A. Ambikapathi, and
C.-Y. Chi. 2014. A signal processing perspective on hyperspectral unmixing. IEEE Signal
Processing Magazine 31(1): 67–81.
Malinowski, E.R. 1977. Determination of the number of factors and experimental error in a data
matrix. Analytical Chemistry 49: 612–617.
Manolakis, D., and G. Shaw. 2002. Detection algorithms for hyperspectral imaging applications.
IEEE Signal Processing Magazine 19(1): 29–43.
Manolakis, D., C. Siracusa, and G. Shaw. 2001. Hyperspectral subpixel target detection using the
linear mixture model. IEEE Transactions on Geoscience and Remote Sensing 39(7):
1392–1409.
Mausel, P.W., W.J. Kramber, and J.K. Lee. 1990. Optimum band selection for supervised
classification of multispectral data. Photogrammetric Engineering and Remote Sensing 56
(1): 55–60.
Mavrovouniotis, M.L., A.M. Harper, and A. Ifarraguerri. 1994a. Classification of pyrolysis mass
spectra of biological agents using convex cones. Journal of Chemometrics 8: 305–33.
———. 1994b. Convex-cone analysis of the time profiles of pyrolysis mass spectra of biological
agents. U.S. Army Edgewood Research, Development and Engineering Center, Technical
Report ERDEC-CR-130.
Mavrovouniotis, M.L., A.M. Harper, and A. Ifarraguerri. 1996. A method for extracting patterns
from pyrolysis mass spectra, Computer assisted analytical spectroscopy, 189–240. Chichester:
Wiley.
Metz, C.E. 1978. ROC methodology in radiological imaging. Investigative Radiology 21:
720–723.
Miao, L., and H. Qi. 2007. Endmember extraction from highly mixed data using minimum volume
constrained nonnegative matrix factorization. IEEE Transactions on Geoscience and Remote
Sensing 45(3): 765–777.
Moon, T.K., and W.C. Stirling. 2000. Mathematical methods and algorithms for signal processing.
Upper Saddle River: Prentice-Hall.
Mustard, J.F., and C.M. Pieters. 1987. Abundance and distribution of ultramafic microbreccia in
moses rock dike: quantitative application of mapping spectroscopy. Journal of Geophysical
Research 92(B10): 10376–10390.
Nascimento, J.M.P., and J.M. Dias. 2005. Vertex component analysis: a fast algorithm to unmix
hyperspectral data. IEEE Transactions on Geoscience and Remote Sensing 43(4): 898–910.
Neville, R.A., K. Staenz, T. Szeredi, J. Lefebvre, and P. Hauff. 1999. Automatic endmember
extraction from hyperspectral data for mineral exploration. In Proceedings of 4th international
airborne remote sensing conference and exhibition/21st Canadian symposium on remote
sensing, Ottawa, Ontario, Canada, 21–24.
Oja, E., and M. Plumbly. 2004. Blind separation of positive sources by globally convergent
gradient search. Neural Computation 16: 1811–1825.
Pauca, V.P., J. Piper, and R.J. Plemmons. 2006. Nonnegative matrix factorization for spectral data
analysis. Linear Algebra and its Applications 416(1): 29–47.
Paylor, D. 2014. Second order statistics target-specified virtual dimensionality. Department of
Computer Science and Electrical Engineering, University if Maryland, Baltimore County, MD.
Paylor, D., and C.-I Chang. 2013. Second-order statistics-specified virtual dimensionality. In SPIE
conference on algorithms and technologies for multispectral, hyperspectral and ultraspectral
imagery XIX (DS122), 29 April–3 May 2013, Baltimore, MD.
674 References
Ren, H., and C.-I Chang. 1998. A computer-aided detection and classification method for
concealed targets in hyperspectral imagery. In IEEE 1998 international geoscience and remote
sensing symposium, Seattle, WA, July 5–10, 1016–1018.
———. 1999. A constrained least squares approach to hyperspectral image classification. In 1999
conference on information science and systems, 551–556. Baltimore, MD: Johns Hopkins
University.
———. 2000a. A generalized orthogonal subspace projection approach to unsupervised multi-
spectral image classification. IEEE Transactions on Geoscience and Remote Sensing 38(6):
2515–2528.
———. 2000b. Target-constrained interference-minimized approach to subpixel target detection
for hyperspectral imagery. Optical Engineering 39(12): 3138–3145.
———. 2003. Automatic spectral target recognition in hyperspectral imagery. IEEE Transactions
on Aerospace and Electronic Systems 39(4): 1232–1249.
Ren, H., Q. Du, J. Wang, C.-I Chang, and J. Jensen. 2006. Automatic target recognition
hyperspectral imagery using high order statistics. IEEE Transactions on Aerospace and
Electronic Systems 42(4): 1372–1385.
Research Systems, Inc. 2001. ENVI user’s guide. Boulder, CO: Research Systems, Inc.
Resmini, R.S., M.E. Kappus, W.S. Aldrich, J.C. Harsanyi, and M. Anderson. 1997. Mineral
mapping with HYperspectral Digital Imagery Collection Experiment (HYDICE) sensor data
at Cuprite, Nevada, USA. International Journal of Remote Sensing 18(17): 1553–1570.
Richards, J.A., and X. Jia. 1999. Remote sensing digital image analysis. New York: Springer.
Roberts, D.A., M. Gardner, R. Church, S.L. Ustin, G. Scheer, and R.O. Green. 1998. Mapping
chaparral in the Santa Monica mountains using multiple endmember spectral mixture model.
Remote Sensing of Environment 65: 267–279.
Robinove, C.J. 1982. Computation with physical values from Landsat digital data. Photogram-
metric Engineering and Remote Sensing 48: 781–784.
Roger, R.E. 1994. A fast way to compute the noise-adjusted principal components transform
matrix. IEEE Transactions on Geoscience and Remote Sensing 32(1): 1194–1196.
———. 1996. Principal components transform with simple, automatic noise adjustment. Interna-
tional Journal of Remote Sensing 17(14): 2719–2727.
Roger, R.E., and J.F. Arnold. 1996. Reliably estimating the noise in AVIRIS hyperspectral
imagers. International Journal of Remote Sensing 17(10): 1951–1962.
Roggea, D.M., B. Rivarda, J. Zhanga, A. Sancheza, J. Harrisb, and J. Feng. 2007. Integration of
spatial–spectral information for the improved extraction of endmembers. Remote Sensing of
Environment 110(3): 287–303.
Rosario D. 2012. A semiparametric model for hyperspectral anomaly detection. Journal of
Electrical and Computer Engineering. 2012: Article ID 425947, 30 p.
Sabol, D.E., J.B. Adams, and M.O. Smith. 1992. Quantitative sub-pixel spectral detection of
targets in multispectral images. Journal of Geophysical Research 97: 2659–2672.
Safavi, H., and C.-I Chang. 2008. Projection pursuit-based dimensionality reduction. In SPIE
conference on algorithms and technologies for multispectral, hyperspectral, and ultraspectral
imagery XIV, March 16–20, Orlando, FL.
Safavi, H., K. Liu, and C.-I Chang. 2011. Dynamic dimensionality reduction for hyperspectral
imagery. In SPIE conference on algorithms and technologies for multispectral, hyperspectral
and ultraspectral imagery XVII, 25–29 April 2011, Orlando, FL.
Scharf, L.L. 1991. Statistical signal processing. Boston: Addison-Wesley.
Schowengerdt, R.A. 1997. Remote sensing: models and methods for image processing, 2nd
ed. New York: Academic.
Schultz, R.C., S.Y. Chen, Y. Wang, C. Liu, and C.-I Chang. 2013. Progressive band processing of
anomaly detection. In SPIE conference on satellite data compression, communication and
processing IX (OP 405), San Diego, CA, August 25–29.
Schultz, R.C., M. Hobbs, and C.-I Chang. 2014. Progressive band processing of simplex growing
algorithm for finding endmembers in hyperspectral imagery. In Satellite data compression,
676 References
Swets, J.A., and R.M. Pickett. 1982. Evaluation of diagnostic systems: methods from signal
detection theory. New York: Academic.
Tarabalka, T., T.V. Haavardsholm, I. Kasen and T. Skauli. 2009. Real-time anomaly detection in
hyperspectral images using multivariate normal mixture models and GPU processing. J. Real
Time Image Processing 4(3): 287–300.
Thai, B., and G. Healey. 1999. Invariant subpixel material identification in hyperspectral imagery.
In SPIE, vol. 3717.
———. 2002. Invariant subpixel material detection in hyperspectral imagery. IEEE Transactions
on Geoscience and Remote Sensing 40(3): 599–608.
Theiler, J., D. Lavernier, N.R. Harvey, S.J. Perkins and J.J. Szymanski. 2000a. Using blocks of
skewers for faster computation of pixel purity In Proceedings of SPIE, vol. 4132, 61–71.
Theiler, J., J. Frigo, M. Gokhle, and J.J. Szymanski. 2000b. FPGA implementation of pixel purity
index method. In Proceedings of SPIE, vol. 4132.
Tompkins, S., J.F. Mustarrd, C.M. Pieters, and D.W. Forsyth. 1997. Optimization of targets for
spectral mixture analysis. Remote Sensing of Environment 59: 472–489.
Tou, J.T., and R.C. Gonzalez. 1974. Pattern recognition principles, 92–94. Reading, MA: Addi-
son-Wesley.
Tu, T.M. 2000. Unsupervised signature extraction and separation in hyperspectral images: a noise-
adjusted fast independent component analysis approach. Optical Engineering 39(4): 897–906.
Tu, T.M., C.-H. Chen, and C.-I Chang. 1997. A posteriori least squares orthogonal subspace
projection approach to weak signature extraction and detection. IEEE Transactions on Geo-
science and Remote Sensing 35(1): 127–139.
Tu, T.M., H.C. Shy, C.-H. Lee, and C.-I Chang. 1999. An oblique subspace projection to mixed
pixel classification in hyperspectral images. Pattern Recognition 32(8): 1399–1408.
Tu, T.-M., C.H. Lee, C.S. Chiang, and C.P. Chang. 2001. A visual disk approach for determining
data dimensionality in hyperspectral imagery. Proceedings of National Science Council 25(4):
219–231.
Tzou, K.H. 1987. Progressive image transmission: a review and comparison of techniques. Optical
Engineering 26(7): 581–589.
Valencia, D., A. Plaza, M.A. Vega-Rodriguez, and R.M. Perez. 2005. FPGA design and imple-
mentation of a fast pixel purity index algorithm for endmember extraction in hyperspectral
imagery. In SPIE East, Boston.
Van Veen, B.D., and K.M. Buckley. 1988. Beamforming: a versatile approach to spatial filtering.
IEEE ASSP Magazine 5: 4–24.
Wang, J. 2006a. Applications of independent component analysis to hyperspectral data exploita-
tion. Ph.D. dissertation, Department of Computer Science and Electrical Engineering, Univer-
sity of Maryland, Baltimore County, Baltimore, MD.
Wang, S. 2006b. Statistical signal processing applications to hyperspectral signature character-
ization. Ph.D. dissertation, Department of Computer Science and Electrical Engineering,
University of Maryland, Baltimore County, Baltimore, MD.
Wang, J., and C.-I Chang. 2004. A uniform projection-based unsupervised detection and classi-
fication for hyperspectral imagery. In IEEE international geoscience and remote sensing
symposium, Alaska, September 20–24.
———. 2005. Dimensionality reduction by independent component analysis for hyperspectral
image analysis. In IEEE International geoscience and remote sensing symposium, Seoul,
Korea, July 25–29.
———. 2006a. A low probability detector-based unsupervised background suppression, target
detection and classification for hyperspectral imagery. In Recent advances in hyperspectral
signal and image processing, ed. C.-I Chang, 141–169. Trivandrum: Research Signpost, India.
———. 2006b. Independent component analysis-based dimensionality reduction with applica-
tions in hyperspectral image analysis. IEEE Transactions on Geoscience and Remote Sensing
44(6): 1586–1600.
678 References
———. 1999b. N-finder: an algorithm for fast autonomous spectral endmember determination in
hyperspectral data. In Image spectrometry V, Proceedings of SPIE, vol. 3753, 266–277.
———. 2004. A proof of the N-FINDR algorithm for the automated detection of endmembers in a
hyperspectral image. In Proceedings of SPIE, vol. 5425, 31–41.
Winter, E.M., M.J. Schlangen, A.B. Hill, C.G. Simi, and M.E. Winter. 2002. Tradeoffs for real-
time hyperspectral analysis. In Proceedings of SPIE, vol. 4725, Algorithms and technologies
for multispectral, hyperspectral, and ultraspectral imagery VIII, 366–371.
Wong, W.W. 2003. Application of linear algebra. http://convexoptimization.com/TOOLS/gp-r.
pdf.
Wu, C.-C. 2006. Exploration of methods of estimation on number of endmember.MS thesis,
Department of Computer Science and Electrical Engineering, University of Maryland, Balti-
more County, Baltimore, MD.
———. 2009. Design and analysis of maximum simplex volume-based endmember extraction
algorithms. Ph.D. dissertation, Department of Computer Science and Electrical Engineering,
University of Maryland, Baltimore County, Baltimore, MD.
Wu, C.-C., and C.-I Chang. 2006. Automatic algorithms for endmember extraction. In SPIE
conference imaging spectrometry XI, SPIE symposium on optics and photonics, vol. 6302,
13–17 August, San Diego, CA.
———. 2012. Iterative pixel purity index. In 4th IEEE GRSS workshop on hyperspectral image
and signal processing—evolution in remote sensing (WHISPERS), 12–14 June, Shanghai,
China.
Wu, C.-C., W. Liu, H. Ren, and C.-I Chang. 2007. A comparative study and analysis between
vertex component analysis and orthogonal subspace projection for endmember extraction. In
Algorithms and technologies for multispectral, hyperspectral, and ultraspectral imagery XIII,
SPIE defense and security symposium, Orlando, FL, April 9–13.
Wu, C.C., S. Chu, and C.-I Chang. 2008. Sequential N-FINDR algorithm. In SPIE conference on
imaging spectrometry XIII, August 10–14, San Diego.
Wu, C.-C., C.S. Lo, and C.-I Chang. 2009. Improved process for use of a simplex growing
algorithm for endmember extraction. IEEE Geoscience and Remote Sensing Letters 6(3):
523–527.
Wu, C.-C., H.M. Chen, and C.-I Chang. 2010. Real-time N-finder processing algorithms. Journal f
Real-Time Image Processing 7(2): 105–129. doi:10.1007/s11554-010-0151-z.
Wu, C.C., G.S. Huang, K.H. Liu, and C.-I Chang. 2013. Real-time progressive band processing of
modified fully abundance-constrained spectral unmixing. In 2013 I.E. international geoscience
and remote sensing symposium, 21–26 July, Melbourne, Australia.
Xiong, W. 2011. Estimation of effective spectral dimensionality for hyperspectral imagery.
Department of Computer Science and Electrical Engineering, University of Maryland, Balti-
more County, Baltimore, MD.
Xiong, W. C.T. Tsai, C.W. Yang, and C.-I Chang. 2010. Convex cone-based endmember extrac-
tion for hyperspectral imagery. In SPIE, vol. 7812, San Diego, CA, August 2–5.
Xiong, W., C.-C. Wu, C.-I Chang, K. Kapalkis, and H.M. Chen. 2011. Fast algorithms to
implement N-FINDR for hyperspectral endmember extraction. IEEE Journal of Selected
Topics in Applied Earth Observations and Remote Sensing 4(3): 545–564.
Yang, S., J. Wang, C.-I Chang, J.L. Jensen, and J.O. Jensen. 2004. Unsupervised image classifi-
cation for remotely sensed imagery. In 49th Annual meeting, SPIE international symposium on
optical science and technology, imaging spectrometry IX (AM105), Denvor, CO, August 2–4.
Yang, H., J. An, and C. Zhu. 2014. Subspace-projection-based geometric unmixing for material
quantification in hyperspectral imagery. IEEE Journal of Selected Topics in Applied Earth
Observations and Remote Sensing 7(6): 1966–1975.
Yu, X., I.S. Reed, and A.D. Stocker. 1993. Comparative performance analysis of adaptive
multispectral detectors. IEEE Transactions on Signal Processing 41(8): 2639–2656.
Zare, A., and P. Gader. 2007. Sparsity promoting iterated constrained endmember detection in
hyperspectral imagery. IEEE Geoscience and Remote Sensing Letters 4(3): 446–450.
680 References
Zenzo, S.D., R. Bernstein Degloria, and H.C. Kolsky. 1987a. Gaussian maximum likelihood and
contextual classification algorithms for multicrop classification. IEEE Transactions on Geo-
science and Remote Sensing GE-25(6): 805–814.
Zenzo, S.D., R. Bernstein, S.D. Degloria, and H.C. Kolsky. 1987b. Gaussian maximum likelihood
and contextual classification algorithms for multicrop classification experiments using the-
matic mapper and multispectral scanner sensor data. IEEE Transactions on Geoscience and
Remote Sensing GE-25: 815–824.
Zhang, Y., and M.C. Amin. 2001. Array processing for nonstationary interference suppression in
DS/SS communications using subspace projection techniques. IEEE Transactions on Signal
Processing 49(12): 3005–3014.
Zhang, X., and C.H. Chen. 2002. New independent component analysis method using higher order
statistics with applications to remote sensing images. Optical Engineering 41: 1717–1728.
Zhao, L., S.Y. Chen, M. Fan, and C.-I Chang. 2014. Endmember-specified virtual dimensionality
in hyperspectral imagery. In 2014 I.E. international geoscience and remote sensing symposium
(IGARSS), Quebec Canada, July 13–18.
Zortea, M., and A. Plaza. 2009. A quantitative and comparative analysis of different
implementations of N-FINDR: a fast endmember extraction algorithm. IEEE Geoscience
and Remote Sensing Letters 6(4): 787–791.
Index
A BIL, 199–205
Abundance nonnegativity constraint (ANC), CBCRM, 423–425
265, 320, 358 CK-AD, 161
Abundance sum-to-one constraint computational complexity, 169–171
(ASC), 265–267, 320, 358 b-simulated image, 171–172
Adaptive GSVA (AGSVA), 336–337 Intel i5-2500 3.3 GHz CPU and 8 GB
Adaptive linear mixing model (ALMM), 270 RAM, 177–178
Adaptive recursive hyperspectral sample mixed and subpanel pixels, 171–172
processing of LSMA per pixel vector, 178–179
(ARHSP-LSMA), 264 target embeddedness, 174–177, 180
Gaussian distribution, 280 target implanted, 172–175, 180
implementation, 276 computing time analysis, 439–441, 446
linear spectral unmixing-based finding CR-AD, 161
signatures, 278–279 features, 158
OSP-based finding signatures, 277–278 GUI design, 444, 447
posteriori probability distribution, 281 HYDICE image, 185–193
randomized decision rule, 281 area under curve, 434, 435, 437–439
Airborne Visible InfraRed Imaging AVIRIS, 435, 440
Spectrometer (AVIRIS), 343, 578 detected abundance fractions, 429–431
anomaly detection, 181–185 K-AD detection map, 160–161, 436,
CEM, 149–156 442, 445–446
virtual dimensionality processed bands, 429, 432–436
LCVF scene, 18–19, 115 R-AD detection map, 161, 435, 436,
2OS-and HOS-based methods, 116 441, 443–444
reflectance and radiance data, 112–114 scene and ground truth map, 428–429
scene, Cuprite, 15–18, 112 3D plots, 429, 432
signal energy and strength, MOCA and matrix inverse calculation
ATGP, 116–118 causal sample covariance/correlation
Analytical Imaging and Geophysics (AIG), 544 matrix, 162–164
Anomaly detection (AD) sample correlation matrix identity, 166
AVIRIS data, 181–185 Woodbury matrix identity, 165
background suppression, 182–184, PHBP-AD, 422–423
192–199 real-time causal K-AD, 168–169