You are on page 1of 167

IMPROVED TARGET RECOGNITION AND TARGET DETECTION

ALGORITHMS USING HRR PROFILES AND SAR IMAGES

A thesis submitted in partial fulfillment


of the requirements for the degree of
Master of Science in Engineering

By

ANINDYA SANKAR PAUL

B.E., Manipal Institute of Technology, India, 2001

2003

Wright State University


Wright State University
School of Graduate Studies
September 8, 2003

I HEREBY RECOMMEND THAT THE THESIS PRESENTED UNDER MY


SUPERVISION BY Anindya Sankar Paul ENTITLED Improved Target Recognition and
Target Detection Algorithms using HRR profiles and SAR images BE ACCEPTED IN
PARTIAL FULFILMENT OF THE REQUIREMENTS FOR THE DEGREE OF Master
Of Science in Electrical Engineering.

______________________

Arnab K. Shaw, Ph.D.


Thesis Director

______________________

Fred Garber, Ph.D.


Department Chair

Committee on
Final Examination

___________________________

Arnab K. Shaw, Ph.D.

___________________________

Atindra K. Mitra, Ph.D.

___________________________

Fred Garber, Ph.D.

___________________________

Kefu Xue, Ph.D.

___________________________

Joseph F. Thomas, Jr., Ph.D.


Dean, School of Graduate Studies
ABSTRACT
Paul Anindya S. M.S.Eg., Department of Electrical Engineering, Wright State University,
2003: Improved Target Recognition and Target Detection Algorithms using HRR profiles
and SAR images.

In this thesis, a new algorithm to improve automatic target recognition techniques on

High Range Resolution (HRR) Profiles is presented and also a number of ways are

investigated for target detection using Synthetic Aperture Radar (SAR) images.

A new 1-D hybrid Automatic Target Recognition (ATR) algorithm is developed

for sequential High Range Resolution (HRR) radar signatures. The proposed hybrid

algorithm combines Eigen-Template based Matched Filtering (ETMF) and Hidden

Markov modeling (HMM) techniques to achieve superior HRR-ATR performance. In the

proposed hybrid approach, each HRR test profile is first scored by ETMF that is then

followed by independent HMM scoring. The first ETMF scoring step produces a limited

number of “most likely” models that are target and aspect dependent. These reduced

numbers of models are then used for improved HMM scoring in the second step. Finally,

the individual scores of ETMF and HMM are combined using Maximal Ratio Combining

to render a classification decision. Classification results are presented for the MSTAR

data set via ROC curves.

An ultra-wideband (UWB) synthetic aperture radar (SAR) simulation technique

that employs physical and statistical models is developed and presented. This joint

iii
physics/statistics based technique generates images that have many of the “blob-like” and

“spiky” clutter characteristics of UWB radar data in forested regions while avoiding the

intensive computations required for the implementation of low-frequency numerical

electromagnetic simulation techniques. Comparative results from three SVD-based

subspace filtering approaches on target detection algorithms are reported. These

approaches are denoted as “Energy-Normalized SVD”, “Condition-Statistics SVD”, and

“Terrain-Filtered SVD”. Approaches towards developing “self-training” algorithms for

UWB radar target detection are investigated using the results of this simulation process.

iv
CONTENTS

1: Introduction 1

1.1 ATR/Target Detection review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.1.1 A review of ATR/Target detection . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.1.2 Moving Target Indicator (MTI) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2

1.1.3 Synthetic Aperture Radar (SAR) . . . . . . . . . . . . . . . . . . . . . . . . . . . 4

1.1.4 High Range Resolution Radar (HRR) . . . . . . . . . . . . . . . . . . . . . . . 6

1.2 Background and previous work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8

1.3 Background on Automatic Target Recognition using HRR profiles . . . . . . 11

1.4 Background on Target Detection on SAR images . . . . . . . . . . . . . . . . . . . . 15

1.5 Thesis Contribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

1.6 Thesis Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17

2: Robust HRR Radar Target Identification by Hybridization of HMM and 19

Eigen Template based Matched Filtering

2.1 ETMF Approach of training and classification . . . . . . . . . . . . . . . . . . . . . . 19

2.1.1 HRR Data Generation and Preprocessing . . . . . . . . . . . . . . . . . . . . . . 20

2.1.2: Normalization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

v
2.1.3: Alignment of HRR Profiles in Range . . . . . . . . . . . . . . . . . . . . . . . . . 22

2.1.4: Eigen-analysis of HRR data for training . . . . . . . . . . . . . . . . . . . . . . 22

2.1.5: Unknown Target Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25

2.1.6: Modified Normalization and Centroid Alignment . . . . . . . . . . . . . . . 26

2.2: HMM approach of training and classification . . . . . . . . . . . . . . . . . . . . . . . 30

2.2.1: Discrete Hidden Markov Model Introduction . . . . . . . . . . . . . . . . . . 31

2.2.1.1: Elements of HMM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

2.2.1.2: Three Basic Problems for HMM . . . . . . . . . . . . . . . . . . . . . . . 34

2.2.1.3: Model Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

2.2.1.4: Optimal State Sequence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

2.2.1.5: Parameter Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

2.2.2: HMM Operation steps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

2.2.2.1: Framing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

2.2.2.2: Clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37

2.2.2.3: HMM Training . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

2.2.2.3.1: Model Optimum Parameter estimation . . . . . . . . . . . . 40

2.2.2.3.2: Optimum state sequence estimation . . . . . . . . . . . . . . 45

2.2.2.4: HMM Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

2.2.2.4.1: The Forward Procedure . . . . . . . . . . . . . . . . . . . . . . . . 48

2.2.2.4.2: The Backward Procedure . . . . . . . . . . . . . . . . . . . . . . 52

2.3: Approach of Combination between ETMF and HMM . . . . . . . . . . . . . . . . 53

2.3.1: Motivation for Hybrid approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53

2.3.2: Proposed Solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

vi
2.3.2.1: Necessity of developing modified hybridization in ETMF ATR 56

2.3.2.2: Number of Subset Model selection and Weight calculation . . . 58

2.3.2.2.1: Subset Model Selection . . . . . . . . . . . . . . . . . . . . . . . . 58

2.3.2.2.2: Weight determination . . . . . . . . . . . . . . . . . . . . . . . . . 61

2.4: Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64

2.4.1: Data Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64

2.4.2: ETMF Simulation Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64

2.4.2.1: Formation of Template Profiles . . . . . . . . . . . . . . . . . . . . . . . . 64

2.4.2.2: Classification using Matched Filter Technique . . . . . . . . . . . . 65

2.4.3: HMM Simulation Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

2.4.3.1: Framing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

2.4.3.2: Clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

2.4.3.3: Training and Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

2.4.4: Single Look Case . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

2.4.4.1: Forced Decision Result . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67

2.4.4.2: Classification in Unknown Target Scenario . . . . . . . . . . . . . . . 73

2.4.4.3: Computational Cost . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86

2.4.5: Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87

3: Time Recursive Multiple Hypothesis Testing 88

3.1: Theory Development . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88

3.2: Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91

3.2.1: ETMF classifier Simulation results . . . . . . . . . . . . . . . . . . . . . . . . . . 92

vii
3.2.1.1: Description of MSTAR Data . . . . . . . . . . . . . . . . . . . . . . . . . . . 92

3.2.1.2: Multilook Performance Results . . . . . . . . . . . . . . . . . . . . . . . . . 93

3.2.2: Hybrid Classifier Simulation results . . . . . . . . . . . . . . . . . . . . . . . . . 95

3.2.3: Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98

4: Improved SAR Target Detection Using Subspace Filtering 99

4.1: Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100

4.2: Ultra-wideband Radar simulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103

4.3: Eigen-Analysis of SAR image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106

4.4: Clutter Suppression Capability of SVD . . . . . . . . . . . . . . . . . . . . . . . . . . . 107

4.5: SVD based SAR Target Detection Algorithms . . . . . . . . . . . . . . . . . . . . . . 108

4.5.1: Energy Normalized SVD (EN-SVD) . . . . . . . . . . . . . . . . . . . . . . . . . 108

4.5.1.1: Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108

4.5.1.2: EN-SVD Training Phase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110

4.5.1.3: EN-SVD Testing Phase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113

4.5.2: Condition-Statistic SVD (CS-SVD) . . . . . . . . . . . . . . . . . . . . . . . . . . 114

4.5.3: Terrain-Filtered SVD (TF-SVD) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115

4.5.3.1: Motivation and Kernel formation . . . . . . . . . . . . . . . . . . . . . . . . 115

4.5.3.2: Implementation steps of TF-SVD . . . . . . . . . . . . . . . . . . . . . . . 118

4.5.4: Simulation results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119

4.5.4.1: UWB SAR simulated image specification . . . . . . . . . . . . . . . . 120

4.5.4.2: Performance Comparison of Target Detection algorithms . . . . 122

4.5.4.2.1: Performance Comparison in Offline training mode . . 122

viii
4.5.4.2.2: Performance Comparison in Self-Training mode . . . . 129

4.5.4.3: Performance Comparison of various techniques . . . . . . . . . . . 132

4.5.4.4: Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135

5: Summary and Future work 137

5.1: Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137

5.1.1: Hybrid ATR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137

5.1.2: Time Recursive Sequential ATR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139

5.1.3: SAR target detection algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139

5.2: Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140

Bibliography 141

ix
List of Figures

1.1 Side looking radar system geometry . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

2.1 Eigen-Template Generation from detected HRR profiles . . . . . . . . . . . 21

2.2 Distribution of Singular values for MSTAR target T72, 1000 sector . . . 23

2.3 Implementation of the Correlation Classifier . . . . . . . . . . . . . . . . . . . . . 25

2.4 Observation and Template Profiles are shown in shaded and blank 27

boxes of different lengths . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

2.5 Shift = -8 of Centroid . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

2.6 Shift = 0 aligned of Centroid . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

2.7 Shift = +8 of the Centroid . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

2.8 Simplified Block diagram of target recognition using HMM . . . . . . . . 30

2.9 Two states Hidden Markov Model with two output symbols, V1 and 32

V2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

2.10 Flow diagram for LBG clustering algorithm . . . . . . . . . . . . . . . . . . . . . 38

2.11 Illustration of the sequence of operation required for the computation

of the joint event that the system is in state Si at time t and state Sj at 43

time t+1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

2.12 Baum-Welch learning algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44

x
2.13 Block Diagram of a Designed HMM recognizer . . . . . . . . . . . . . . . . . 47

2.14 Illustration of the sequence of operations required for the computation

of the (a) forward variable and (b) backward variable . . . . . . . . . . . . . . 49

2.15 State lattice used to derive the forward/backward recursion . . . . . . . . . 51

2.16 Data flow in the proposed hybrid algorithm . . . . . . . . . . . . . . . . . . . . . 57

2.17 In 100 aspect case, this plot shows the HMM recognition rate with 60

number of HMM model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

2.18 This figure is used to determine the most effective W1/ W2 so that the 62

combined ETMF+HMM recognition rate is the highest . . . . . . . . . . . .

2.19 Bar plot representation of ETMF, HMM and Hybrid classifier 72

performances . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

2.20 ROC curves for Probability of declaration vs Conditional Probability 78

of Correct Classification (Single profile) . . . . . . . . . . . . . . . . . . . . . . . .

2.21 ROC curves for Probability of False Alarm vs Probability of 78

Declaration (Single profile) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

2.22 ROC curves for Probability of declaration vs Conditional Probability 80

of Correct Classification (3 profile average) . . . . . . . . . . . . . . . . . . . . .

2.23 ROC curves for Probability of False Alarm vs Probability of 80

Declaration (3 profile average) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

2.24 ROC curves for Probability of declaration vs Conditional Probability 81

of Correct Classification (5 profile average) . . . . . . . . . . . . . . . . . . . . .

2.25 ROC curves for Probability of False Alarm vs Probability of 81

Declaration (5 profile average) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

xi
2.26 ROC curves for Probability of Declaration vs Conditional Probability 82

of Correct Identification (Combined result) . . . . . . . . . . . . . . . . . . . . . .

2.27 ROC curves for Probability of False Alarm vs Probability of 82

declaration (Combined result) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

2.28 ROC curves for Probability of False Alarm vs Conditional Probability 83

of Correct Classification (Combined result) . . . . . . . . . . . . . . . . . . . . .

3.1 Block diagram for time recursive multiple hypothesis Combiner . . . . . 91

3.2 ROC curves for Probability of detection vs Conditional Probability of 93

Correct Classification (time recursive ETMF) . . . . . . . . . . . . . . . . . . . .

3.3 ROC curves for Probability of False Alarm vs Probability of 94

declaration (time recursive ETMF) . . . . . . . . . . . . . . . . . . . . . . . . . . . .

3.4 ROC curves for Probability of False Alarm vs Probability of 95

declaration (time recursive hybrid) . . . . . . . . . . . . . . . . . . . . . . . . . . . .

3.5 ROC curves for Probability of Declaration vs Conditional Probability 96

of Correct Identification (time recursive hybrid) . . . . . . . . . . . . . . . . . .

3.6 ROC curves for Probability of False Alarm vs Conditional Probability 96

of Correct Identification (time recursive hybrid) . . . . . . . . . . . . . . . . . .

4.1 Block Diagram for UWB SAR Simulation . . . . . . . . . . . . . . . . . . . . . . 102

4.2 Eigen Spectrum of Target and Clutter blob . . . . . . . . . . . . . . . . . . . . . . 110

4.3 SAR image feature extraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111

4.4 Filter kernel for “Terrain-Filtered SVD” . . . . . . . . . . . . . . . . . . . . . . . . 116

4.5 Sample Filter Histogram for TF-SVD . . . . . . . . . . . . . . . . . . . . . . . . . . 117

xii
4.6 Sample UWB SAR Simulation test Image . . . . . . . . . . . . . . . . . . . . . . . 119

4.7 Sample UWB SAR Simulation Clutter only image . . . . . . . . . . . . . . . . 120

4.8 UWB SAR simulation image after performing EN-SVD . . . . . . . . . . . 122

4.9 UWB SAR simulation image after performing CS-SVD . . . . . . . . . . . . 124

4.10 UWB SAR simulation image after performing Euclidean masking 125

operation in TF-SVD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

4.11 Final UWB SAR simulation image after performing TF-SVD . . . . . . . 126

4.12 Performance comparison of EN-SVD, CS-SVD and TF-SVD in 127

offline training-real time testing mode . . . . . . . . . . . . . . . . . . . . . . . . . .

4.13 Performance comparison of EN-SVD, CS-SVD and TF-SVD shown 128

in logarithmic scale in offline training-real time testing mode . . . . . . .

4.14 Performance comparison of EN-SVD, CS-SVD and TF-SVD in self- 130

train mode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

4.15 Performance comparison of EN-SVD, CS-SVD and TF-SVD in self- 132

train mode (logarithmic scale) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

4.16 ROC Performance comparison of various techniques . . . . . . . . . . . . . . 133

4.17 ROC Performance comparison of various techniques (logarithmic 134

scale) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

xiii
List of Tables

I Organization of a Confusion Matrix . . . . . . . . . . . . . . . . . . . . . . . . 68

II Summary of Forced Decision Results . . . . . . . . . . . . . . . . . . . . . . . 69

III Confusion matrix for ETMF with single profile testing . . . . . . . . . 69

IV Confusion matrix for HMM with single profile testing . . . . . . . . . 69

V Confusion matrix for Hybrid algorithm with single profile testing . 70

VI Confusion matrix for ETMF with three profile average testing . . . . 70

VII Confusion matrix for HMM with three profile average testing . . . . 70

VIII Confusion matrix for Hybrid algorithm with three profile average 70

testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

IX Confusion matrix for ETMF with five profile average testing . . . . 70

X Confusion matrix for HMM with five profile average testing . . . . . 70

XI Confusion matrix for Hybrid algorithm with five profile average 71

testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XII Evaluation parameter computation from the confusion matrix . . . . 74

XIII Confusion matrix (Unknown rejection threshold about 0.6) for 75

ETMF based classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XIV Confusion matrix (Unknown rejection threshold about 0.6) for 76


Hybrid classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

xiv
XV Confusion matrix (Unknown rejection threshold about 0.6) for 76

ETMF based classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XVI Confusion matrix (Unknown rejection threshold about 0.6) for 76

Hybrid classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XVII Confusion matrix (Unknown rejection threshold about 0.6) for 77

ETMF based classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XVIII Confusion matrix (Unknown rejection threshold about 0.6) for 77

Hybrid classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XIX Conditional Probability of Correct Classification of Hybrid and 84

ETMF classifiers at Pd = 0.7 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XX Probability of Declaration of Hybrid and ETMF classifiers at Pfa = 84

0.4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XXI Conditional Probability of Correct Classification of Hybrid and 85

ETMF classifiers at Pfa = 0.4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XXII Improvement of Probability of Declaration of Hybrid classifiers 97

due to time recursive multilook approaches at Pfa = 0.4 . . . . . . . . .

XXIII Improvement of Conditional Probability of Correct Classification 97

of Hybrid classifiers due to time recursive multilook approaches at

Pd = 0.7 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

XXIV Improvement of Conditional Probability of Correct Classification 97

of Hybrid classifiers due to time recursive multilook approaches at

Pfa = 0.4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

xv
ACKNOWLEDGEMENTS

It is my pleasure to acknowledge and thank people who helped me accomplish my

goal to pursue graduate studies. First I would like to thank my parents, Goutam K. Paul

and Sipra Paul, for their constant support and encouragement. They have made lots of

sacrifices to help me with my education, for which I will always be grateful.

I would like to thank Professor Arnab K. Shaw, WSU, and Dr. Atindra K. Mitra,

WPAFB/SNRR, for their guidance and encouragement throughout my thesis. I would

also like to thank Dr. Kefu Xue and Dr. Fred Garber for agreeing to be on my thesis

committee.

I would also like to thank Thomas L. Lewis, WPAFB/SNRR for assisting me to

generate the simulated target-clutter image.

I would like to acknowledge my friends, Koel Das and Sivaram Bandaru for

helping me with my thesis preparation.

Lastly, I would wish to thank all the faculty members of the Electrical

Engineering Department at Wright State University for their generous help and

tremendous support through the course of my M.S. program.

xvi
1: INTRODUCTION
The objective of Automatic Target Recognition (ATR) algorithms is to correctly identify

an unknown target from sensed radar signatures [1-4], whereas in target detection case,

the requirement is to detect target from clutter. The need for ATR and target detection

technology is evident from various “friendly fire” incidents. The most popular algorithm

for ATR is the template-matching algorithm. Given a sensed signature from an unknown

target, the ATR systems compare the observed signatures with a set of stored target

hypotheses. The target decision is based on some form of optimum similarity between the

observed signature and one of the stored targets. Template based ATR provides

encouraging results as demonstrated in the work of Novak, et al. [5], Mirkin [6] and

many others [7-9]. Whereas in target detection case, the classifier is trained to determine

the threshold, which is a discriminant factor between target and clutter. Based on this

threshold the classifier will perform target detection while nullifying clutter.

In the next subsections a brief review of ATR/target detection and its background and

previous work are depicted.

1.1: ATR/Target Detection Overview

1.1.1: A Review of ATR/Target Detection

The present era of limited warfare demands precision strikes for reduced risk and cost

efficient operation with minimum possible collateral damage. In order to meet such

1
exacting challenges, Automatic Target Recognition (ATR)/Target Detection capability is

becoming increasingly important to the Defense community. The overall goals are to

analyze image data using digital computers in order to detect, classify and recognize

target signatures automatically, i.e., with minimum possible human assistance. The image

data for processing may be generated by one of many possible imaging sensors including

radar, optical, infrared or others. Hence target detection/recognition is considered to be

one of the most challenging among current research problems because the system

developers have little control over the possible target scenario and the operational

imaging condition [14-17]. Also, compared to the diversity of possible images during

operations, only a relatively smaller subset of images may be available at the

development or training stage. Furthermore, the operational target detection/recognition

algorithms may have to deal with intelligent adversary attempting to defeat the system, as

opposed to amore controlled environment during development.

Traditionally, air to ground acquisition of ground target information is categorized

into two general areas: Moving Target Indication (MTI) and Synthetic Aperture Radar

(SAR) [18-22]. The original purpose for developing these radar technologies had been to

achieve all weather and all day/night imaging, i.e., to transcend traditional photographic

camera based imaging that must rely on sunlight and is susceptible to clouds, fog or

precipitation.

1.1.2: Moving Target Indicator (MTI)

Most surface and airborne radar systems operate in an environment where the clutter

return obscures targets of interest [23]. If the target is moving relative to the clutter it is

2
possible to filter out the undesired clutter return by exploiting the differential doppler

frequency shift produced by relative target to clutter radial motion. Systems following

this principle are called Moving Target Indicator (MTI) radar.

MTI has the capability to detect target reflections [24] having differential radial

motion with respect to the clutter. The clutter causing background may be either terrain,

sea, weather or chaff [25-26]. MTI’s are operated with either fixed based or a moving

platform such as an aircraft or a satellite. Considering detection of low flying aircraft’s,

i.e. the radar is surface based, flying over terrain through possible weather disturbances.

In such an event, MTI rejects the returns from terrain and weather while retaining the

return from the aircraft. This property gives it good detection capabilities for air borne

targets. In cases where the target is surface based, as in Air to Ground ATR application,

the ground clutter are stronger than the expected target return. The ground clutter

extends out to a range where terrain features that cause the clutter are masked due to

earth's curvature. In such cases, the ground clutter extends to the full operating range of

the radar. This makes MTI without any recognition capabilities.

MTI is a mature radar technology that allows airborne sensors to survey large

areas of land and it has coarse target detection and range determination capabilities. It

makes use of target movement for image formation and hence, it is highly effective for

distinguishing moving targets from ground clutter. However, a major drawback of the

MTI technology is its lack of any target recognition capability.

3
1.1.3: Synthetic Aperture Radar (SAR)

Although the major emphasis of this thesis is to utilize HRR profiles, as described in the

previous chapter, it may be pointed out here that data collection as well as most of the

front-end processing for HRR is identical to that of SAR. Hence, in this Section, SAR

image formation is described in considerable detail.

Figure 2.1 shows the side-looking radar system wherein an aircraft carry on-board

a SAR imager [27-28] illuminates a patch of ground having a target with certain

surroundings. The beam of the radar looks out to the side of the aircraft, in a direction

orthogonal to its flight path. This direction of radiation propagation is referred to as the

Range direction and the direction parallel to flight path is called the Cross-range

direction. During the aircraft's movement, it periodically transmits pulses of microwave

energy that impinge on the patch containing the target. Each of these pulses is

subsequently reflected back to and received by the radar, where demodulation is

performed. The assemblage of data collected and pre-processed in this manner is called a

phase history and is passed on to the processor for image reconstruction. The processor

could either be located on the ground or on board the flying aircraft. This processor gives

out as output the electromagnetic reflectivity of the illuminated ground patch. The

reflectivity is a two-dimensional function having dimensions as Range and Cross-range.

Although a SAR picture looks entirely different from an optical photograph, the

key features are easily recognizable. SAR is coherent radar that employs signal

processing and motion compensation to provide a high spatial resolution estimate of the

scenes reflectivity, also commonly known as Radar Cross Section (RCS). Motion sensors

are used to measure platform flight characteristics so that non-ideal flight path generated

4
phase errors can be removed during image formation processing. Platform or target

motion creates scene aspect variations, leading to a differential doppler signature of

scatterers in the antenna footprint. The doppler signatures are subsequently exploited to

achieve enhanced Cross-range resolution. Doppler frequency is 1/(2π(dφ/dt)), where φ =

4πR/λ. This is the fundamental behind SAR imaging concept (also commonly known as

Range/Doppler imaging).

The reasons for using SAR images over optical ones are summarized below.

• It is able to image a surface with very fine resolution of a few meters to coarse

resolution of a few kilometers.

• It can provide imagery to a given resolution independently of altitude, limited only by

the transmitter power available.

• A number of fundamental parameters such as polarization and look angle can be

varied to optimize the system for a specific application.

• Imaging is independent of solar illumination (availability or angle) because the

system provides its own source of illumination.

• It can operate independently of weather conditions if sufficiently long wavelengths

are chosen.

• It operates in a band of electromagnetic spectrum different from the bands used by

visible and infrared (IR) imageries.

5
1.1.4: High Range Resolution (HRR)

MTI and SAR are active Doppler systems that transmit and receive electromagnetic

waveforms in the microwave bands that have superior penetrating capabilities than visual

frequency bands. These radar technologies are being researched and developed over

several decades now and both concepts have some share of strengths and weaknesses.

MTI makes use of target movement for image formation and hence, it is highly effective

for distinguishing moving targets from ground clutter. It is a mature radar technology that

allows airborne sensors to survey large areas of land and it has coarse target detection and

range determination capabilities. However, although very useful for target detection, the

MTI technology lacks target recognition capability. In case of SAR, in contrast, ground

target information is available for processing in both range and cross-range domains, and

it has excellent target recognition and identification capabilities. However, processing

requirements for SAR is considerably high, preventing it from being used as a wide area

surveillance technology.

Unlike SAR and MTI, the HRR technology considered in this work would rely on

processing high resolution `Range Profiles', as distinguished from traditional SAR-ATR

that utilizes SAR image data. Its potential target recognition capability promises to bridge

the gap between the wide area surveillance target detection capabilities of MTI and the

very narrowly focused target identification capabilities of SAR.

HRR images are used to overcome the disadvantages of SAR data whereas

moving targets are concerned. In case of SAR images, the ability to achieve high Cross-

Range resolution is limited by the migration of scatterers into neighboring resolution

cells.

6
Figure 1.1: Side looking radar system geometry

Secondly, even a Cross-Range resolution of 1 ft can require large angular aperture,

resulting in significant blurring due to scattered migration. This becomes evident at low

frequencies since a large coherent processing angle is required for a given Cross-Range

resolution. Moreover, the image blurring becomes significant as the migration of

scatterers approaches the desired resolution.

All these factor make recognition hard for moving targets. In case of HRR

profiles, all the information in range is still present, but the cross-range blurring is not

present. This makes HRR as the most feasible choice as far as moving target is concerned

and HRR radar sensor has wide application in target tracking.

7
1.2: Background and Previous Work

Most research in the field of target recognition address data study, theoretical formulation

and algorithm development. Clearly, important milestones [29-30] have been reached in

these areas. However, barring some notable exceptions [31-33] most existing target

detection/recognition algorithms are meant to be implemented using 2-D Synthetic

Aperture Radar (SAR) image data. These algorithms are critically dependent on

appropriate target and sensor models. The major limitation of detection/identification

using SAR is its failure to recognize correctly in case of moving targets due to blurring

caused in the Cross-range domain. This problem makes SAR-target recognition

unsuccessful in case of moving targets. The other field in which much research is done is

target detection/recognition using Moving Target Indicator (MTI). MTI radar is very

good for detection but fails due to coarse recognition capabilities. In fact, most well

established algorithms are mathematically and computationally so comprehensive that it

would be quite impractical to implement those in on-line applications. This problem

grows when the number of targets to be detected becomes large.

The previous work on ATR encompasses a variety of approaches. SAR-

detection/estimation is one of the most important ones [34-36]. An accurate clutter model

had been suggested for precise target detection [37]. The power spectral density (PSD) of

the clutter was estimated such that a multi-dimensional matched filter could be designed

for detection. Another approach [34] has been used for model-based ATR/detection

techniques. The basic paradigm involves detection and feature extraction such that they

can be used in hypothesis using target identities. If the hypothesis is satisfied, the target is

termed as recognized else it is reformed and used to improve the predicted signature.

8
Morgan, et al. [35] has used the Classical Bayesian detection and decision theory for

model-based ATR. It was proposed that when the model tends to represent the

uncertainties in target type, shape, surround, scatterers and feature extraction, then

classical theory yields model based ATR techniques. The concept was extended to use of

model-based templates for SAR-ATR [36]. Mahalanobis [38] has discussed the use of a

correlation filter in SAR-ATR at the recognition stage.

The previous work on detection/recognition also includes using Multi-resolution

Wavelet Decomposition [39-41]. The Wavelet Transform has been found to be highly

effective for image analysis, data and image compression, feature analysis, and many

other applications [42-44]. It has also been used for speckle reduction of SAR images

[45]. Image compression is achieved by successive Wavelet Decomposition of the image

using a pyramid scheme. Peterson et al. [46] has developed a technique for classifying

different objects in natural imagery by employing a wavelet transform and training a

neural network on certain wavelet transform coefficients in pattern recognition context.

Tagaliarini et al. [47] also incorporated the use of Wavelets with Neural Networks. In his

work, the filter coefficients are a linear combination of wavelet coefficients and can give

rise to an energy distribution that makes recognition easier when compared with that of

conventional wavelets.

The use of Eigen vectors corresponding to an Eigen value problem has been

extensively utilized in many applications like Sonar, SAR etc. Bottcher et al. [48] has

presented the optimal method for term expansions based on the optimum eigen function

related to surface of the object. Here, the conversion of Fredholms integral equation of

first kind was done as an eigen value problem of a related hermition operator. This led to

9
target identification by solving the classical scattering theory of waves. Work on ATR

has also been done using Hidden Markov Models (HMM). HMM has been found to be

extremely successful in speech recognition [49] and it has also found some use in SAR

target detection [50]. Liao et al. [8] extracted features from each of the HRR waveforms

via the RELAX algorithm before feeding those to HMM.

Another approach to detection/recognition is by computer simulation [51],

wherein the elements of the complex system are implemented as interacting software

objects. New methods have been proposed for use as these software objects. The target

recognition is performed by a family of 2D cluster filters. Artificial Intelligence [52] has

been used in ATR applications to reduce the search combinatorics. These methods use

domain specific information for robust physical description of the images.

HRR-ATR has been used to solve the problem of moving target recognition [53-

54]. ATR using HRR profiles has been tried using Neural Networks [55-56]. Yiding et al.

used the property of the distinction of Doppler modulation echo for different targets in

HRR profiles for target recognition. The echo spectral density is obtained by the Fourier

transform. Following that, the choice of the total spectral energy and the four segment

spectral energy as characters is done for use in Neural Networks for ATR. Xun et al. [55]

have used the Matrix Pencil method for scattering centers extraction from full

polarization multi-frequency scattering returns. Feature Extraction is done by using

transient polarization response. Finally, the classification of selected features is done

using Multi-resolution Neural Network. Worrell [56] has used the mean-based templates

for feature extraction. Jacobs et al. [52] has chosen a deterministic Gaussian model for

each Range profile. The likelihood functions under each model for varying orientations

10
and target types are compared. The limit on the orientation estimator performance is

described in terms of Hilbert-Schmidt bound on the estimation error. Stewart et al. [57]

has compared the different classification approaches for HRR profiles. The intrinsic

dimensionality of the signatures was obtained using kth nearest neighbors. The two

classifiers compared were the Gaussian classifier and synthetic discriminant function

(SDF) classifier. In his work, he found that the Gaussian correlation classifier performed

better in presence of white noise while the SDF approach worked better for large angle

bin size.

In a detection/estimation algorithm importance must be given to the fact that how

the target orientation phase behaves to changes in the feature extraction, especially in

case of moving targets.

1.3: Background on Automatic Target Recognition using HRR profiles

For several years, Automatic Target Recognition has been studied for Moving Target

Indicator (MTI) and Synthetic Aperture Radar (SAR) images. MTI and SAR are active

Doppler systems that transmit and receive electromagnetic waveforms in the microwave

bands that have superior penetrating capabilities than visual frequency bands. Though

they are much superiors to optical images they have certain drawbacks when used for

recognition of moving targets. MTI makes use of target movement for image formation

and hence, it is highly effective for distinguishing moving targets from ground clutter but

it lacks target recognition capabilities. In case of SAR, in contrast, ground target

information is available for processing in both range and cross-range domains, and it has

excellent target recognition and identification capabilities. However, processing

11
requirements for SAR is considerably high, preventing it from being used as a wide area

surveillance technology. Moreover, the performance of SAR target detection algorithms

degrade when the target is moving because SAR images cannot be formed properly for

moving targets due to blurring caused in the cross-range domain.

Unlike SAR and MTI, the HRR technology would rely on processing High Range

Resolution (HRR) radar signatures, as distinguished from traditional SAR-ATR that

utilizes SAR image data. The information contained in this signature is the radar

scattering characteristics of the target as a function of range along the line of sight of the

radar. It’s potential target recognition capability promises to bridge the gap between the

wide area surveillance target detection capabilities of MTI and the very narrowly focused

target identification capabilities of SAR. Also there is considerable saving in front end

processing in HRR profile generation which require 1-D FFT operation as opposed to

SAR’s use of 2-D FFT.

The primary difficulty associated with the HRR sensor for ATR is that it collapses

three-dimensional information into a single dimension, as opposed to 2D information in

SAR, making HRR-ATR a more challenging task. Recently Target Detection using HRR

profiles achieved lots of attention in literature. Nguyen et al. [7] developed a

superresolution technique for HRR ATR with High Definition Vector Imaging (HDVI),

where a super-resolution technique is applied to the HRR profiles before the profiles are

passed through ATR classification. A statistical feature based classifier developed by

Mitchell and Westerkamp [9] for robust HRR radar target identification showed that the

amplitude and location of HRR signature peaks could be used as features for target

classification.

12
Currently, one of the priority research initiatives of the Air Force is to develop an

advanced air-to-ground HRR ATR program. The ultimate program objective is to

transition mature HRR-ATR technology into operational Air Force airborne attack and

surveillance platforms. The new HRR-ATR technology can be applied into a system

approach and it is expected to vastly improve Air Force's ability to detect, recognize, as

well as identify time-critical military targets. ATR performance with HRR is found to be

excellent for stationary targets, as discussed in the later chapters. It is expected to be

superior for moving targets which cause blurring Synthetic Aperture Radar (SAR) images

making recognition a difficult task.

Research on HRR-ATR requires a multifaceted approach is essential in order to

harness recent advances from multiple disciplines. At the initial stage, complete

characterization of the HRR-profile data was conducted encompassing both theoretical

and implementation aspects. This included though not limited to, correlation analysis,

histogram analysis, sector generation and matching, feature extraction, principal

component analysis, signature generation, recognition using Matched Filtering and Least

Squares. Once the interpretation of the basic characteristics of the HRR profiles was

complete, the accumulated insights were eventually gathered systematically in the ATR

algorithms developed. Different ATR approaches were studied to compare the

performance of different algorithms.

Our previous work [11-12,58-59] demonstrated that effective HRR-ATR

performance can be achieved if the training templates are formed via Singular Value

Decomposition (SVD) of detected HRR profiles and the classification is performed using

Normalized Matched Filtering (MF). It was demonstrated in [11-12,58-59] that a

13
significant proportion (>90%) of target energy is accounted for by the dominant

Eigenvector of the range-space correlation matrix. More interestingly, it was shown that

the range and angle basis spaces are numerically decoupled in the form of left and right

eigenvectors, respectively. This enabled us to exploit the decoupled range information

exclusively for the purpose of target recognition. The theoretical results were also

presented to demonstrate that the range space eigenvectors constitute the "optimal"

features in the range domain. Basis space decomposition via SVD is also shown to be

useful for suppression of clutter from measured profile data by eliminating the

eigenvectors corresponding to smaller singular values, which represent noise or clutter

sub-spaces. In [12], it was demonstrated certain limitations of the use of Power

Transform when the observation profiles are noisy. Specifically, it was shown that

significant signature information might be lost due to the application of Power Transform

on detected noisy profiles, leading to considerable reduction in ATR performance.

Hybridization of multiple optimization techniques has also been attempted for

HRR ATR. In [58], the entire 360-degree of a target vehicle circumference was divided

into several optimum-sized sub-targets and templates were constructed from these sub-

targets. Then the result of template matching was combined using Bayesian updating to

arrive at the final target classification.

Earlier research works on HRR-ATR focused on simulated XPATCH data [59-

60]. But this thesis concentrates on recognition of stationary targets using the MSTAR

data.

14
1.4: Background on Target Detection on SAR images

Synthetic Aperture Radar (SAR) imagery is commonly used as a tool in detecting,

classifying, recognizing and possibly identifying mobile or stationary targets.

Recognizing target from SAR images is an important, yet challenging application if the

target is hindered under outliers. To date, the authors have engaged in research and

published results on a number of approaches to target detection [11] in the ultra-

wideband (UWB) SAR area. The approaches presented include detailed discussion on a

number of aspects of ultra-wideband radar target detection and algorithm development. A

bi-modal technique for modeling ultra-wideband radar clutter was proposed. An approach

to developing a new class of rank order filters, known as, “discontinuity filter” for ultra-

wideband radar target detection applications was presented. These approaches mainly

concentrate on the investigation of algorithms that implement elaborate off-line training

as well as the development of rank-order filtering algorithms that are designed for basic

UWB SAR sensor phenomenology and at the same time do not require an extensive off-

line training step. Both of these approaches have been shown to generate an acceptable

level of performance under certain conditions that are of interest for UWB SAR

applications.

1.5: Thesis Contribution

In this thesis HRR-ATR performance has been analyzed for Moving & Stationary Target

Acquisition and Recognition (MSTAR) data using hybridization of Hidden Markov

Model (HMM) and Eigen Templates based Normalized Matched filter (ETMF) based

ATR algorithm. The following contributions were made to the existing ATR techniques:

15
A new hybrid 1-D ATR approach is presented where the HRR test profiles are first

scored by ETMF and then the most likely HMM models determined by ETMF are

used for HMM scoring at the second step. Final ATR decision is based on proper

weight combination of the two individual scores. Performance comparison results are

provided for Forced Decision as well as for Unknown Target scenarios. The unknown

target scenario is simulated using the Leave One Out Method (LOOM) [4]. The

performances of ATR algorithms are compared in terms of the Receiver Operating

Characteristics (ROC) curves.

In this paper, the proposed hybrid algorithm is extended for moving target case,

which will facilitate simultaneous, multiple target tracking. For Continuous-ID and

joint tracking, the single look ETMF and HMM hybrid technique needs to be applied

time-recursively to update the multiple ID hypothesis as new range profiles are

observed over time. The proposed approach would be a recursive version of the

block-processed stationary multi-look approach [2] that has shown considerable

success in identifying stationary targets.

In addition to ATR, considerable improvement in SAR target detection field is also

performed.

A set of results from an investigation of an approach denoted as “self-training

algorithms for ultra-wideband SAR target detection” is presented. Under this

16
approach, a number of categories of algorithms are investigated that implement “self-

training” procedures. These procedures are developed such that a set of localized

regions within a given SAR image are sampled in real-time for purposes of obtaining

low-order and robust real-time clutter models. These real-time models are applied in

a sliding-window type target detection paradigm for clutter cancellation and target

detection. Results are presented from the analysis of three new categories of

algorithms that were developed specifically for this investigation. These three

categories of algorithms denoted as “Energy-Normalized SVD” (EN-SVD),

“Condition-Statistic SVD” (CS-SVD), and “Terrain-Filtered SVD” (TF-SVD) are

generating satisfactory simulation results for severe UWB SAR impulsive-type

clutter. Though offline training is required for both EN-SVD and CS-SVD to perform

satisfactory level, the third approach TF-SVD is a notable step to develop a self

training algorithm system i.e. where no offline training is required and the algorithm

will learn as it flies on the observation image.

1.6: Thesis Outline

A brief overview of the thesis is as follows: Section 2 describes the hybrid approach of

ETMF and HMM. Section 2.1 gives a brief description of ETMF approach; section 2.2

provides a brief overview of HMM training and classification. Section 2.3 explains in

detail the process of combining between ETMF and HMM. Section 2.4 provides the

HMM simulation parameters and also shows the ATR performance results for both

ETMF and HMM individually and the resulting hybrid technique. Section 2.5

summarizes the results obtained.

17
Section 3 is devoted to explore the performance capability of the proposed time

recursive multiple ID hypothesis. Section 3.1 briefly summarizes the approach and

assumptions, Section 3.2 compares the performance between single profile hypothesis

and time recursive multi profile hypothesis. Section 3.3 summarizes the performance

improvement due to time recursive target ID updating approach.

In Section 4 a number of methodologies to develop a “self-training” algorithms

for UWB radar target detection are investigated. The SAR simulation algorithm is

discussed in detail in section 4.1. A brief discussion of eigen analysis on SAR and clutter

suppression capability of SVD are presented in section 4.2. The “Energy-Normalized

SVD”, “Condition-Statistic SVD”, and “Terrain-Filtered SVD” algorithms are discussed

in section 4.3 and comparative detection results are presented in section 4.4 along an

analysis and discussion.

Section 5 presents the conclusion, possible future application and the summary of

this work.

18
2: Robust HRR Radar Target Identification by Hybridization of HMM

and Eigen Template based Matched Filtering

A new hybrid Automatic Target Recognition (ATR) algorithm is developed for

sequential HRR radar signatures. The proposed hybrid algorithm combines ETMF and

HMM techniques to achieve superior HRR-ATR performance. In the proposed hybrid

approach, each HRR test profile is first scored by ETMF which is then followed by

independent HMM scoring. The first ETMF scoring step produces a limited number of

“most likely” models that are target and aspect dependent. These reduced number of

models are then used for improved HMM scoring in the second step. Finally, the

individual scores of ETMF and HMM are combined using Maximal Ratio Combining to

render a classification decision. Classification results are presented for the MSTAR data

set via ROC curves.

2.1: ETMF Approach of training and classification

In ETMF approach of target classification, a new air-to-ground HRR-ATR

algorithm is proposed, where the template features are obtained via Singular Value

Decomposition (SVD) of HRR profiles and the unknown target classification is

performed using normalized Matched Filtering. The SVD operation projects the

information content in a detected HRR profile matrix onto orthogonal basis spaces. This

19
is also known as Karhunen-Loeve Transformation or Principal Component Analysis.

More interestingly, when SVD is applied to a HRR profile matrix, which is range vs.

aspect, it is shown that the range and angle basis spaces are numerically decoupled in the

form of left and right eigen vectors, respectively. This enables us to exploit the decoupled

range information exclusively for the purpose of target recognition.

The Theoretical results presented in [11-12] demonstrated that the range-space

eigen vectors constitute the "optimal" features in the range domain. In addition, SVD

analysis of a large class of MSTAR targets indicates [12] that over 95% of target energy

is accounted for the largest singular value only, further justifying the proposed utilization

of significant range-space eigen-vectors as templates. Basis space decomposition via

SVD is also shown to be useful for suppression of clutter from measured profile data by

eliminating the eigen-vectors corresponding to smaller singular values, which may

represent noise or clutter sub-spaces.

2.1.1: HRR Data Generation and Preprocessing

Most work on Automatic Target Recognition (ATR) has been performed using Synthetic

Aperture Radar (SAR) images. ATR using SAR images performs poorly in case of

moving targets due to blurring caused in the cross-range domain. The HRR-ATR

technology relies on processing high resolution 'Range Profiles', as distinguished from

traditional SAR-ATR that utilizes SAR image data. In HRR based ATR systems there is

a considerable saving in front-end processing in producing HRR profiles which require

only 1-D FFT operation, as opposed to SAR's use of 2-D FFT. The processing factor

becomes significant in case of on-line processing because in order to produce a single

20
SAR image, radar returns must be generated over a relatively large sector of angles. With

HRR profiles, only a relatively small number of angles would be sufficient to perform

ATR. Figure 2.1 shows the process of generating HRR profiles from Complex Phase

History (CPH). As shown, SAR image can be obtained from the HRR profiles by taking

Fourier transform in the angle-domain to produce the cross-range information.

The Range Swath to be imaged is defined a-priori based on Altitude and

depression angle of radar. This makes a fixed sampling window. The two primary HRR

waveforms for SAR systems are the Frequency stepped and Linear Frequency

modulation. The Range resolution (∆R) is determined by the radar RF bandwidth. Thus,

the resultant received signal (Y ( τ j )) in each Range gate would be

N 2
j4 πR i
Y(τ j ) ∝ ∑
i =1
σi e λ
(2.1)

Where σi is the RCS of elemental scatterers in Range gate, Ri is the Range and N is the

number of scatterers in a Range gate.

Complex phase
IFFT (r2+x2)1/2 SVD
history (CPH) Complex Detected Eigen
HRR HRR profiles Templates

Fig. 2.1: Eigen-Template Generation from detected HRR profiles

21
Note that no power transform operation is performed on magnitude HRR as in [12] we

proved that, power transform severely degrade ATR performance if noise is embedded

with Complex Phase History data.

2.1.2: Normalization

The template profiles of all the targets are normalized to have same length (i.e. energy),

while preserving their angular separation and relative variations in scattering returns.

2.1.3: Alignment of HRR Profiles in Range

The HRR profiles of the Segmented Data set provided by AFRL (TRUMPETS) are not

aligned in Range. Hence each Profile of 1-Degree Sector should be aligned with respect

to each other. This alignment is achieved by taking a profile as a reference and shifting

the adjacent profile till maximum correlation was achieved. This procedure is repeated

until all the profiles in a sector have been aligned. Though this procedure of aligning the

HRR profiles is fairly accurate, but it is not fully perfect.

2.1.4: Eigen-analysis of HRR data for training

Singular Value Decomposition (SVD) is a very effective and robust tool for decomposing

any matrix into orthogonal basis spaces. Let Y be an NXM of detected range profiles at

M angular looks containing N range gates each. The SVD operation produces basis

decomposition in the form of three matrices,

• Y ∈ ℜ NXM: Detected HRR Profile Matrix, N = No. of Range profiles and M = No.

of Angular looks

22
M

Y 
→ UΛV =
SVD T
∑λ u v
i =1
i i
T
i (2.2)

where,

• Range-Space (Left) Eigen Vectors : (Use as Features)


NXM
U = EV[YYT ] = [u1 ......un ] ∈ ℜ (2.3)

• Angle-Space(Right) Eigen Vectors : (Discard)

V = EV[Y T Y] = [ v1 ....v m ] ∈ ℜ NXM (2.4)

• Singular Values :

Λ = Diagonal[λ11 ......λ MM ] ∈ ℜ NXM (2.5)

• Range and Angle sub-spaces are decoupled via SVD.

Eigen Value Distribution


6

e
d
ut 3
i
n
g
a
M
2

0
0 5 10 15 20 25 30 35
Number of Eigen Values

Fig. 2.2: Distribution of Singular values for MSTAR target T72, 1000 sector

23
Where, EV[.] denotes the operation “Eigen-Vectors of”. For Range vs. Angle

HRR data, the left eigen vectors (U) span the orthogonal basis space in the range domain

while the right eigen vectors (V) span the angle space. The middle matrix Λ is diagonal

containing M (N>M is assumed here) singular values in decreasing order,

λ11 ≥ λ 22 ... ≥ λ MM , where λ ii denotes the weights associated with i-th eigenvector. Larger

Singular values imply significant contribution of that particular eigen-vector in forming

the target signal. Hence these are denoted as “signal subspace” eigenvectors whereas

those corresponding to the smaller singular values are denoted as “noise or clutter

subspace”. Figure 2.2 displays the distribution of singular values for a typical MSTAR

targets in a particular degree range. In that case, it is seen that only the highest singular

value ( λ11 ) makes up more than 96% of the total energy of the distribution. Interestingly

the range space in U and the angle space eigenvectors in V appears in decoupled form

after the SVD operation is applied to Y as shown in equation (2.3). It can be concluded,

the HRR profile matrices are close to rank one, which implies that u 1 , the left-

eigenvector corresponding to the largest (or dominant) singular value ( λ1 ) ought to

contain the essential range information of the underlying target. Hence, here it is

proposed to use the dominant range-space (left) eigenvector as the feature template.

24
2.1.5: Unknown Target Classification

Template

||<------ --------------------------------Overlap Region ----------->|


Test Profile (shift = -8)

|<-----------------------------Overlap Region ------------------------->|


Test Profile (shift = 0)

|<--------------------------Overlap Region-------------------------->|
Test Profile (shift =+ 8)

Fig. 2.3: Implementation of the Correlation Classifier

The unknown target classification is performed using Normalized Matched Filtering.

Given observed (or, test) range profile(s) of an unknown target, the ultimate

objective of classification is to determine which target class it belongs to. This is

accomplished by comparing the observed profile with all the available templates, which

are assumed to have been formed beforehand using training data set. The decision

determines the target type for which the correlation between its template ( mi ) and the

observation (a) profile is maximized among all template choices. However as the

observation profile a and all the template may not be exactly aligned, the correlations

have to be calculated with various lag values and the maximum correlation among all

lags for each target type has to be determined. For each target, there are usually a large

25
number of templates at different aspects. In our simulations with MSTAR database,

correlation lag values up to ±8 were used.

The maximum correlation value among all templates within ± D° of target aspect

(assumed known or estimated by an MTI tracker) for each target is determined. This

process is repeated for all target classes, with each class being assigned its maximum

correlation out of all lags for aspect angles within ± D° . Finally, the target class having

the maximum correlation value among all classes is termed the matched target class. In

our simulations, correlation lag values up to ±5o of the true aspect was used because it is

assumed that the MTI tracker (running in conjunction with HRR-mode radar) would

provide a reasonably good aspect estimate.

2.1.6: Modified Normalization and Centroid Alignment

To improve the performance of the ATR algorithm it is important to include that portion

of the Observation and Template profiles which contains significant portion of the target

signature information. Therefore, if the Observation and Template profiles are not pre-

aligned it is important that they be aligned prior to using them in the classifier. In this

work the Centroid of a range profile was used as the reference in aligning the

Observation and the Template profiles.

As described in the previous section, the Matched Filter Classifier assumes that

both the Observation and the Template profiles are normalized to have equal lengths.

However, while correlating the template and test profiles to find the best match, one of

the profile vectors is shifted to the left and right of the Centroid to obtain the maximum

correlation. When the observation profile is shifted over the Template profile, the region

26
of overlap between the two would change with each shift. However, the norms of the

overlapped regions of the observation and template may also change with each particular

shift. Hence, using a stored template profile originally normalized over its entire length

will not be appropriate if used as is. In order to satisfy MF’s requirement that both

template and observation have identical lengths, it is important that only the overlapping

parts of both the profiles is normalized prior to correlating the vectors, as described next.

Let the test and template profiles be represented by narrow (shaded) and wide

rectangles, respectively, as depicted in Figure 2.4. The lengths are shown different

intentionally, as the test and template could be of different lengths. Different heights are

used primarily to differentiate between the test and template. It has no other implication.

A
OBSERVATION PROFILE Template Profile

Fig. 2.4: Observation and Template Profiles are shown in shaded and blank boxes of different lengths

Next for better understanding, the correlation process with overlap normalization

is described in detail. The test profile was shifted over the template and correlated. In the

next figure, it is assumed that the shift is –8 with respect to the centroid. Clearly, the

entire lengths of neither test nor the template are overlapping. Hence, it doesn’t make any

sense to normalize over entire lengths of the template or test, because the correlation is

27
occurring only over the overlapped (shown in stripe) region. It will be more appropriate

to ensure that the norms within the overlap region of the vectors are kept the same.

Hence, we re-normalize both vectors only over the overlapped parts (in stripe) before we

perform correlation.

Overlap region to be normalized

Fig. 2.5: Shift = -8 of Centroid

Next, the case when both test and templates are aligned on the Centroid is

depicted. In this case, the entire length of the template is overlapping some middle

portion of the test. Hence, once again, we re-normalize only within the overlapping

regions to ensure that both vectors have same lengths. It may be noticed that the length of

the overlapped portion (in stripe) of the vectors is longer than the previous case.

28
Overlap region to be normalized

OVERLAP REGION

Fig. 2.6: Shift = 0 aligned of Centroid

Next, the +8 shift case from centroid is shown. Again, the overlapped regions

have changed for both. Again, only the striped regions are normalized for both vectors

before correlating.

Overlap region to be normalized

Fig. 2.7: Shift = +8 of the Centroid

29
2.2: HMM approach of training and classification

Some recent work has shown encouraging promise for the use of HMM in HRR target

recognition [8]. The airborne radar transmits microwave pulses at constant depression

angle. Each pulse is reflected from the target and gets back to the radar receiver.

Preprocessing is performed on the scattered waveforms and the output achieved is

sequence of range scattered pulses which is termed as High Range Resolution (HRR)

profiles. The HRR signatures characterize the target at a specific airborne sensor

orientation. In the MSTAR data collection studies, it is assumed that the depression angle

of airborne radar with respect to the target is constant and the target sensor orientation is

modeled as the change of azimuthal orientations. Though there is significant variability in

HRR signatures at different orientations, the scattered range field can be assumed to be

stationary over small angular sectors. Each such angular region is termed as a “state”. As

the target orientations are unknown in addition to target identity, theses information can

be assumed as hidden and HMM can be used to model and characterize the sequence of

scattered waveforms. Figure 2.8 shows a framework for HRR-based target recognition

using HMM.

HRR profiles
(Training data) Training Model
(HMM)

Code Book

HRR profiles Testing Target


(Test data) Classification

Fig. 2.8. Simplified Block diagram of target recognition using HMM

30
This subsection describes an introduction to Hidden Markov Models and algorithms for

evaluation, HMM training using Forward-Backward algorithm and HMM classification

using maximum likelihood viterbi searching.

2.2.1: Discrete Hidden Markov Model Introduction

It is often convenient to think of HMM as a collection of interconnected states. Like the

classical Markov model, we use a transition probability to provide the probability of a

transition from one state to another. Unlike a classical Markov model, a Hidden Markov

Model introduces an output probability density function (Pdf) to define the conditional

probability that a symbol is generated from a finite set of symbols, given that we are in a

particular state. From the probabilistic perspective, HMM’s characterize a stochastic

process with an underlying Markovian finite-state structure that may only be observed

indirectly (hence the “hidden” nature of the model). At any given time, it is unknown to

an outside observer what state the process is in, but it can be observed through the

sequence of symbols emitted from the states.

We will limit our consideration to the first-order Hidden Markov Model, where

state dependencies are on the immediate predecessor only. Another assumption made in

this discussion is output-independence, which means the output symbol probability

depends only on the current state at this observation frame and is conditionally

independent from the previous state and the past symbols emission. This second

assumption may degrade the experimental realism of HMM’s, but it reduces the number

of parameters required by the model and allows the use of efficient evaluation and

training algorithms in the synthesis and learning phase.

31
2.2.1.1: Elements of HMM

Fig. 2.9: Two states Hidden Markov Model with two output symbols, V1 and V2

Figure 2.9 shows a Hidden Markov Model with two output symbols, V1 and V2. This

simple model is used to explain the elements of HMM. The parameters of the HMM that

can generate the output symbols V1 and V2 are shown in Equation (2.6) and (2.7).

N = 2, M = 2 (2.6)

1  0.6 0.4   0.8 0.2


π =  , A =   , B=  (2.7)
0 0.2 0.8   0.3 0.7 

Following are the definitions for each parameter:

N: the number of states in the model. We will denote the individual state as

S = {s1 ,s 2 ,s 3 ,...s N } , and the state at time t as qt.

32
M: the number of distinct observable symbols per state or the size of the codebook.

We denote the individual symbols as V = {v1 , v 2 , v3 ,..., v M } , and the observation

symbol at time t as Ot.

A: NxN matrix representing the state transition probabilities, i.e. the probability to

make a transition from state si to state sj:

A = {a ij} where a ij = prob ( q t = s j | q t -1 = s i ) 1 ≤ i, j ≤ N (2.8)

And

N
a ij ≥ 0, ∑a
j=1
ij = 1 ,1 ≤ a ij ≤ N (2.9)

B: NxM matrix which specifies the observation symbol probability distribution in the

state sj:

B = {b j (k)} where b j (k) = prob(v k = t | q t = s j ), 1 ≤ j ≤ N, 1 ≤ k ≤ M (2.10)

And

M
b j (k) ≥ 0, ∑ b (k) = 1 ,
k =1
j 1≤ j≤ N (2.11)

π: N-element vector indicating the initial state probability distribution:

π = {πi } , where πi = prob(q1 = si ), 1 ≤ i ≤ N (2.12)

The complete parameter set λ of HMM requires the specification of two model

parameters (N and M), the specification of the observation symbols, and the specification

33
of the three probability matrices A, B and π. The compact but convenient notation is used

to represent the parameter set of HMM model.

λ = ( A, B , π ) (2.13)

2.2.1.2: Three Basic Problems for HMM

The use of HMM models in real-world applications requires the solution of the following

three problems related to the set of their parameter descriptors:

Probability Evaluation: Given a model and a sequence of observations, how do we

efficiently evaluate the probability that the model generated the observations?

Optimal State Sequence: Given a model and a sequence of observations, how do we

determine an optimal state sequence in the model that generated the observations?

Training: Given a model and a set of observations, how do we adjust the model

parameters of λ to maximize the probability of generating the observations?

In the following sections, the solutions proposed for each of these three basic problems

are reported.

2.2.1.3: Model Evaluation

The evaluation problem can be stated as: given the observation sequence

O = O 1O 2 ...O T , and a HMM model λ = ( A, B, π ) , compute P (O λ ) , the probability

that the observed sequence is produced by the model. The most straightforward way to

compute this is to enumerate all possible paths (state sequences) of length T that generate

observation sequence O i.e. sum of all their probabilities.

34
P ( O | λ ) = ∑ P ( Q | λ ) P ( O | Q, λ ) (2.14)
all Q

Where Q is the state sequence: Q = q1q 2 L q T and q1 is the initial state. The first factor in

Equation (2.14) can be re-written by applying the two Markov assumptions:

T
P ( Q | λ ) = ∏ P(q t | q t −1 ) = πq1 a q1q 2 a q 2q 3 L a q T−1qT (2.15)
t =1

The second factor in Equation (2.14) can be re-written by applying the output-

independence assumption:

P ( O | Q, λ ) = bq1 ( O1 ) ⋅ b q 2 ( O 2 ) L bq T ( OT ) (2.16)

Substituting Equation (2.16) and (2.15) into (2.14), we have:

P ( O | λ ) = ∑ P ( Q | λ )P ( O | Q, λ ) (2.17)
all Q

= ∑
q1 ,q 2 ,...q T
πq1 b q1 (O1 ) a q1q2 b q2 (O 2 )...a qT−1qT b qT (O T ) (2.18)

From Equation (2.18) we can directly calculate the P (O λ ) from the HMM parameters,

but the computation is unfeasible even for small values of N and T because the

computational complexity increases exponentially with T. Fortunately, because of our

assumptions, there is a more efficient algorithm called Forward-Backward procedure

to compute P (O λ ) recursively.

35
2.2.1.4: Optimal State Sequence

Given a model λ and a sequence of observations O = O1O2 ...OT , one problem that needs

to be addressed is the estimate of the best state sequence Q = q1q 2 ...q T (or the most likely

state path) corresponding to the given observation sequence. A dynamic programming

method called Viterbi algorithm [60] is used to choose the optimal state sequence, i.e., to

maximize P (Q O , λ ) , which is equivalent to maximizing P (Q, O λ ) .

2.2.1.5: Parameter Estimation

The training problem involves adjusting the model parameters in order to maximize the

probability of the training observation sequences being produced by the model. The

iterative procedure called the Baum-Welch algorithm is used to choose the maximum

likelihood model parameter λ such that its likelihood function P(O | λ ) is locally

maximized.

2.2.2: HMM Operation steps

2.2.2.1: Framing

Here, each HRR profile is blocked into frames of N samples, with adjacent frames being

separated by M (M<N). The first frame consists of the first N samples. The second frame

begins Mframe samples after the first frame, and overlaps it by N- M samples. Similarly,

the third frame begins 2M samples after the first frame (or M samples after the second

frame) and overlaps it by N- 2M samples. This process continues until all the profile is

accounted for within one or more frames. As each HRR profile consists of a small

36
number of samples, framing is pretty useful to duplicate the information contained in a

single profile and learning the HMM. Framing is done on HRR profiles used both for

training and testing.

2.2.2.2: Clustering

The next specific step is to make target and aspect specific Vector Quantization (VQ)

codebook. The importance of codebook is to cluster the data in feature space. A

commonly used LBG algorithm [61] for clustering a set of L frame vectors into a set of C

codebook vectors was formally implemented by the following recursive algorithm:

1) Design a 1-vector codebook; this is the centroid of the entire set of training vectors

(hence, no iteration is required here).

2) Double the size of the codebook by splitting each current codebook yn according to the

rule:

y +n = y n (1 + ε) (2.19)

y −n = y n (1 − ε) (2.20)

Where n varies from 1 to the current size of the codebook, and ε is a splitting parameter

or learning parameter (we choose ε = 0.01).

3) Nearest-Neighbor Search: for each training vector, find the codeword in the current

codebook that is closest (in terms of similarity measurement, here we have taken as

Euclidean distance as similarity measure), and assigns that vector to the corresponding

cell (associated with the closest codeword).

4) Centroid Update: update the codeword in each cell using the centroid of the training

vectors assigned to that cell.

37
Iteration 1: repeat steps 3 and 4 until the average distance falls below a preset threshold.

Iteration 2: repeat steps 2, 3 and 4 until a codebook size of C is designed.

Figure 2.10 shows, in a flow diagram, the detailed steps of the LBG algorithm.

“Cluster vectors” is the nearest-neighbor search procedure that assigns each training

vector to a cluster associated with the closest codeword. “Find centroids” is the centroid

update procedure. “Compute D (distortion)” sums the distances of all training vectors in

the nearest-neighbor search so as to determine whether the procedure has converged.

Find
centroid
Yes No
m<M Stop

Split each
centroid

m = 2*m

Cluster
vectors

Find
centroids

Compute D
(distortion)

No D'−D Yes
D’ = D <ε
D

Fig. 2.10: Flow diagram for LBG clustering algorithm

38
Intuitively, the LBG algorithm designs a C-vector codebook in stages. It starts first by

designing a 1-vector codebook, then uses a splitting technique on the codewords to

initialize the search for a 2-vector codebook, and continues the splitting process until the

desired C-vector codebook is obtained. Clustering is done on HRR frames used for both

training and testing.

Euclidean distance is computed for each HRR frame with all codebook centers.

The codebook centers, which are the nearest of the frames, emerge as winner and the

indices of those winner centers are sent to HMM optimum parameter estimation process

during training mode and maximum likelihood computation process during testing mode.

2.2.2.3: HMM Training

HMM training procedure consists of the following 2 operational steps:

HMM model parameter estimation i.e. estimating parameter λ such that its

corresponding likelihood function P(O | λ ) is locally maximized, where

O = O1O 2 L OT and λ = ( A, B, π ) i.e. convenient notation used to represent HMM

parametric model.

HMM Optimal State Sequence estimation i.e. given a model λ and a sequence of

observations O = O1O 2 L OT , one problem that needs to be addressed is the estimate

of the best state sequence Q = q1q 2 L q T (or the most likely state path) corresponding

to the given observation sequence.

The two training steps are described in the following section.

39
2.2.2.3.1: Model Optimum Parameter estimation:

In the proposed HMM training, one HMM model is designed for each target of interest.

Each target HMM model is subdivided into its corresponding aspects. A target model is

partitioned into N distinct states, denoted by the set S={ s1 , s2 ,...,s N }, where the total

number of states N can be determined from azimuthal partitioning. A HMM model can

be denoted as λ = ( A, B, π) . The state transition probabilities are denoted as, A = {aij},

where, aij = P (st+1 = p| st = q), the probability of being in state p at time t+1, given that at

time t the state is at q. The state output probability B={bij}, bij = P(uk at t| st = q), where

U={ u1 , u 2 ,..., u N } is the observation likelihood sequence from all N states at time t. State

initial probability π = { πi }, πi = P(s1 = p), the probability of being in state p at the

beginning of the experiment. By definition, A and π are initialized as:

Number of occurences of {O1 ∈i}


πˆ i = , 1≤ i ≤ N (2.21)
Total Number of occurences of O1

Number of occurances of {O t ∈ i and Ot +1 ∈ j } for all t


â ij = , 1 ≤ i ≤ N and 1 ≤ j ≤ N (2.22)
Number of occurances of {O t ∈i}for all t

In the ATR application, these initializations are dependent on the angular extent of each

state and the difference between target orientations over successive measurements. If θi

represents angular extent of state i and ∂θ the change of target orientations over 2

consecutive measurements, the initial state transition probabilities can be defined as:

ai,i-1 = ai,i+1 = ∂θ / (2 θi ) and ai,i = ( θi - ∂θ )/ θi (2.23)

where i is the state index.

40
As we have no prior information about the targets, the initial state probability for each

target is uniformly distributed:

θ j,i
πj = N
(2.24)
∑ θ j,i
i =1

where the subscripts i and j are the state and target indices respectively.

If at a particular time instant t, Ot is the trained HRR profile, our primary task in the

training process is to optimize it’s corresponding target HMM model λ = ( A, B, π) , so that

the likelihood P (Ot| λ ) is maximized. The optimal A, B and π matrices for each target

can be estimated using the well-known recursive method called Baum-Welch re-

estimation algorithm [62]. This is a Bayesian approach where an initial HMM model is

improved upon by using recursive iterative formulas.

The primary objective of Baum-Welch estimation algorithm is to maximize

P(O| λ ) by adjusting the parameters of λ = ( A, B, π) . Consequently, the Baum-Welch

optimization is considered as Expectation Maximization (EM) criterion. The following

recursive equations in the Baum-Welch estimations describe a complete updating cycle:

First, state occupation probability at time t given the observation sequence and

HMM model is defined as:

γ t (i) = P(q t = s i | O , λ ) i.e. the probability of being in state i at time t, given the target

observation id O = O1, O2…OT and the model λ . qt is the state observed at t-th instant

and si is the i-th state of the target (i = 1,2,...,N).

Using Bayes rule,

P(q t = si , O | λ ) α t (i)βt (i)


γ t (i) = = (2.25)
P(O | λ ) P(O | λ )

41
Where, α t (i) is the forward variable and can be defined as:

α t (i) = P (O1, O2… Ot, qt = si| λ ) i.e. is the probability of partial observation sequence up

to time t and the state si at time t, given the HMM model λ . (2.26)

In a similar manner the backward variable β t (i) can be defined as:

β t (i) = P (Ot+1,Ot+2…OT| qt = si, λ ) i.e. is the probability of observation sequence from

time t+1 to T, given the state si at time t and the model λ . (2.27)

ξ t (i, j) = P(q t = s i , q t +1 = s j | O , λ ) i.e. the probability of being in state i at time t and

making a transition to state j at time t+1, given the observation target id O = O1,O2,…,OT

and the HMM model λ . Again, applying Bayes rule,

P(q t = s i , q t +1 = s j , O | λ )
ξ t (i, j) = (2.28)
P(O | λ )

But O can be defined as O = O1, O2… OT. Hence the Numerator can be written as,

Numerator = P(q t = s i , O1 , O 2 ..., O t , O t +1 , O t +2 ..., O T , q t+1 = s j | λ )

= P(q t = s i , O1 , O 2 ..., O t | λ ) P(O t +1 , O t +2 ..., O T , q t +1 = s j | λ )

= Term1×Term2 (2.29)

Where, Term1= α t (i) [from equation 2.26] (2.30)

From Term2 it is evident that we have to choose the observation id Ot+1 from state sj.

Hence we have,

P(O t +1 , O t + 2 ..., OT , q t +1 = s j | λ ) = P(q t +1 = s j , O t +1 | λ )P(O t +1,..., O T | q t +1 = s j , λ )

= aijbj(Ot+1) β t +1 ( j) (2.31)

Combining (2.29), (2.30) and (2.31)

42
α t (i)a ijb j (O t +1 )β t +1 ( j)
ξt (i, j) = (2.32)
P(O | λ )

Fig. 2.11: Illustration of the sequence of operation required for the computation of the joint event that the

system is in state Si at time t and state Sj at time t+1(Reproduced from [62])

Figure 2.11 shows the computation of ξt (i, j) based on the forward and backward

variables.

From equation 2.32, α t (i) accounts for O1…OT, aij accounts for transition

probability from state i to state j, b j (O t +1 ) picks up the symbol O t +1 from state j and

β t +1 ( j) accounts for the remaining observation sequence Ot+2,…OT.

The expected number of transitions from state si during the time instant t is:

T −1
∑ γ t (i) (2.33)
t =1

43
The expected number of transitions from state si to state sj during the time instant t is:

T −1
∑ ξt (i, j) (2.34)
t =1

The iterative formula for π , A and B may be computed as:

πˆ i = γ t (i) ,1 ≤ i ≤ N (2.35)

T −1
∑ ξ t (i, j)
â ij = t =1
T −1
,1 ≤ i ≤ N (2.36)
∑ γ t (i)
t =1

T
∑ γ t ( j)
t =1

Ot = k
b̂ j (k) = T
,1 ≤ i ≤ N (2.37)
∑ γ t ( j)
t =1

Fig. 2.12 Baum-Welch learning algorithm

Figure 2.12 summarizes the Baum-Welch algorithm.

44
The EM algorithm itirates by using the new HMM parameters λ computed in (2.35)-

(2.37) to recompute the parametres defined in (2.26)-(2.32). These iterations go on until

λ converges. The HMM parameters can be iteratively computed as long as P (O| λ t +1 ) >

P (O| λ t ). When the inequality becomes equal or if the right hand side becomes less than

the left, we can conclude that λ is converged and the estimated λ̂ maximizes P(O| λ )

i.e. upon convergence of λ , we can conclude that the likelihood is maximum of a given

set of training HRR waveforms O = {O1O2...OT}, as applied to it’s associated HMM

model.

2.2.2.3.2: Optimum state sequence estimation:

Given a model λ and a sequence of observations O = O1O 2 L OT , the problem is to

estimate the best state sequence Q = q1q 2 L q T corresponding to the given observations.

A famous dynamic programming method called Viterbi algorithm [60] was used to

choose the optimal state sequence, i.e., to maximize P (Q O, λ ) , which is equivalent to

maximizing P (Q, O λ ) . We define the maximum probability along a single best path at

time t, which accounts for the first t observations and ends in state sj given the HMM

model λ, as

δt (i) = max P[q1q 2 L q t = s i , O1O 2 L O t | λ ] (2.38)


q1q 2 Lq t −1

We also define an array ψ t ( j) to keep track of the argument, which maximizes δt −1 (i)a ij

for each t and j. The Viterbi algorithm obtains a solution recursively as follows:

45
Initialization:

δ1 (i) = P(O1 , q1 | λ ) = π1bi (O1 ), where 1 ≤ i ≤ N (2.39)

ψ1 (i) = 0 (2.40)

Recursion:

δ t ( j) = max P[δ t −1 (i)a ij ]b j (O t ), where 2 ≤ t ≤ T and 1 ≤ j ≤ N (2.41)


1≤ i ≤ N

ψ t ( j) = arg max[δt −1 (i)a ij ], where 2 ≤ t ≤ T and 1 ≤ j ≤ N (2.42)


1≤i ≤ N

Termination:

P* = max[δT (i)] (2.43)


1≤ i ≤ N

q*T = arg max[δT (i)] (2.44)


1≤i ≤ N

Path Backtracking:

q ∗t = ψ t +1 (q *t +1 ), t = T − 1, T − 2, L , 1 (2.45)

From the above induction, we can see that the Viterbi searching is a time synchronous

searching algorithm. Before going to time t+1, the process at time t is completed. For

time t, each state is updated by the best score from the states at time t-1. After finishing

the forward state estimation, Viterbi algorithm backtracks to optimize the states for the

given HRR sequences.

46
2.2.2.4: HMM Classification

λ1 HMM Model 1

Likelihood
computation

λ2
Observation
sequence Collection of
Likelihood
Computation likelihood’s
Framing Vector (P(O|λ)) indicating
Quantization the similarity
Test between the test Target index
HRR HMM model K HRR profiles and corresponding to
Feature the (PHMM)max
vectors training models
λK

Likelihood
Computation

Fig. 2.13: Block Diagram of a Designed HMM recognizer

In the training procedure, one HMM model is optimized for each target of interest. In the

testing process, a given set of sequential HRR profiles under test are submitted to all

HMM models, and the data is associated with that target for which the respective HMM

yields the maximum likelihood, i.e. for each test HRR profile at time t, we computed

P(Ot| λ k ), where k, the index, varies from 1 to the total number of HMM models. The test

profile will be assigned to that target class whose corresponding HMM models produces

the largest likelihood. The classification procedure can be understood from the block

diagram 2.13.

47
The HMM classification problem can be mathematically stated as: given the

observation sequence O = O1O 2 L OT , and HMM models λ κ = ( A k , B k , π κ ) , where k

varies from 1 to total number of HMM models (which is same as number of trained

targets), compute likelihood P ( O t λ k ) for each HMM model and classify the unknown

observation profile at time t to that known target class, whose likelihood is the largest.

The efficient algorithm called forward-backward procedure will calculate P ( O t λ k )

recursively for each HMM model and for profiles observed at each time instant. The

Forward and Backward recursion procedures are individually described below:

2.2.2.4.1: The Forward Procedure

Let’s define the forward variable as

α t (i) = P(O1O 2 L O t , q t = si | λ ) (2.46)

i.e., the probability of the partial observation sequence, O1O 2 L O t , (until time t) and

state si at time t, given the model λ.

The forward and backward recursion procedures are depicted in the next page.

48
Fig. 2.14: Illustration of the sequence of operations required for the computation of the (a) forward variable

and (b) backward variable (Reproduced from [62])

49
We can compute α t (i) inductively, as follows:

Initialization: The initial forward variable α1 (i) is the joint probability of the state i,

time t = 1 and the initial observable symbol O1 by the given model λ.

α1 (i) = πi bi (O1 ) where 1 ≤ i ≤ N (2.47)

Induction: From Figure 2.14 a)

It can be seen that state j can be reached at time t+1 from the N possible states i,

1 ≤ i ≤ N at time t, and the induction is:

α t +1 ( j) = P ( O1O2 L O t +1 , q t +1 = s j | λ ) (2.48)

N 
=  ∑ α t (i)a ij  b j (O t +1 ), where 1 ≤ t ≤ T − 1, 1 ≤ i, j ≤ N (2.49)
 i =1 

Termination: The probability of the complete observations sequence is the sum of all

N final forward variables:

N N
P ( O | λ ) = ∑ P ( O1O2 L O T , q T = si | λ ) = ∑ P ( O, q T = si | λ ) (2.50)
i =1 i =1

N
= ∑ αT (i) (2.51)
i =1

The forward probability calculation is based on the lattice or trellis structure showed in

Figure 2.15. Since there are only N states, all possible state sequence will remerge into

these N states irrespective of the length of the observation sequence.

50
S S S S

α(y1t −1 , j)
β(y Tt+ 2 | j)
j j j j
α(y1t ,i)
β(y Tt+1 | i)

a(i | j) i a( j | i)

2 2 2 2
b(y t +1 | j)
b(y t | i)

1 1
1 1

t−2 t −1 t t +1 t+2
yt−2 y t −1 y t y t +1 y t+ 2

Fig. 2.15: State lattice used to derive the forward/backward recursion

In Figure 2.15 above, the time instants are denoted as t and the observation sequence at a

particular time instant t is denoted as yt.

51
2.2.2.4.2: The Backward Procedure

Figure 2.14 (b) illustrates the backward procedure. We define a backward variable:

βt (i) = P(O t +1O t + 2 L OT ,q t = si | λ ) (2.52)

i.e., the probability of the partial observation sequence from t+1 to the end, given state si

at time t and the model λ. βt (i) was computed using the following recursive steps:

Initialization:

βT (i) = 1, where 1 ≤ i ≤ N (2.53)

Induction:

βt (i) = P(O t +1O t + 2 L OT ,q t = si | λ ) (2.54)

N
= ∑ a ijb j (O t +1 )β t +1 ( j), where t = T − 1, T − 2, L ,1, 1 ≤ i ≤ N (2.55)
j=1

Termination: The probability of the complete observations sequence is the sum of all

N final backward variables:

N N
P ( O | λ ) = ∑ P ( O T O T −1 L O1 , q1 = s i | λ ) = ∑ P ( O , q1 = s i | λ ) (2.56)
i =1 i =1

N
= ∑ β1 (i) (2.57)
i =1

The introduction of forward and backward variables provides us with an efficient

algorithm to compute the probability of the observation sequence given the model. The

classification algorithm will assign the observation sequence O to that target class whose

corresponding HMM model produces the highest P ( O | λ ) .

52
2.3: Approach of Combination between ETMF and HMM

2.3.1: Motivation for Hybrid approach

As, single ETMF or single HMM failed to provide satisfactory target recognition

performance (shown in later section), concentration was directed to hybrid those single

approaches and apply in target recognition environment.

The motivation came from the following two reasons:

Each ATR algorithm has inherent performance limitation. ETMF or HMM cannot

exceed it’s maximum capability and perform better than that. Hence single algorithm

ATR classification is not always the correct way to get above satisfactory level

output.

The major advantage of the hybrid technique is: if one algorithm fails to get target ID

correctly, it is quite possible that another algorithm might work. In the hybrid

algorithm as two independent classifications algorithms are running parallely, the

probability of having mistake for both techniques at the same time/same place will

reduce drastically. The challenge is to design a hybrid system that can recognize the

target correctly even if only one algorithm between the two makes a correct decision.

In other words, the main question to be answered during the design phase is how to

harness capabilities of multiple ATR algorithms to achieve further performance

improvement.

53
2.3.2: Proposed Solution

There were hybrid algorithms in the literature [63] applied in speech recognition, where

two algorithms were allowed to run parallely in order to achieve a combined

classification decision. In those approaches, two classification algorithms run parallely on

the same set of data and produce discriminant scores independently. Next, the

discriminant scores produced by those approaches were combined with proper weights,

which are derived from the data set, and classification was performed on the combined

set of discriminant results. The primary advantage of the hybrid algorithm is improved

classification performance can be achieved compared to each algorithm alone, but the

disadvantage of this technique is improved classification is extracted at the expense of

computational efficiency. As ATR is a real time process, computational efficiency to

classify unknown target is a major area of concern. Considering the computational need

and performance improvement, in this paper, the hybrid algorithm presented is modified

version of what being already seen in the literature. In this proposed hybrid algorithm the

two classification algorithms are not allowed to run parallely, in fact the operation of two

ATR algorithms are rather serial and the operation of the second algorithm is much

dependent on the successful act of the first. The major steps of the proposed algorithm are

described below:

In the first step, one ATR algorithm is allowed to choose several Most Likely training

models, which can produce correct classification. Each most likely models are

assigned discriminant scores. The discriminant scores attached with each training

model indicate how closely that model matches with the unknown observation

sequence.

54
Hence the task of first ATR algorithm can be diagonized as :

Selection of a smaller subset of most likely solutions.

Assigning a discriminant score with each one at the smaller subset indicative of

how closely the models match with the unknown observation sequence.

Next, smaller subset model information is passed to the other ATR algorithm.

The task of the second ATR algorithm is to perform classification of the observation

waveforms with only those smaller subset of models and also assign discriminant

scores with each member of the smaller subset group indicating how closely the

match is between the known trained model and the unknown observation sequence.

Hence for a single observation sequence we have two sets of discriminants scores

obtained from the two sets of ATR algorithms.

Next, the two sets of discriminant scores are combined with weights estimated from

the data and accuracy of one ATR algorithm over the other and finally the unknown

observation sequence is decided as that of the known trained target whose model

produces the highest combined discriminant score.

In unknown target scenario, the highest combined discriminant score achieved for the

given profile is first compared with a predetermined threshold. If the discriminant

score exceeds the threshold a target decision is made otherwise it is concluded as

“unknown”.

In the proposed hybrid approach, ETMF and HMM techniques perform the functionality

as first and second ATR algorithms respectively. The necessity and usefulness of

modified hybridization in our case is described in detail in the next section.

55
2.3.2.1: Necessity of developing modified hybridization in ETMF ATR

In the ETMF testing phase, each HRR profile is tested with 3-aspect angle templates

( ± 1 o aspects) per target producing 3K discriminant values, where K is the number of

trained target templates. Here we are assuming that MTI radar is also in operation and

that is in synchronous with HRR radar. As MTI radar is pretty accurate in calculating the

aspect information of the unknown observation sequence, we can limit ourselves in ± 1o

model searching for each target, when an unknown observation sequence comes in. The

ETMF algorithm chooses the target corresponding to the highest discriminant correlation

value. However, it will take wrong decisions in some cases leading to False Alarm or

Missed target. One effective way to reduce the probability of wrong decision is to couple

the ETMF technique with some other independent ATR technique(s) and then make the

final target classification decision based on a combined result. The inherent assumption

here is that the probability of failure in classifying a target by both ATR techniques at the

same time would be very low if the target classification methods are independent of each

other. Since HMM is now well established in speech and pattern recognition, the ETMF

results are combined with that of HMM and the classification decision is taken based on

the combined performance. But here is the problem, incorporating HMM to classify

unknown targets has a huge drawback. As during testing, HMM has to go through

forward-backward recursive procedure to find the likelihood, it is time consuming. If the

number of target templates are large, it goes beyond real time. Hence the tricky issue was

how to implement HMM in real time ATR environment.

That is the reason why we modified the hybrid approach and instead proposed a

new method to hybridize two independent classification algorithms. ETMF algorithm

56
selects a small number of most likely models that are most similar to the unknown

observation sequence. In the next stage, HMM likelihood computation is limited to those

numbers of models and it computes discriminant scores for those models only. By doing

this, a lot of computation time was saved as HMM need not perform matching with all

it’s models between ± 1o . Here it is assumed that ETMF is accurate enough to pass those

models where the actual solution is embedded to HMM. This assumption is based on

experimental observation applied on training profiles. It was observed in case of ETMF

that though sometimes it misses the correct target and assigns highest discriminant score

to the wrong one, but the correct target was always among the top few output

discriminants. Hence we decided to use only a subset of the higher discriminants values

for HMM to score. This method not only saves computational efficiency and implements

the hybrid system in real time but also provides better recognition rate compared to the

hybrid technique used in this same environment.

ETMF Correlation Probability


Conversion P

Model information Combined


Score
HRR
Profile Clustering HMM Probability (P)
Indices Target index

Fig. 2.16: Data flow in the proposed hybrid algorithm

The block diagram 2.16 depicts the proposed combining procedure between ETMF and

HMM.

57
The detailed description about number of subset model selection and weight calculations

to combine ETMF and HMM output discriminant scores are provided in the later section.

2.3.2.2: Number of Subset Model selection and Weight calculation

2.3.2.2.1: Subset Model Selection

The proposed hybridization technique is described in detail in the last two sections. In a

recent hybridization technique in literature, all the HMM models are used in scoring [63].

In our proposed hybrid technique, only few number of HMM models that can most likely

generate the correct solution are used in discriminant scoring. The specific models that

HMM will use in scoring are to be determined by the ETMF algorithm. As HMM has

models for each target and aspect (same as ETMF), only those HMM models are used to

score the observed HRR profiles that give high discriminant scores in ETMF testing. It is

important to emphasize that the exact number of HMM models to be used for scoring is

determined from the training HRR profile data. To find out the optimum models that

HMM will use in classification, the training data was split into training and testing sets.

The “testing set” was formed by randomly selecting 25% of the HRR profiles from each

target in the training database. The remaining 75% of HRR profiles from each target are

regarded as training set and are used to generate the Eigen-templates and HMM models.

At each aspect the “test” set was then used to calculate the HMM classification

performance by varying number of HMM models to be used in scoring. The number of

HMM model goes accordingly to the maximum correlation score assigned by ETMF on

58
each model. In other words, at every aspect, classification for the same aspect test

waveforms are performed using HMM models from 1 to 3K (K is the number of targets

in the database) successively according to the ETMF discriminant scores. The target-

aspect model that is assigned highest similarity by ETMF is used first on HMM

classification. Then the same aspect all test data is classified again by the top two

discriminant target aspect model. This process will go on until all 3K HMM models are

used in classifications. In order the system not to be biased for the specific test samples,

ten different Monte Carlo simulations are performed by taking ten random collections of

“testing set”. The % recognition rate shown in the Figure 5 is the average of all ten HMM

recognition results. The number of HMM models those gives the highest average

recognition rate in that aspect is saved and is used during real testing. The same process

is repeated for all aspects from 10 to 3600. Hence HMM model selection is an aspect

dependent process. Figure 2.17 shows a case of optimal model selection for HMM

classification.

59
Fig. 2.17: In 100 aspect case, this plot shows the HMM recognition rate with number of HMM model
(In our simulation as we used 4 targets, maximum no.of HMM model is possible to score is 12)

From Figure 2.17, it can be seen that for train HRR profiles in the 100 aspect case, HMM

recognition rate is maximum when we used 3 most likely models, assigned highest

similarity by ETMF. Hence, 100 aspect unknown observational profiles are classified

using the 3 HMM model which would get highest ranking in the preceeding ETMF step.

This technique of reduced HMM model scoring provides a 10-15% improvement in PCC

as compared to the case when all the HMM models are applied for scoring purposes.

60
2.3.2.2.2: Weight determination

In order to facilitate the hybridization with HMM, the correlation based ETMF scoring

was converted to probability measures by normalization of the correlated scores. If the

correlation is small, then the probability that the profile is the desired target will also be

small, and as the correlation increases, the probability increases and approaches one.

Assuming ETMF and HMM techniques are statistically independent, the joint probability

can be formed as:

PCombined = PETMF,HMM = PETMF PHMM (2.58)

where PCombined, PETMF and PHMM are discriminant likelihood’s indicating similarity

between the train templates and the unknown observation obtained from the hybrid

combination, ETMF algorithm and HMM algorithm respectively. Equation (2.58)

indicates the discriminant score achieved from the hybrid combiner is equal to the

multiplication of that obtained individually from ETMF and HMM. This condition is

achievable because ETMF and HMM similarity measures are independent of each other.

As the HMM output generates log-likelihood, the Probability expression in equation

(2.58) is converted to log-likelihood as,

log(PCombined ) = log(PETMF ) + log(PHMM ) (2.59)

This would be the simplest combination technique where the weightings on both ETMF

and HMM are equal. However, since the performance of ETMF in classifying targets is

much superior compared to that of HMM (Further improvement in HMM performance

may be achieved by feeding longer sequence of HRR profiles for estimating the HMM

training model parameters [62], which may not preferable for moving targets), the

61
proposed hybridization approach applies different weights to the log discriminant scores

before adding them to form the combined discriminant score.

The way of combining the two ATR algorithms is as follows:

log(Pcombined ) = W1 log(PETMF ) + W2 log(PHMM ) (2.60)

The weights W1 and W2 are determined from the training data and they are target and

aspect independent. For the 4-targets MSTAR data used in our simulation, the best choice

for weight proportions is 75% and 25% for ETMF and HMM, respectively. This decision

was made from Figure 2.18 (Combined recognition rate vs. W1/ W2).

Fig. 2.18:This figure is used to determine the most effective W1/ W2 so that the combined
ETMF+HMM recognition rate is the highest.

62
To find out the optimum W1/ W2, the training data was again split like optimum model

selection into training and testing sets. The “testing set” was formed by randomly

selecting 25% of the HRR profiles from each target in the training database. The

remaining 75% of HRR profiles from each target are regarded as training set and are used

to generate the Eigen-templates and HMM models. The “test” set was then used to find

the combined classification performance by varying W1/ W2 (as seen from the Figure

2.18). In order the system not to be biased for the specific test samples, ten different

Monte Carlo simulations are performed by taking ten random collections of “testing set”.

The % combined recognition rate shown in Figure 2.18 is the average of all ten combined

W1 3
recognition rates. It is found out that for = , the combined ETMF and HMM
W2 1

produced the maximum recognition rate. Hence, the same 3:1 ratio of W1/ W2 was later

used in actual test HRR profile scoring. Hence the final ETMF+HMM combined

discriminant score can be determined as:

log(Pcombined ) = 0.75log(PETMF ) + 0.25log(PHMM ) (2.61)

which ensures that the summation of W1 and W2 is equal to 1 i.e. W1+W2=1. The hybrid

ATR algorithm selects the highest logarithmic combined score and classifies the target

accordingly. In the simulation section it is shown that the combined ETMF-HMM hybrid

approach performs better than either technique used independently.

63
2.4: Simulation Results

2.4.1: Data Set

The proposed algorithm was trained and tested using MSTAR data set (without power

transform) containing HRR profile sets of 4 ground military vehicles (BMP2, T72, 2S1

and BRDM2) at 170 depression angle over 360 degree of aspect angles. The MSTAR data

set typically consists of SAR image chips. These images have been converted to the HRR

profile sequences through 1-D inverse Fourier Transform in the cross-range domain and

several filtering operations. The raw HRR profiles used in these studies were formed

from SAR imagery under the Defense Advanced Research Projects Agency (DARPA)

TRUMPETS Program and provided to Wright State University by Air Force Research

laboratories (AFRL), WPAFB, Dayton, OH [10].

2.4.2: ETMF Simulation Parameters

2.4.2.1: Formation of Template Profiles

ETMF based templates are generated from the MSTAR 170 depression training data set.

The templates at each target-aspect are formed by computing the eigen vector ( u 1 )

corresponding to the highest eigen value of the range space correlation matrix R̂ , where

ˆ = YYT , where Y={r1, r2,…, rN} i.e. the range profile matrices at each aspect.
R

Templates are formed at each aspect of all train targets.

64
2.4.2.2: Classification using Matched Filter Technique

In the ETMF technique, the observational profile a is correlated along ±8 shift with the

template profiles generated from the training HRR dataset. Before correlation, two

important preprocessing steps are performed which ultimately enhances the recognition

rate. The test profile is centroid aligned with respect to the templates and the overlap

region between test and templates are normalized prior to performing correlation. In the

Matched filter based classifier, each test profile is correlated against ± 10 HRR templates.

The matched correlation classifier is described in detail in [11-12].

2.4.3: HMM Simulation Parameters

2.4.3.1: Framing

Each HRR profile is blocked into frames of N samples, with adjacent frames being

separated by M (M<N). In the HRR framing part, N=12 and M=8 were selected.

2.4.3.2: Clustering

Vector Quantization (VQ) rule is followed for clustering the HRR frames. The commonly

used LBG algorithm for clustering a set of L frame vectors into a set of C codebook

vectors was formally implemented. Clustering is performed with cluster center (C)=32

(for this data C=32 was found to be a reasonable trade-off between performance and

computation efficiency). The clusters are target and aspect dependent. As MSTAR

dataset has K=4 targets with aspect angle data from 00 to 3600, we have 360 × 4 cluster

center matrix. To make the clusters more robust, HRR frames of a particular degree are

65
mixed with adjacent degree HRR frames before clustering and clustering is performed on

the combined HRR frame matrices, i.e. a particular degree cluster not only consists

profile information for that particular degree but also it’s nearest neighbors. For example,

the 20 cluster will have information not only for 20 but also for 10 and 30, and so on.

2.4.3.3: Training and Classification

In the next step of this classification algorithm, one HMM is designed for each target, i.e.,

there is a total of K HMM models, where K=4 in our simulation. Another important

factor for HMM design is to select the number of hidden states. The total training HRR

profiles are divided into N states using a uniform angular decomposition (for N states,

each state has an angular extent of 3600/N). In this algorithm, N=120 was used, so that

each state can account for 30 angular extent.

2.4.4: Single Look Case

Performance Improvement by Hybrid over ETMF and HMM

In this thesis the result of the proposed ETMF–HMM hybrid algorithm is compared with

those of the individual techniques in single look case. The results are presented for three

different classifier approaches, Hybrid classifier, ETMF based classifier and HMM based

classifier. The classifiers are designed in such a way that they can reject unknown targets

while maintaining high known target classification decision. The MSTAR data set used

to compare classifier performances is without power transformed because in [12] it is

66
shown Power transform degrades the classifier performance severely for noisy scattered

profiles.

The objective of the research is to demonstrate the performance enhancement of

the hybrid classifier over that of it’s component algorithms. Simulation results are

presented in the form of confusion matrices and are given in tables II-XI. In these tables,

the performance results for a single independent look are presented. Multilook techniques

can certainly improve the recognition rate further [6-7]. But as conventional Multilook

techniques are difficult to implement in tracking, results are presented in a new time

recursive multi target hypothesis updating procedure, which is time recursive version of

the stationary multilook approach and tracker friendly. Two types of classification tests

are performed, namely Forced decision and unknown target scenario. The former

assumes that all test targets belong to one of the known training target classes, which is

the unlikely in practical situation. The later makes no such assumption, so each test target

needs to be compared with a threshold before making any decision whether it is known or

unknown.

2.4.4.1: Forced Decision Result

In the Forced Decision case, the classification of the target class with the largest

likelihood determines decision for an observation profile. Table II summarizes the Forced

decision result for the above mentioned three algorithm. In table II, PCC comparison is

made among the proposed hybrid technique, ETMF alone and HMM alone for single

look-single profile, single look-3 profile average and single look-5 profile average

measures. It is found there, PCC of the ETMF-HMM hybrid technique is the highest for

67
all three cases. The PCC for ETMF, HMM and hybrid displayed in table I is actually the

average of all four target’s individual PCC’s. It can be also seen that the performance of

the HMM-only case is relatively poor, but when it is combined with the ETMF technique,

the overall PCC improves.

Averaging of several profiles certainly improve the recognition rate further as

observed from table II. However, for moving targets and tracking applications the

position of target changes with time. Hence, it is advantageous to use a single profile at a

time to obtain instantaneous track information. This is the reason for us to show the

classifier result on single look-single profile case. The classifier performance of ETMF,

HMM and hybrid techniques are also shown for 3 and 5 profile averaging. The results

demonstrate that averaging of profiles improves recognition rate significantly. In all 3

cases the hybrid technique outperforms the stand-alone methods.

Table I describes the organization of the confusion matrix for 4 target class data set

comprising of BMP2, T72, 2S1 and BRDM2.

Targets BMP2 T72 2S1 BRDM2

BMP2 P(BMP2|BMP2) P(T72|BMP2) P(2S1|BMP2) P(BRDM2|BMP2)

T72 P(BMP2|T72) P(T72|T72) P(2S1|T72) P(BRDM2|T72)

2S1 P(BMP2|2S1) P(T72|2S1) P(2S1|2S1) P(BRDM2|2S1)

BRDM2 P(BMP2|BRDM2) P(T72|BRDM2) P(2S1|BRDM2) P(BRDM2|BRDM2)

Table I: Organization of a Confusion Matrix

68
In the Forced decision case, The performance of three different classifiers is compared in

terms of Probability of Correct Classification (PCC). PCC is defined as the Probability of

correctly classifying a known target. In table I all the diagonal entries P(BMP2|BMP2),

P(T72|T72), P(2S1|2S1) and P(BRDM2|BRDM2) represents PCC for targets BMP2, T72,

2S1 and BRDM2. More the PCC, the better the classifier is in recognizing the unknown

range profile. Based on table I, the classification performance of Hybrid, ETMF and

HMM is compared in terms of Forced decision confusion matrices.

ATR Single 3 profile 5 profile


Classifiers profile averaging averaging
HMM 66.67% 82.17% 87.34%
ETMF 81.50% 91.04% 94.57%
Hybrid 85.55% 93.94% 96.63%

Table II: Summary of Forced Decision Results

Target BMP2 T72 2S1 BRDM2


BMP2 0.7727 0.0876 0.0682 0.0715
T72 0.0780 0.8330 0.0435 0.0455
2S1 0.0649 0.0334 0.8233 0.0784
BRDM2 0.0557 0.0474 0.0700 0.8269

Table III: Confusion matrix for ETMF with single profile testing (PCC=81.5%)

Target BMP2 T72 2S1 BRDM2


BMP2 0.6178 0.1302 0.1326 0.1194
T72 0.1315 0.7321 0.0830 0.0534
2S1 0.1207 0.0813 0.6725 0.1255
BRDM2 0.1214 0.0966 0.1404 0.6416
Table IV: Confusion matrix for HMM with single profile testing (PCC=66.67%)

69
Target BMP2 T72 2S1 BRDM2
BMP2 0.7731 0.0657 0.0845 0.0767
T72 0.0461 0.9233 0.0213 0.0093
2S1 0.0365 0.0134 0.8698 0.0603
BRDM2 0.0868 0.0378 0.0690 0.8364

Table V: Confusion matrix for Hybrid algorithm with single profile testing (PCC=85.55%)

Target BMP2 T72 2S1 BRDM2


BMP2 0.8724 0.0465 0.0413 0.0398
T72 0.0244 0.9515 0.0201 0.0040
2S1 0.0243 0.0078 0.9343 0.0336
BRDM2 0.0431 0.0265 0.0472 0.8832

Table VI: Confusion matrix for ETMF with three profile average testing (PCC=91.035%)

Target BMP2 T72 2S1 BRDM2


BMP2 0.7526 0.0556 0.1024 0.0894
T72 0.0548 0.8719 0.0450 0.0283
2S1 0.0409 0.0230 0.8697 0.0664
BRDM2 0.0546 0.0417 0.1112 0.7925
Table VII: Confusion matrix for HMM with three profile average testing (PCC=82.1675%)

Target BMP2 T72 2S1 BRDM2


BMP2 0.9160 0.0225 0.0262 0.0353
T72 0.0094 0.9781 0.0080 0.0045
2S1 0.0168 0.0040 0.9390 0.0402
BRDM2 0.0324 0.0111 0.0321 0.9244
Table VIII: Confusion matrix for Hybrid algorithm with three profile average testing (PCC=93.9375%)

Target BMP2 T72 2S1 BRDM2


BMP2 0.9182 0.0212 0.0350 0.0256
T72 0.0024 0.9855 0.0065 0.0056
2S1 0.0142 0.0054 0.9605 0.0199
BRDM2 0.0365 0.0101 0.0349 0.9185

Table IX: Confusion matrix for ETMF with five profile average testing (PCC=94.5675%)

70
Target BMP2 T72 2S1 BRDM2
BMP2 0.8271 0.0334 0.0750 0.0645
T72 0.0398 0.9226 0.0222 0.0154
2S1 0.0305 0.0122 0.9105 0.0468
BRDM2 0.0505 0.0208 0.0955 0.8332
Table X: Confusion matrix for HMM with five profile average testing (PCC=87.3350%)

Target BMP2 T72 2S1 BRDM2


BMP2 0.9490 0.0220 0.0252 0.0038
T72 0.0011 0.9962 0.0017 0.0010
2S1 0.0124 0.0092 0.9681 0.0103
BRDM2 0.0152 0.0103 0.0225 0.9520
Table XI: Confusion matrix for Hybrid algorithm with five profile average testing (PCC=96.6325%)

In tables III-XI the individual performance confusion matrices for ETMF, HMM and

Hybrid classifiers are displayed in single profile, 3 profile averaging and 5 profile

averaging environment respectively. The row targets of the confusion matrices denote the

train templates and the column targets represent observation profiles. In the Forced

Decision case, as it is assumed that each observation sequence belongs to any of the

known target templates, the major disadvantage is it’s inability to distinguish non-targets

from real targets. So Forced Decision classifier has limited practical uses. Though very

little application in airborne radar, Forced Decision is often used to measure strength of

ATR algorithms. From the Forced Decision result, it can be concluded that HMM

classifier performance is not at per compared to ETMF or hybrid classifier. As each HRR

profile contains large variability, longer sequence of profiles is needed to train HMM.

This is quite obvious when we observe table II. HMM classification rate improves from a

meagre 66.67% to high 82.17% (an improvement of about 15.5%) as we proceed from

71
single test profile classification to three profile average classification. Averaging or

sequencing of waveforms improves HMM classifier performance but then also the

performance results are below compared to that of ETMF or Hybrid.

120

100

80
HMM
PCC

60 ETMF
ETMF+HMM
40

20

0
1 3 5
No. of Profiles

Fig. 2.19: Bar plot representation of ETMF, HMM and Hybrid classifier performances (In the x-axis, 1,3 and 5
represents classifier performance on single profile, 3 profile average and 5 profile average observation profiles

Figure 2.19 shows a bar plot depicting the classifier performance comparison

between ETMF, HMM and Hybrid techniques. It can be seen that as the observation

72
profile averaging increases performance enhancement of the hybrid algorithm becomes

significantly smaller over it’s single algorithm components.

2.4.4.2: Classification in Unknown Target Scenario

In this case, the hybrid algorithm is applied to make classification decisions in the

unknown target scenario, which is simulated by rotating the target class using the LOOM

[10] approach. Thresholds for distinguishing between known and unknown targets are

generated according to LOOM approach explained in [11-12]. As the HMM classifier did

not perform as well as ETMF in Forced decision case, it is regarded that HMM classifier

is not at all a major force here and the hybrid classifier performance is compared with

that of ETMF only. The performance results of the classifier are summarized by ROC

curves. The performance parameters chosen are Conditional Probability of Correct

Classification (PCCCond), Probability of Target Detection (Pd) and Probability of false

Alarm (Pfa). Pd is the probability that a target declaration is made provided a known target

is observed. Known target observations classified as unknown are termed as “Miss” and

denoted as Pmiss.

Pmiss = 1- Pd (2.62)

PCCCond is the Probability of correctly classifying a target provided a target declaration is

made. Pfa is the Probability of classifying unknowns as known target. The primary aim of

any ATR algorithm is to minimize Pfa. The Hybrid and ETMF Classifier performances

are also shown through confusion matrices. Table XII describes the organization of the

confusion matrix for 4 target class data set comprising of BMP2, T72, 2S1 and BRDM2.

73
The performance comparison of Hybrid and ETMF ATR classifications is made

through confusion matrix like the one in Table XII. It is distinguishable from the Forced

decision confusion matrix shown in table I in the sense that an extra column and row

labeled Unknown in table XII. The Unknown column represents Miss decisions (Pmiss),

i.e. when the input unknown observation profile was tested against all the templates, none

of resulting discriminant scores were above the LOOM threshold. The Unknown row

represents the targets, which has no signature present in the training dataset. LOOM [22]

is responsible to create the unknown row. The most desirable results would be zeros in

the entire Unknown column and row entries except the lower right-hand corner, which

should be 1 i.e. unknown targets are 100% classified as unknown. This means that all of

the unknown observations were rejected and all of the known observations were

identified and recognized.

Target BMP2 T72 2S1 BRDM2 Unknown

BMP2 PCC Pmis-id Pmis-id Pmis-id Pmiss

T72 Pmis-id PCC Pmis-id Pmis-id Pmiss

2S1 Pmis-id Pmis-id PCC Pmis-id Pmiss

BRDM2 Pmis-id Pmis-id Pmis-id PCC Pmiss

Unknown Pfa Pfa Pfa Pfa 1- Pmiss

Table XII: Evaluation parameter computation from the confusion matrix

74
From the confusion matrix template shown in table XII, Pd , PCCCond and Pfa is calculated

as follows:

4 4 4
Pd = ∑ PCCi, j + ∑ ∑ Pmis −idi , j (2.63)
i, j=1 i =1 j=1

where i and j are row and column indices respectively. The upper limit of the summation

in equation (41) indicates there are 4 target templates present in the training dataset.

4
∑ PCCi, j
i, j=1
PCC Cond = ∑ (2.64)
( )
4
∑ 1 − Pmissi ,j = 5
i =1

4
Pfa = ∑ Pfai = 5,j (2.65)
j=1

Based on table XII, performances of Hybrid and ETMF classifiers are displayed in tables

XIII-XVIII.

Classifier Performance Results for ETMF processed single look-single HRR

observations:

BMP2 T72 2S1 BRDM2 Unknown


BMP2 0.6007 0.0375 0.0412 0.0560 0.2646
T72 0.0111 0.6995 0.0280 0.0185 0.2429
BRDM2 0.0277 0.0158 0.7027 0.0449 0.2089
2S1 0.0485 0.0399 0.0875 0.6090 0.2151
Unknown 0.1760 0.0540 0.0685 0.0807 0.6208

Table XIII: Confusion matrix (Unknown rejection threshold about 0.6) for ETMF based classifier

75
Classifier Performance Results for Hybrid single look-single HRR observations:

BMP2 T72 2S1 BRDM2 Unknown


BMP2 0.6347 0.0275 0.0539 0.0450 0.2389
T72 0.0174 0.7278 0.0073 0.0045 0.2430
BRDM2 0.0198 0.0055 0.7338 0.0249 0.2160
2S1 0.0358 0.0165 0.0474 0.6497 0.2506
Unknown 0.1275 0.0650 0.0686 0.0610 0.5779

Table XIV: Confusion matrix (Unknown rejection threshold about 0.6) for Hybrid classifier

Classifier Performance Results for ETMF processed single look-3 profile average

observations:

BMP2 T72 2S1 BRDM2 Unknown


BMP2 0.6448 0.0575 0.0212 0.0806 0.1959
T72 0.0241 0.7299 0.0566 0.0123 0.1771
BRDM2 0.0388 0.0285 0.7227 0.0395 0.1705
2S1 0.0392 0.0485 0.0721 0.6474 0.1928
Unknown 0.1345 0.0647 0.0822 0.1144 0.6042

Table XV: Confusion matrix (Unknown rejection threshold about 0.6) for ETMF based classifier

Classifier Performance Results for Hybrid single look-3 profile average


observations:

BMP2 T72 2S1 BRDM2 Unknown


BMP2 0.6997 0.0263 0.0358 0.0342 0.2040
T72 0.0104 0.7904 0.0040 0.0026 0.1926
BRDM2 0.0126 0.0025 0.7888 0.0205 0.1756
2S1 0.0237 0.0148 0.0408 0.7092 0.2115
Unknown 0.1405 0.0902 0.0889 0.0807 0.5997

Table XVI: Confusion matrix (Unknown rejection threshold about 0.6) for Hybrid classifier

76
Classifier Performance Results for ETMF processed single look-5 profile average
observations:

BMP2 T72 2S1 BRDM2 Unknown


BMP2 0.6892 0.0423 0.0188 0.0956 0.1541
T72 0.0207 0.7694 0.0511 0.0195 0.1393
BRDM2 0.0266 0.0122 0.7640 0.0255 0.1717
2S1 0.0300 0.0376 0.0601 0.6990 0.1733
Unknown 0.1200 0.0521 0.0700 0.1574 0.6005

Table XVII: Confusion matrix (Unknown rejection threshold about 0.6) for ETMF based classifier

Classifier Performance Results for Hybrid single look-3 profile average

observations:

BMP2 T72 2S1 BRDM2 Unknown


BMP2 0.7325 0.0335 0.0398 0.0276 0.1666
T72 0.0150 0.8193 0.0130 0.0060 0.1467
BRDM2 0.0215 0.0085 0.8090 0.0450 0.1160
2S1 0.0311 0.0209 0.0491 0.7371 0.1618
Unknown 0.1505 0.0899 0.0546 0.0945 0.6105

Table XVIII: Confusion matrix (Unknown rejection threshold about 0.6) for Hybrid classifier

The confusion matrices shown in table XIII-XVIII are generated by fixing the unknown

target rejection threshold at about 0.6 i.e. in 60% of the cases unknown targets are

regarded as unknown.

From each set of confusion matrix, performance parameters PCCCond, Pd and Pfa

are calculated. By varying the unknown rejection threshold over a wide range, a number

of confusion matrices and it’s associate parameters can be generated. The performance

results are expressed in terms of the parameters PCCCond, Pd and Pfa and shown via ROC

curves: Pd vs Pfa and PCCCond vs Pd.

77
PCCCond

Fig. 2.20: ROC curves for Probability of declaration vs Conditional Probability of Correct Classification

Fig. 2.21: ROC curves for Probability of False Alarm vs Probability of declaration

78
Since there are three different performance parameters PCCCond, Pd and Pfa, two ROC

curves are drawn to quantize the performances of Hybrid and ETMF. The First one is

showing Pd vs Pfa and the second one is Pd vs PCCCond. In Pd vs Pfa curve, the desirable

performance is to acheive high Pd while keeping Pfa at lowest possible. In Pd vs PCCCond

curve, the acceptable performance will be to acheive high PCCCond with Pd. Figure 2.20

and 2.21 depicts the classification performance on single profile. From Figure 2.20, it can

be seen that if the observed profile is identified as target, the probability that the Hybrid

algorithm classifies it in the correct target group is more than that of ETMF algorithm.

But in Figure 2.21, the hybrid algorithm generates more False Alarm at a fixed

probability of detection. This should not be accepted in ATR applications, as the

advantage of the Hybrid Classifier is limited to the known target cases only and moreover

it often makes mistakes in classifying unknown objects. This result is not surprising as

HMM performance is vey much poor if classification is performed on single observed

profile. The performance degradation of HMM on single profile is also seen in Forced

Decision case. So single profile classification is actually not the correct way to compare

the classification strength of the Hybrid algorithm. Hence it is decided to plot ROC

curves by forming the test profile using 3 profile averages and 5 profile averages.

79
PCCCond

Fig. 2.22: ROC curves for Probability of declaration vs Conditional Probability of Correct
Classification (each test profile is formed by averaging 3 observed range profiles)

Fig. 2.23: ROC curves for Probability of False Alarm vs Probability of Declaration (each test profile
is formed by averaging 3 observed range profiles)

80
PCCCond

Fig. 2.24: ROC curves for Probability of declaration vs Conditional Probability of Correct Classification
(each test profile is formed by averaging 5 observed range profiles)

Fig. 2.25: ROC curves for Probability of False Alarm vs Probability of Declaration (each test profile is
formed by averaging 5 observed range profiles)

81
PCCCond

Fig. 2.26: ROC curves for Probability of Declaration vs Conditional Probability of Correct
Identification (Single profile testing, average of 3 profile testing and average of 5 profile testing)

Fig. 2.27: ROC curves for Probability of False Alarm vs Probability of declaration (Single profile
testing, average of 3 profile testing and average of 5 profile testing).

82
PCCCond

Fig. 2.28: ROC curves for Probability of False Alarm vs Conditional Probability of Correct
Classification (Single profile testing, average of 3 profile testing and average of 5 profile testing).

Figure 2.22 and 2.23 shows the ROC comparison of Hybrid and ETMF algorithms on 3

profile average observation sequence. Here also the Hybrid technique generates more

False alarm at a fixed probability of declaration compared to that of ETMF. But the

encouraging part is that the false alarm gets reduced compared to what is observed in

Figure 2.21 at same Pd. That is what we expected. As the HMM classification

performance improves with profile averaging or profile sequencing, the hybrid algorithm

should improve it’s false alarm performance over profile averaging. The encouraging

result in 3 profile averaging prompts us to go forward in 5 profile averaging. As it is seen

in Figure 2.25, finally the hybrid classifier Pfa vs Pd ROC curve overtakes that of ETMF

classifier. Though the improvement is minor, but the trend is clearly visible. If we go

more and more test profile averaging, the hybrid classifier will outperform ETMF

83
classifier in a large scale. But we limit ourselves upto 5 profile averaging, as one of our

objective was to develop a classifier which will perform joint target tracking with

recognition. More profile averaging would hinder our tracking objective.

Next, the performances of Hybrid and ETMF classifiers are summarized in tabular forms.

At Pd = 0.7, PCCCond of Hybrid and ETMF are tabulated below:

Pd = 0.7 Hybrid Single Hybrid 3 Hybrid 5


Profile profile avg. profile avg.
PCCCond 91.15% 93.27% 94.94%

Pd = 0.7 ETMF Single ETMF 3 ETMF 5


Profile profile avg. profile avg.
PCCCond 87.20% 90.68% 92.91%

Table XIX: Conditional Probability of Correct Classification of Hybrid and ETMF classifiers at Pd = 0.7

At Pfa = 0.4, Pd of Hybrid and ETMF are tabulated below:

Pfa = 0.4 Hybrid Single Hybrid 3 Hybrid 5


Profile profile avg. profile avg.
Pd 75.20% 81.95% 89.75%

Pfa = 0.4 ETMF Single ETMF 3 ETMF 5


Profile profile avg. profile avg.
Pd 80.50% 83.30% 87.98%

Table XX: Probability of Declaration of Hybrid and ETMF classifiers at Pfa = 0.4

84
Figure 2.26 and 2.27 summarizes the performance parameters of Hybrid and ETMF on

single profile testing, 3 profile average testing and 5 profile average testing. In Figure

2.28, it is tried to compare the false alarm performance with Conditional Probability of

Correct Classification.

At Pfa = 0.4, PCCCond of Hybrid and ETMF are tabulated below:

Pfa = 0.4 Hybrid Single Hybrid 3 Hybrid 5


Profile profile avg. profile avg.
PCCCond 91.14% 94.15% 95.10%

Pfa = 0.4 ETMF Single ETMF 3 ETMF 5


Profile profile avg. profile avg.
PCCCond 86.40% 90.30% 93.10%

Table XXI: Conditional Probability of Correct Classification of Hybrid and ETMF classifiers at Pfa = 0.4

From table XIX and XXI it can be concluded that, for unknown target scenario at

constant false alarm, the hybrid technique with single profile achieves similar level of

performance as that attained by the ETMF algorithm using at least three profiles. These

results are promising as to develop a system that can perform joint tracking with

recognition, there is not much chance to perform classification on averaged profiles and

even if profile averaging can be done, it has to be done based on limited number of

profiles. So we are searching for an algorithm which can correctly classify targets as well

as that of other algorithms but with less number of profile averaging or sequencing. The

proposed Hybrid algorithm fulfills the criteria. From table XX it can be observed that at

constant false alarm, the probability of detection of the hybrid algorithm is much worse

85
compared to that of ETMF when target decision is made based on single profile. But as

profile averaging increases, the performance of hybrid is getting better and ultimately if

target decision is made on 5 profile average, the hybrid algorithm outperforms the ETMF

as far as probability of detection is concerned. The tables XIX, XX and XXI are

calculated based on a single operating point from the ROC curves shown previously. The

specifications assumed here as Pd = 0.7 and Pfa = 0.4 are arbitrary. Any other

specifications can easily be met by moving around on previously defined ROC curves. If

the specifications cannot be met from those ROC curves, one can consider

averaging/sequencing more number of test profiles to improve the classification

performance as it is clearly established in this paper that averaging/sequencing improves

performance of the classifier.

2.4.4.3: Computational Cost

Computational cost is always an important criteria as far as ATR is concerned as there is

the need to perform target recognition in the Aircraft with limited resources. The ETMF

classifier, we estimated is pretty fast and can be made operational in real time. That’s the

disadvantage of the hybrid classifier. As ETMF is combined with HMM and HMM takes

longer time to compute it’s likelihood, Hybrid classifier demands more computing power

than ETMF alone to operate in real environment.

The actual computation of the discriminants for the ETMF classifier varies

between 0.0050 s/signature to 0.0160 s/signature. Mostly the computation time in ETMF

case clusters around 0.0110 s/signature. In case of Hybrid classifier, discriminant

computation varies between 0.0320 s/signature to 0.0520 s/signature and mostly it is

86
clustered around 0.0460 s/signature. In summary, we can tell the ETMF-HMM Hybrid

classifier is about 4.2 times more computation intensive than ETMF alone. The reason

behind for such delay of Hybrid classifier is due to HMM Forward recursive likelihood

computation. Hence to implement the proposed Hybrid algorithm in real time, high

computive environment is a must.

2.4.5: Conclusion

The main objective of this research was to demonstrate that improved classification of

targets (PCC) could be achieved by hybridizing multiple HRR-ATR algorithms. We have

demonstrated that the proposed hybrid ETMF-HMM technique improves PCC when

compared to what is achievable by any one of the algorithms applied individually. It was

also demonstrated that instead of classifying with a single profile, the performance of the

proposed hybrid technique improves considerably if profile averaging is done while

making classification decisions. For forced-decision case, our results show that the hybrid

technique improves performance by 4.5%. Though this is not a significant improvement,

the usefulness of the Hybrid algorithm can be understood from the results obtained in

unknown target scenario. For unknown target scenario, the hybrid technique with single

profile achieves similar level of performance as that attained by ETMF algorithm using at

least three profiles. The results of this research will be integrated into current simulation

efforts for the research and development of a joint tracking based ATR system.

87
3: TIME RECURSIVE MULTIPLE HYPOTHESIS TESTING

3.1: Theory Development

It was seen that averaging over several profiles or multilook processing improves the

ATR classifier performance [7]. The multilook approach [7] is quite renowned to

improve the target recognition rate compared to single profile ATR. The main

disadvantage of multilook ATR is it is suitable for stationary targets, hence to combat

joint tracking with continuous target recognition for moving targets, the single look

ETMF needs to be applied time recursively as new profiles are observed over time. Here

we propose a time recursive sequential updating technique that will improve the ATR

performance and as well as will facilitate tracking. For sequential processing, when

multiple HRR profiles are observed over time, no particular target decision is taken based

on a single profile. Instead the maximum matched correlations for each target are saved

and combined later with those for other profiles coming in sequence. The main

characteristics for this sequential multiple hypothesis technique is the hypothesis score

for range profiles is always updated in a time recursive manner, hence the target decision

made at time t is dependent on all the hypothesis score from initial upto time t. To

facilitate tracking it is preferable to convert the correlations into probabilities, which can

be interpreted as target hypothesis measure at the particular time instant.

88
Let c1,c2,…cK be the discriminant scores for the K targets. If the discriminants are

correlation scores, the corresponding probability measures at the t-th time instant can be

found as,

ck
Ptk = K
, k = 1, 2,...K (3.1)
∑c
k =1
k

The detailed algorithm followed are discussed next in four steps:

Determining hypothesis measures over multiple range profiles:

Over a small time window N, the chosen ATR algorithm determines the target hypothesis

measures for multiple range HRR profiles at times t, t-1,…,t-N+1.

Combining multiple hypothesis measures:

Corresponding hypothesis measures for multiple range profiles observed at times t, t-

1,…,t-N+1 can then be combined to form the final multiple hypothesis measures for each

target as follows:

t
k
Pt|t,...,t − N +1 = ∏
l = t − N +1
Plk , k = 1, 2,..., K (3.2)

Where, N is the window of the time sequences. This technique is different than profile-

averaging approach or stationary multilook technique. Here, instead of averaging a

number of profiles, we combine the individual discriminant scores of a sequence of HRR

test profiles.

89
Classification based on combined hypothesis score:

The classification ID decision ( It|t,...,t − N +1 ) at the t-th time instant would correspond to the

maximum among these quantities.

It|t ,...,t − N +1 = max(Pt|tk ,...,t − N +1 ), k = 1, 2,..., K (3.3)

Here, it can be observed that the classification decision at time t depends not only on the

hypothesis score for the range profile at time t, but also on all the stored hypothesis score

for the range profiles observed at times t, t-1,…,t-N+1.

Updating combined hypothesis score:

As a new range profile is observed at time (t+1), the hypothesis can be updated as,

Ptk+1|t +1,...,t − N +1 = Pt|tk ,...,t − N +1Ptk+1 , k = 1, 2,...K (3.4)

For moving targets, it is important to use only a small window of range profiles as

considerably earlier observations may bias the decisions. The multiple hypotheses that is

limited to a window of N time sequences can be achieved as follows:

Ptk+1
Ptk+1|t +1,...,t − N = Pt|t,...,t
k
− N +1 , k = 1, 2,...K (3.5)
Ptk− N +1

The multiple hypothesis probabilities can be utilized by the tracker as confidence level

for each target under track. The block diagram next describes the time recursive multiple

hypothesis testing in schematic.

90
1
P t
Single profile ID 2
P
Range Profile at time t
Max
I t |t
t

M 1
1 Maximum P K P k | k ,K , k − N + 1
Templates t
Matched KN discriminants
2
Filter for K targets
discriminants being tracked
M
Hypothesis
P k | k ,K , k − N + 1
Range Profile at
Max M
time t-1
1 I t −1|t −1 M Sequential
K
Maximum Hypothesis Multiple P k | k ,K , k − N + 1
Train Matched Hypothesis
models Filter
discriminants Combiner
KN for K targets 1
being tracked Pt− N +1
Max
Range
profiles
Range
P t2− N + 1 I t |t ,K , t − N +1
Profile at
Max
It−N+1|t−N+1 M
time t-N+1
1
Maximum
P tK− N + 1
Matched
Filter
discriminants
Templates for K targets M
KN
being tracked Hypothesis

Fig. 3.1:Block diagram for time recursive multiple hypothesis Combiner

3.2: Simulation Results:

The strength of the proposed time recursive multilook approach is tested in two different

ATR classifiers: ETMF based classifier and ETMF-HMM Hybrid classifier. Also the

two-classifier algorithms are allowed to run on two different MSTAR data sets in order to

prove that the proposed time recursive multiple ID updating approach is independent of

classifiers and data sets to use.

91
3.2.1: ETMF classifier Simulation results

3.2.1.1: Description of MSTAR Data

The raw HRR profiles used in these studies were formed from SAR imagery provided to

Wright State University by Air Force Research laboratories (AFRL), WPAFB, Dayton,

OH [13]. These Data were collected under the MSTAR program and were distributed in

the form of two separate CDs to be used for template formation (17o) and testing (15o).

The data set contains SAR images of twenty-two military vehicles at two different

depression angle mode 15o and 17o over a total of 360o aspect angle. The target set

includes nine distinct military vehicles namely 2S1, BMP2, BRDM2, BTR60, BTR70,

D7, T72, ZIL131 and ZSU234 in addition to three variants of BMP2 and eleven variants

of T72. These images were converted to a sequence of HRR waveforms through the

IFFT operation in the cross range domain. The HRR signatures have a bandwidth of

600MHz centered at 10GHz.

The 17o depression angle images from nine of the twenty-two targets were used to

construct the HRR classifier template profiles: 2S1, BMP2 (SN #9563), BRDM2,

BTR60, BTR70, D7, T72 (SN #S7), ZIL131 and ZSU234. All the military vehicles used

in the template are distinct.

The HRR test profiles were generated from MSTAR target images taken at the

15o depression angle. The Eigen Template based ATR classifiers were tested on all

twenty-two targets covering the whole 360o aspect angle. Test targets, which are directly

included in the train database are called direct representative targets. The targets BMP2

(SN #9566), BMP2 (SN #C21), T72 (SN #132), T72 (SN #812), T72 (SN #A04), T72

92
(SN #A05), T72 (SN #A07), T72 (SN #A10), T72 (SN #A32), T72 (SN #A62), T72 (SN

#A63) and T72 (SN #A64) were not included in the training database but have

representative signature in the training set. These targets are referred as independent

targets.

3.2.1.2: Multilook Performance Results

The ETMF classifier performance is evaluated among the single profile hypothesis

(N=1), 3 profile hypothesis (N=3) and 5 profile hypothesis (N=5) through a set of ROC

curves shown in Figure 3.2-3.3. The time window size N is kept small as considerably

earlier decisions may bias the classification.

PCCcond

Fig. 3.2: ROC curves for Probability of detection vs Conditional Probability of Correct Classification

93
Fig. 3.3: ROC curves for Probability of False Alarm vs Probability of declaration

Figure 3.2 provides the relationship between Probability of declaration (Pd) and

conditional PCC. It is observed that for a fixed probability of declaration, (PCC)cond for

multiple profile hypothesis ETMF is greater than that of single look ETMF. Figure 3.3

provides a relationship between probability of declaration i.e. the probability to recognize

a test profile as target and probability of false alarm. The primary objective of all

automatic target detection schemes is to minimize the false alarm while improving the

probability of correct identification. It is observed here, for a fixed false alarm Pfa=0.4,

(Pd)single look = 60.52%, whereas (Pd)3 profile recursive = 76.85% and (Pd)5 profile recursive =78.59%.

Considering the performance of all the ROC curves it can be concluded that the proposed

94
time recursive multiple profile hypotheses improves the classifier performance compared

to single look ATR technique.

3.2.2: Hybrid Classifier Simulation results

The Hybrid ATR classifier performance is evaluated among the single profile hypothesis

Hybrid (N=1), 3 profile hypothesis Hybrid (N=3) and 5 profile hypothesis Hybrid (N=5)

through a set of ROC curves shown in Figure 3.4-3.6. The time window size N is kept

small as considerably earlier decisions may bias the classification.

Fig. 3.4: ROC curves for Probability of False Alarm vs Probability of declaration (single profile test,
Time recursive 3 profile updating test and time recursive 5 profile updating test)

95
PCCcond

Fig. 3.5: ROC curves for Probability of Declaration vs Conditional Probability of Correct Identification
(single profile test, Time recursive 3 profile updating test and time recursive 5 profile updating test)

PCCcond

Fig. 3.6: ROC curves for Probability of False Alarm vs Conditional Probability of Correct Identification
(single profile test, Time recursive 3 profile updating test and time recursive 5 profile updating test)

96
Figure 3.4-3.5 provides the relationship between Pfa- Pd and Pd-PCCcond. It is observed

that for a fixed probability of declaration, (PCC)cond for multiple profile hypothesis

ETMF is always greater than that of single look ETMF. Figure 3.6 provides a relationship

between Pfa- PCCcond. The primary objective of all automatic target detection schemes is

to minimize the false alarm while improving the probability of correct identification.

The performance comparison of single look Hybrid technique with Multilook

time recursive Hybrid techniques (N=3 and N=5) are displayed in the following tabular

form:

Pfa = 0.4 Hybrid Single Hybrid 3 Hybrid 5


Profile profile profile
recursive recursive
Pd 75.20% 83.15% 87.65%

Table XXII: Improvement of Probability of Declaration of Hybrid classifiers due to time recursive
multilook approaches at Pfa = 0.4

Pd = 0.7 Hybrid Single Hybrid 3 Hybrid 5


Profile profile profile
recursive recursive
PCCCond 91.15% 93.55% 96.10%

Table XXIII: Improvement of Conditional Probability of Correct Classification of Hybrid


classifiers due to time recursive multilook approaches at Pd = 0.7

Pfa = 0.4 Hybrid Single Hybrid 3 Hybrid 5


Profile profile profile
recursive recursive
PCCCond 91.14% 93.50% 95.25%

Table XXIV: Improvement of Conditional Probability of Correct Classification of Hybrid


classifiers due to time recursive multilook approaches at Pfa = 0.4

97
Considering the performance of all the ROC curves it can be concluded that the proposed

time recursive multiple profile hypotheses improves the classifier performance compared

to single look ATR technique. The time recursive profile updating technique is of great

help to tracking as here target ID is updating recursively with time and no profile

averaging is required.

3.3: Conclusion

It was demonstrated that instead of classifying with a single look, the performance of the

ATR technique improves considerably if time recursive sequential processing is done

while making classification decisions. The time recursive sequential processing is

important for performance improvement if the target is moving and changes position

instantly with time. As profile averaging is difficult for moving target, the only way to

improve target recognition performance is time recursive profile sequencing technique,

proposed in this thesis. The results of this research will be integrated into current

simulation efforts for the research and development of a joint tracking based ATR

system.

98
4: IMPROVED SAR TARGET DETECTION USING SUBSPACE

FILTERING

An ultra-wideband (UWB) synthetic aperture radar (SAR) simulation technique that

employs physical and statistical models is developed and presented. This joint

physics/statistics based technique generates realistic images that have many of the “blob-

like” and “spiky” clutter characteristics of UWB radar data in forested regions while

avoiding the intensive computations required for the implementation of low-frequency

numerical electromagnetic simulation techniques. The proposed image simulation

technique allows system designers to test in a variety of operating environments.

Firstly, the developed SAR target detection algorithms are trained in offline mode

to learn the clutter statistics. But as it is not always possible to learn beforehand,

approaches towards developing “self-training” algorithms for UWB radar target detection

are also investigated using the results of this simulation process. These adaptive

approaches employ some form of modified Singular Value Decomposition (SVD)

algorithm where small blocks of data in the neighborhood of a sliding test window are

processed in real-time to estimate clutter characteristics. These real-time clutter models

are then used to cancel clutter in the sliding test window.

99
4.1: Introduction

Synthetic Aperture Radar (SAR) imagery is commonly used for detecting, classifying and

recognizing mobile or stationary targets. Recognizing target from SAR images is an

important yet challenging problem especially if the target is hindered under outliers. In

previous work, it has been proposed a number of approaches to target detection [64-65]

in the ultra-wideband (UWB) SAR area. The approaches presented include detailed

discussion on a number of aspects of ultra-wideband radar target detection and algorithm

development. A bi-modal technique for modeling ultra-wideband radar clutter was

proposed in [64-65]. An approach to developing a new class of rank order filters, known

as, “discontinuity filter” for ultra-wideband radar target detection applications was

presented. These approaches mainly concentrate on the investigation of algorithms that

implement elaborate off-line training as well as the development of rank-order filtering

algorithms that are designed for basic UWB SAR sensor phenomenology but do not

require extensive off-line training. Both of these approaches have been shown to

generate acceptable levels of performance under certain conditions that are of interest for

UWB SAR applications.

This section of the thesis presents theoretical development and its associated

simulation results of a new ultra-wideband SAR target detection algorithm in offline and

self-training mode. A number of ways are investigated that implement self-training

procedures, where a set of localized regions within a given SAR image are sampled in

real-time for purposes of obtaining low-order and robust real-time clutter statistics. The

real time clutter models are applied in a sliding-window type target detection paradigm

for clutter cancellation and target detection. Results are presented from the analysis of

100
three new categories of algorithms that were developed for this investigation, denoted as

“Energy-Normalized SVD” (EN-SVD) “Condition-Statistic SVD” (CS-SVD), and

“Terrain-Filtered SVD” (TF-SVD). It is demonstrated that all three algorithms can

generate satisfactory results for severe UWB SAR impulsive-type clutter. Initially, all

three algorithms were allowed to run in offline training mode and later concentration is

given in developing self-training algorithm systems. The self-training mode can be

characterized as, no offline training is required and the algorithm will learn as it flies on

the observation terrain. Comparative results of the three SVD-based approaches to

adaptive and self-trained target detection algorithms are reported. The results indicate that

the CS-SVD and TF-SVD approaches, where a pre-filter is applied in an effort to

eliminate severe clutter discretes that adversely effect performance, appears promising

for the purposes of developing self-training algorithms. Development of self-training

algorithms are given emphasis because in some applications the target detection classifier

needs to learn while “on the fly” due to a lack of accurate off-line training data.

Simulated SAR impulse clutter is generated by modulating a “top hat” model for the SAR

video phase history with a K-distributed amplitude. Targets are synthesized within the

SAR image via the application of a dihedral model.

4.2: Ultra-wideband Radar simulation

Figure 1 shows a block diagram of the UWB SAR simulation developed for this

investigation. Two types of clutter were embedded into the SAR image: forest clutters

and residual background clutters. The central scattering structure for UWB forest clutter

101
elements are modeled by applying a physics-based parametric radar scatter center model

[66] that is based on the geometrical theory of diffraction.

Fig. 4.1: Block Diagram for UWB SAR Simulation

102
The general scattering model is described by [66]

 AHH
k
 α
n
 HV   f  i  2πf  4πf
E(f , φ) = ∑
k =1
 Ak   j  sin c 
 AVV   fc   c

Lk sin ( φ − φk )  exp  − j
  c
[ xk cos φ + yk sin φ]  (4.1)

 k 

Here we are assuming that the wavelength of the incident excitation is much smaller

compared to the target extent, so that the effective backscattered field is the vector sum of

the entire backscattered field emitted from each scatter center. The backscattered field E

is a complex function of frequency f and target aspect angle φ.

In equation (1), E (f, φ) is the set of backscattered field in all 3 polarization’s HH,

HV and VV and fc is the center frequency. The model parameters

{A k , x k , y k , α k }kn =1 characterize the n individual scattering centers. For the kth scattering

center, (xk,yk) gives the slant plane location with respect to a zero-phase reference, αi

characterizes the geometry and Ak is a complex scalar characterizing the magnitude and

phase. The round trip propagation delay from each kth scattering center is denoted as:

tk = 2/c(xkcosφ+ yksinφ) (4.2)

For this application, targets are modeled as long dihedrals and clutter scattering

centers are modeled as short tophats. For a single-dihedral target model with HH

polarization, A HV
k
and A VV
k
are zero, and αi is equal to one . The synthetic images were

created using Ohio State’s models, followed by SAR image formation. Tree trunks were

modeled as tophat reflectors and targets were modeled as varying length dihedrals. The

forest was created by randomly placing trees in the image. The reflectivity of the trees

are K-distributed, and the locations are uniformly distributed. The tree canopy was

modeled with a Gaussian distribution with the mean and the canopy radius both

103
proportional to the trunk amplitude. A 2D Hanning window was used to smooth the

transition between the canopy and the background. Finally, Gaussian noise was added to

represent the noise floor.

In order to model the basic canonical clutter structure as tophats, the φ

dependence in the argument of the sinc term in equation (4.1) is removed and the sinc is

varied as a function of the dihedral length, L, only. The total number of synthetic

scatterers in the SAR video phase history (VPH) computation is denoted by n. As shown

in the VPH block of Figure 4.1, the VPH is synthesized by breaking the summation in

equation (4.1) into two parts and then selecting the parameters for n scatterers: long

dihedrals for targets (15 targets is the default number with length varying from 0.5 to 1.5)

and tophats for clutter (n-15 number clutter scatterers). Here the chosen n is 600, a

default number. The emphasis was given to have a high clutter to target scatterer ratio,

which is 20:1 in our simulated image. The x, y locations of the clutter scatterer centers

are assumed to be uniformly distributed.

In addition, for this simulation approach, we make the heuristic argument that the

number of “scatterers” in the neighborhood of a tophat scattering center (due to canopy

scattering, etc.) is random. This allows us to generalize statistical arguments from the

statistical signal processing literature [67] that indicate that, if the number of scatterers in

an “elementary cell” is finite and random, the number of scatterers is roughly Poisson

distributed. The expected value of this Poisson distribution is typically also assumed

random with a gamma distribution, and the resulting intensity distribution is typically

assumed K-distributed. K-distribution was chosen for clutter intensity modeling due to

the assumption that neighboring clutters are correlated. For this simulation, we generalize

104
the concept of an “elementary cell” to include the region in the immediate local

neighborhood of a tophat scattering center including statistical angle-dependent scattering

due to the neighboring canopy region. The parameters of the K-distributed intensity are

randomly distributed to generate a severe clutter environment that has many instances of

impulsive “spikes” and “blobs”. To this end, a challenging simulation environment was

generated by implementing a joint physics-based (equation 4.1) and statistical (K

distributed tophat scattering centers) approach that avoids intensive numerical

electromagnetic computations.

The proposed approach of simulating targets in clutter environment offers a

number of practical advantages. Firstly, it is cost-effective, as highly expensive data

collection is not needed. The software-based capability allows us to vary clutter severity,

the number of targets and separations between targets etc. to resemble any desired

realistic UWB images. Hence, it is possible to test the effectiveness of various target

detection algorithms for many variations of target and clutter operating scenarios.

The x, y coordinates of the VPH model in equation (4.1) are further synthesized

and transformed into airborne SAR range coordinates and then processed by a wideband

wide-angle SAR image formation algorithm known as “wavefront processing” [68].

After SAR image formation with wavefront processing, the resulting SAR image is post-

processing for additive noise and for local variations in the residual clutter level in

accordance with the following two-attenuation model [69].

The radar signal attenuation due to foliage can be summarized for the case of typical

woods in the temperature zone as:

 sin π 4 
Two-Way Attenuation (dB) = β f α   (4.3)
 sin γ 

105
where, f is the frequency in Megahertz, γ is grazing angle, and, for HH polarization, the

median attenuation is for α = 0.79 and β = 0.044 and the 90th percentile attenuation is for

α = 0.79 and β = 0.077. The α and β values are chosen according to the forest clutter

model as described in [69].

Two samples of UWB SAR imagery generated with this simulation process are

illustrated in sub-section 4.5 for the purposes of testing and training. In the next two sub-

sections the SVD operation on SAR images and its effectiveness in SAR clutter

suppression are discussed.

4.3: Eigen-analysis of SAR image

Given an N x N SAR image data window Y, basic singular value decomposition (SVD)

operation can be mathematically formulated as follows [11-12]:

M
SVD
Y → UΛV T = ∑λ u v
i =1
i i
T
i (4.4)

where,

Range-Space (Left) Eigen Vectors :

U = EV[YY T ] = EV[R] = [u1 ......u N ] ∈ℜ N× N (4.5)

Angle-Space(Right) Eigen Vectors :

V = EV[Y T Y] = [ v1 ....v N ] ∈ ℜN×N (4.6)

Singular Values :

Λ = Diagonal[λ11 ......λ MM ] ∈ ℜN×N (4.7)

The EV[·] operator represents “the eigenvectors of” the matrix inside the brackets. For

Range Vs. Crossrange SAR data, the left eigen vectors (U) span the orthogonal basis

106
space in the range domain while the right eigen vectors (V) span the Crossrange domain.

The middle matrix Λ is diagonal containing N singular values in decreasing order,

λ11 ≥ λ 22 ... ≥ λ NN , where λ ii denotes the weights associated with i-th eigenvector.

4.4: Clutter Suppression Capability of SVD

A critical problem faced in ATR is clutter noise that often corrupts SAR images.

Noise/clutter reduction is a common problem faced in many different fields of radar,

sonar, speech, image etc., where SVD has been found to be highly effective for noise

suppression and also for improved estimation of underlying system parameters. The

range-crossrange orthogonal basis spaces with dominant singular values correspond to

the “Signal subspace”, whereas those corresponding to the smaller singular values are

referred as “noise-subspace” (or “clutter-subspace”). Hence in order to suppress noise a

common approach has been to reconstruct the original matrix, after "zeroing-out" what

are considered to be the noise or clutter subspaces i.e. the clutter reduced HRR matrix can

be reconstructed as,

N
ˆ =
Y ∑ λˆ uˆ vˆ
i =1
i i i
T
(4.8)

where, λ̂ i is the estimated set of dominant eigen values, ûi and v̂ i are the estimated set

of range domain and crossrange domain eigenvectors corresponding to the dominant

eigenvalues, respectively. This approach is known as Eigen Filtering or Subspace

Filtering. In the next section we discuss the three SVD-based algorithms that are

developed to adaptively cancel UWB radar impulse clutter based on the approach of

decoupling clutter from Signal subspace using subspace filtering.

107
4.5: SVD based SAR Target Detection Algorithms

Theoretical approach of the three proposed SVD based algorithm EN-SND, CS-SVD and

TF-SVD are descibed next.

4.5.1: Energy Normalized SVD (EN-SVD)

4.5.1.1: Motivation

The proposed algorithm for energy normalized SVD technique is based on Eigen

filtering. Eigen filtering is an effective way to decouple target from clutter. In this

algorithm the sum of the ratio of eigenvalues between localized windows are taken as a

metric to discriminate between target and clutter. This idea is taken from the

homogeneity test, described in [70] where it is shown that the generalized inner product

(GIP), which is a test of eigen structure and detecting forms of non-homogeneity is the

sum of the ratio of the eigenvalues of the range-crossrange space correlation matrix.

Homogeneity test was chosen as the discrimination factor between target and clutter due

to the homogeneous nature of clutter scatterers i.e. energy of the clutter scatterers are

uniform, whereas targets are mostly rank one i.e. the highest eigenvalue contains

significant amount of energy compared to those lesser ones. In UWB environment, the

motivation of developing SAR target detection algorithm based on eigen decomposition

came from the plotting of eigenvalues of a 15 × 15 target only image block with that of

same size clutter block. From Figure 4.2, it can be verified that in case of target, the

highest eigenvalue contains significant amount of energy compared to it’s smaller ones,

but in case of clutter, atleast 4-5 eigenvalues are necessary to completely describe the

108
clutter blob. Hence for this reason, the magnitude of the highest eigenvalue of the target

blob is far too high compared to that of clutter blob. This information can be used as the

discriminant factor between target and clutter image block and the idea of the proposed

EN-SVD algorithm took place.

In summary, the proposed EN-SVD algorithm tabulates histograms of the sum of

eigenvalue ratios from pairs of neighboring image windows. Statistics from this

“training” histogram are used to determine a threshold. During the “testing” phase, sums

of eigenvalue ratios are computed for each localized test clutter window with respect to

the localized test window with the minimum energy (i.e. minimum sum of eigenvalues).

The minimum energy is chosen as a reference because background clutter energy is

assumed to be lower than that of target. If the test “sum of eigenvalue ratio” metric is

below the train threshold, the value in the given test window are “zeroed out. If this test

metric exceeds the train threshold, the data in the test window is reconstructed from the

pre-dominant SVD eigenvalues and eigenvectors. This process attempts to eliminate the

effect of background clutter while keeping the targets intact.

The detailed version the algorithm is described in the next section.

109
Fig. 4.2: Eigen Spectrum of Target and Clutter blob

4.5.1.2: EN-SVD Training Phase

The detailed steps for the training phase are outlined as follows:

1) The train clutter only image is processed as segmented 3m × 3m regions that are each

divided into n=9, m × m windows.

Figure 4.3 shows the image segmentation and feature extraction process in graphical

manner. A11, A12..., A33 are the respective feature sets extracted from each m × m region

of the images respectively.

110
A11 A12 A13
m

A22
m

A21 A23
3m

A31 A33
A32
3m

Fig. 4.3: SAR image feature extraction

2) The highest m/3 eigenvalues are calculated for each m × m window within a given

3m × 3m region. m/3 number is chosen as it is seen experimentally that the highest

m/3 eigenvalues contains more than 95% of energy irrespective of target or clutter in

each m × m region. In the EN-SVD approach, as sum of ratio of the eigen values are

chosen to be the discriminant factor between target and clutter, hence eigen values are

treated as feature components and the feature vectors A11, A12..., A33 each contains

the top m/3 eigenvalues extracted from each m × m window.

111
3) Sum of the ratio of A11 with all other (n-1) feature vectors are calculated. This yields

a vector B1 of elements (n-1) “sum of eigenvalue ratio” metric values.

 A A A A 
B1 =  ∑ 11 , ∑ 11 ,..., ∑ 11 ,..., ∑ 11  (4.9)
 A12 A13 A 22 A 33 

4) Similarly sum of the ratio of A12 with all other (n-2) feature vectors (n-2, because

A11
is already done in step (3)) are calculated and stored in B2.
A12

 A A A A 
B 2 =  ∑ 12 , ∑ 13 ,..., ∑ 12 ,..., ∑ 12  (4.10)
 A13 A 21 A 22 A33 

A 32
This process will go on, until we get, B n −1 = (4.11)
A 33

5) All individual vectors B1, B2,..., Bn-1 are concatenated to generate a large B vector.

B = [B1, B2,..., Bn-1] (4.12)

6) Next, in a similar procedure B vector is created for a number of different clutter only

m × m images and they continuously concatenated with previous B vectors.

7) With all the elements of B a histogram is formed and T=mean+sigma (sigma signifies

one standard deviation of the histogram) of the histogrammed output will be denoted

as feature threshold to discriminate between target and clutter. It is to be noted that,

more number of m × m clutter only regions are used to train the EN-SVD classifier,

more robust the classifier will be to discriminate between target and clutter.

112
4.5.1.3: EN-SVD Testing Phase

The detailed steps for the testing phase are outlined as follows:

1) The test SAR image is processed as segmented 3m × 3m regions that are each divided

into n=9, m × m windows same as train-clutter image. The respective highest m/3

eigenvalue feature sets are calculated from each small m × m window and denoted as

A11, A12..., A33 in the same way as shown in Figure 4.3.

2) As eigenvalues are directly proportional to energy, the m × m window with least

energy is found out. Let the feature vector of the least energy window is denoted as

Akl (1 ≤ k ≤ n / 3,1 ≤ l ≤ n / 3) .

3) Sum of the ratio of Akl with all other (n-1) feature vectors are calculated. This yields a

vector C of elements (n-1) “sum of eigenvalue ratio” metric values.

 A A A A 
C =  ∑ kl , ∑ kl ,..., ∑ kl ,..., ∑ kl  (4.13)
 A11 A12 A 22 A 33 

Hence, C will be a (n-1) element array.

4) Each (n-1) metric value in vector C is compared with the predetermined threshold T

obtained from the training phase.

If T ≥ C( j) , 1 ≤ j ≤ n − 1

Mapping is done from that corresponding j to the m × m window, which is

responsible for generation of C(j) with the lowest energy window. The corresponding

m × m window is zeroed out assuming that window consists of clutter only.

113
Else T ≤ C( j) , 1 ≤ j ≤ n − 1

Again the correspondence of C(j) to m × m window is found out. This time it is

assumed that the window consists of a whole/part of a target and the data in that

window is replaced with the dominant order eigenvalue-eigenvector expansion.

5) The entire operation described in steps (1)-(4) is repeated throughout the whole image

in a sliding window type fashion.

4.5.2: Condition-Statistic SVD (CS-SVD)

In this approach, the training data is statistically pre-filtered before running the EN-SVD

from section (4.5.1). The need for the CS-SVD can be observed in the simulation section,

as the performance of the EN-SVD is not satisfactory in ROC-curve comparison. This is

due to significant amount of target information loss in EN-SVD case. The problematic

area detected is the presence of large bright clutter blobs in the image, so before

implementing EN-SVD on the SAR test image, the bright clutter blobs should be

removed as much as possible by statistical filtering.

Although many possibilities for a statistical filter can be investigated, for this

particular investigation a simple mean filter is applied. The main intention here is to

demonstrate the performance improvement of EN-SVD by statistical filtering. As it is

observed that mxm window is sufficient to capture the whole clutter blob, mean is

computed over a sliding mxm training regions of interest. Then, the mean of the

resulting mean-filter output over all training regions is used as a “blob” threshold for the

training. In other words, any mxm training region whose mean exceeds the mean

114
threshold is denoted as severe “blob-type” impulse clutter and are zeroed out from the

training set. Hence what we get is a filtered train image. Next the training algorithm of

the EN-SVD are applied on that image to generate the threshold T.

For testing, the same mean prefiltering steps are followed on the observation/test

image before applying the same EN-SVD testing algorithm. As shown in the simulation

section, this approach can allow EN-SVD to more effectively cancel the “typical clutter”

while preserving more targets and can yield enhanced target detection performance.

4.5.3: Terrain-Filtered SVD (TF-SVD)

4.5.3.1: Motivation and Kernel formation

The terrain-filtered SVD approach is a combination of convolutional filtering and energy

normalized SVD approach. The motivation of this algorithm remains the same as that of

CS-SVD. Instead of applying a pre-filter in CS-SVD case, here an Euclidean mask is

used to suppress the bright clutter blobs. In this approach a convolutional prefiltering step

is performed instead of mean prefiltering before applying the EN-SVD. Here, a

convolutional spatial filter kernel is designed that computes an intensity-weighted

distance metric from the kernel center in an effort to pre-filter severe “blob-like” and

“sparsely-impulsive” clutter discretes. The particular filter kernel designed for this

investigation is illustrated in Figure 4.4. This convolutional pre-filter is implemented as a

sliding window over the observation/test image. This kernel function is calculated by

setting all pixels within an 11-pixel radius of the kernel center equal to zero and by

equating the remainder of the pixels equal to the Euclidean distance between the kernel

115
center and the pixel. The test image is convoluted over pixel by pixel sliding terrain and

all the convoluted outputs are used to generate a histogram of convoluted outputs.

The primary motivation for the formation of such type of Euclidean mask came

due to the discrete nature of clutter blob. The scattering centers of the clutter blobs are

distant apart unlike that of targets where they are quite coherent. The designated mask,

assuming the target is at center, tried to estimate the density around the target. If target

remains at the center of the sliding mask and there is no bright scatterers around the

target, the convoluted output will be zero as the center target is already masked by the

kernel. Hence if the convoluted output generates a large value, that means there are

discrete scatterers present around the target at large distance (because here we are

multiplying the scatterer intensity with distance from the center) and it is needed to

suppress those discrete blobs in order to improve the target detection performance of EN-

SVD.

Fig. 4.4: Filter kernel for “Terrain-Filtered SVD”

116
Figure 4.5 shows the histogram of the data at the output of the terrain prefilter, along with

the “center”, Sigma-1, and Sigma-2 values for this simulation. The resulting histogram

filter data from the test image is processed in accordance with the following procedure:

Sigma1 Sigma2
Center

Fig. 4.5: Sample Filter Histogram for TF-SVD

The maximum value of the filter output histogram distribution is denoted as the “Center”

and Sigma-1/Sigma-2 are standard deviation of all the filtered values to the left/right of

the histogram “Center”. A given test image pixel is set equal to zero if the corresponding

filter output is beyond Sigma-1 to the left of the histogram “Center” or is beyond “Sigma-

2” to the right of the “Center”. It is assumed that those test image pixels correspond to

either blob type clutter or discrete residual clutter. But those test image pixels whose

corresponding filter output is in between sigma-1 and sigma-2 of the histogrammed plot

are kept intact assuming target information with clutter are embedded in those region and

117
EN-SVD will be performed on this image to nullify further clutter and improve target

detection performance.

This filter has the effect of filtering-out a significant number of severe “blob-like”

and “sparsely-impulsive” clutter discretes.

4.5.3.2: Implementation steps of TF-SVD

The overall TF-SVD algorithm can me implemented in the following three steps

described next:

a) The sliding-window terrain filter kernel is applied on the train clutter only image and

store histogram of filter output.

b) Center test image pixels corresponding to the image-mask region is set equal to zero

in accordance with histogram Sigma-1 and Sigma-2 criteria as described in the kernel

formation section. We can assume mostly all the clutter blobs and discretes are

nullified after this approach. Then EN-SVD algorithm is applied on the train clutter

only image to generate the target-clutter discrimination threshold T. EN-SVD

algorithm is applied on the same image to eliminate the clutter blobs that might be

present in between Sigma-1 and Sigma-2 of the histogrammed prefilter output.

c) For testing, the same Euclidean masking steps are followed on the observation/test

image before applying the same EN-SVD testing algorithm. As shown in the

simulation section, this approach can also allow EN-SVD to more effectively cancel

118
the “typical clutter” while preserving more targets and can yield enhanced target

detection performance.

In the simulation section it is also shown that these approaches has potential to act as self

training algorithms i.e. where no separate train image is required to learn the algorithm,

the algorithm will learn itself while on the fly.

4.5.4: Simulation results

Fig. 4.6: Sample UWB SAR Simulation test Image (target with clutter, where targets are boxed
with red rectangle)

119
Fig. 4.7: Sample UWB SAR Simulation Clutter only image

4.5.4.1: UWB SAR simulated image specification

In this section, a set of comparative results are presented by running the EN-SVD, the

CS-SVD, and TF-SVD algorithms on the sample simulation test image of Figure 4.6. The

synthetic image was created using Ohio State’s models, followed by SAR image

formation. This sample image shows many instances of “spiky” and “blob-like” realistic

UWB radar clutter. The targets, modeled as dihedrals, are shown along three rows on the

top, middle, and bottom of the simulation. There are altogether fifteen targets in the

image, whereas five targets are placed on each row. The bottom row has relatively long

targets (length L=1.5), the middle row has targets of relatively medium length (length

120
L=1), and the top row has short targets (length L=0.5). The target backscatter level or

reflectivity varies from left to right. The reflectivity was varied based on the length of the

targets. The reflectivity was calculated as: f = αL , where α varies from 2 to 10. The

UWB radar clutter was modeled in two phases: 1) bright or blob clutter modeled as tree

trunks and tree canopies 2) residual clutter. Tree trunks were modeled as tophat reflectors

and targets were modeled as varying length dihedrals. The forest was created by

randomly placing trees in the image. The reflectivity of the trees are K-distributed, and

the locations are uniformly distributed. The tree canopy was modeled with a Gaussian

distribution with the mean and the canopy radius both proportional to the trunk

amplitude. A 2D Hanning window was used to smooth the transition between the canopy

and the background. Finally, Gaussian noise was added to represent the noise floor.

All the proposed SVD based target detection algorithm is trained first by feeding

the clutter only image shown in Figure 4.7. From that clutter only image, the SAR target

detection algorithms would learn the clutter statistics and determine the threshold to

distinguish between target and clutter. Then the performance of all three-detection

algorithms is tested in Figure 4.6 and their performance comparison is made through a set

of ROC curves. It is to be noted that the clutter only simulation UWB image (Figure 4.7)

and clutter with target UWB simulation image (Figure 4.6) were generated using two

different simulation runs and conditions are put such that random numbers which

determine position and intensity of clutter blobs/residual clutter varies with different

simulation runs. This was done intentionally to make those two images uncorrelated, as

finally the proposed algorithm will be tested in self-training mode. Each segmented

window size is taken m=15 as the maximum size of a clutter blob is 15x15.

121
4.5.4.2: Performance Comparison of Target Detection algorithms

4.5.4.2.1: Performance Comparison in Offline training mode

All three SAR target detection algorithms EN-SVD, CS-SVD and TF-SVD are first

applied on the train clutter only image (Figure 4.7) to generate the threshold, which is the

distinguishing factor between target and clutter. Then all those three algorithms are tested

in the target-clutter image (Figure 4.6) and nullify clutter based on the predetermined

threshold.

Fig. 4.8: UWB SAR simulation image after performing EN-SVD on Figure 4.6

122
The performance of the EN-SVD algorithm on Figure 4.6 is displayed in Figure 4.8.

From the figure, it can be seen that although the algorithm nullifies significant amount of

clutter from the neighborhood of targets, severe target information is also lost. Especially

the targets which have low backscatter energy or have smaller length suffers the most.

The colored boxes are drawn around the remains of the visible target. Those targets,

which can be visible almost completely, are surrounded by a red box and the partial

visible targets are surrounded by yellow rectangles. From Figure 4.8, it can be seen that

only two targets are fully recovered and five other targets are partially recovered after

implementing EN-SVD on Figure 4.6. This is highly unwanted, as the EN-SVD detection

algorithm looses a number of targets almost completely. This will definitely reduce the

performance of EN-SVD and this is the motivation for go beyond EN-SVD and the

algorithms of CS-SVD and TF-SVD come into picture. The reason for worse

performance of EN-SVD is quantified due to the presence of bright clutter blobs whose

reflectivity is comparable with that of targets. Hence because of the high reflectivity

clutter blobs, the threshold obtained from the training process is relatively high, which is

nullifying not only significant amount of clutter but also large portions of targets. Hence

the necessity is to implement some pre-filtering/masking approach so that the high-

energy clutter blobs are suppressed before applying EN-SVD. The CS-SVD does just

that, it just has a pre-filtering overhead with EN-SVD. Figure 4.9 displays the image after

applying CS-SVD algorithm. From there it is observed that not only it zeros out the

clutter blobs and clutter discretes, but also it preserves significant amount of target blobs

compare to Figure 4.8.

123
Estimated UWB Image

Figure 4.9: UWB SAR simulation image after performing CS-SVD on Figure 4.6

Figure 4.9 is the displayed SAR image after implementing CS-SVD on Figure 4.6. Here

also the partial/full remain targets are boxed with color rectangles. The red rectangle

surrounds those targets whose maximum portions are still visible in the image and the

yellow rectangle covers the targets whose significant portions are removed but still can

be distinguishable as a target. From Figure 4.9, it can be observed that the CS-SVD

124
algorithm retains at least six targets in full and three targets in partial i.e. much

improvement compared to EN-SVD alone.

Fig. 4.10: UWB SAR simulation image after performing Euclidean masking operation in TF-SVD

Figure 4.10 shows the image after implementing the TF-SVD pre-masking operation on

Figure 4.6. It is observed that the Euclidean masking operation preserved almost all

targets except the leftmost in the top row. The main importance part here is all the large

bright blob type clutters are nullified from the image. Now the EN-SVD algorithm will

125
be applied on Figure 4.10. As the performance detoriating blob type clutter are not there

in Figure 4.10, certainly performance improvement is expected.

Fig. 4.11: Final UWB SAR simulation image after performing TF-SVD

From Figure 4.11, it can be observed that finally the TF-SVD algorithm retains almost all

the targets either full or partial. The targets whose most parts are visible are surrounded

126
by red rectangle and those targets which looses a significant part but still can be

recognized as target is shown in yellow box.

Till now we compare the performance of EN-SVD, CS-SVD and TF-SVD

qualitatively. From the qualitative result, it is expected that both CS-SVD and TF-SVD

will exceed the performance of EN-SVD. The quantitative performance measurement of

the three algorithms is performed using ROC curves.

Fig. 4.12: Performance comparison of EN-SVD, CS-SVD and TF-SVD in offline training-real time
testing mode

127
Fig. 4.13: Performance comparison of EN-SVD, CS-SVD and TF-SVD shown in logarithmic scale in
offline training-real time testing mode

The ROC curves shown in probability of detection as a function of varying the false

alarm threshold in Figure 4.12 and 4.13. In Figure 4.13, the x-axis is plotted in

logarithmic scale. The ROC result confirms the visual inference that CS-SVD and TF-

SVD outperforms the baseline EN-SVD in terms of target detection performance. For

probability of false alarm at 0.1, probability of target detection in case of EN-SVD is

0.66, whereas at the same false alarm probability, CS-SVD provides 0.72 and TF-SVD

128
provides 0.8 as probability of detection. In case of EN-SVD, probability of detection goes

to 1 (100% detection) at a high false alarm rate of 0.78, but in case of CS-SVD and TF-

SVD, 100% detection is achieved only at a small false alarm of 0.2 and 0.22. From the

above ROC plots, it can be concluded that CS-SVD and TF-SVD clearly outperforms the

baseline EN-SVD.

4.5.4.2.2: Performance Comparison in Self-Training mode

Next, the three proposed SAR target detection algorithm EN-SVD, CS-SVD and TF-

SVD are compared in self-training mode. In the self-training mode, the three algorithms

will make use of only the simulated target-clutter image (shown in Figure 4.6), as it will

learn while on the fly. Here, the assumption is there is no absolute requirement of training

image, as the algorithms will learn on the fly while performing target detection. The great

advantage of the self train algorithm compared to other offline training techniques is 1)

here the algorithm need not be familiar with the environment 2) the algorithm will learn

the clutter statistics while simultaneously performing target detection, hence it performs

efficient time and resource management. Clearly, development of self training algorithm

is a practical requirement, as sometimes it is impossible to collect training or

representative data a priori and the target detection algorithm needs to function in an

unknown environment.

The functionality of the proposed target detection algorithms EN-SVD, CS-SVD

and TF-SVD are analyzed in self-training mode via a set of ROC curves.

129
Fig. 4.14: Performance comparison of EN-SVD, CS-SVD and TF-SVD in self-train mode

From Figure 4.14 it can be seen that the target detection algorithms are compared in local

self train mode and global self train mode. In global self train mode, first the whole test

image is generated. Then the threshold to discriminate target with clutter is computed

based on the whole test image. Lastly the algorithm again works on the image and

nullifies clutter while keeps target. It is to be noted that, as each algorithm works on

segmented images, the threshold remains constant for all image segments here. This is a

disadvantage as the global threshold is based on the clutter statistics of the whole region,

130
it may not reflect the statistics of area-localized clutter. So the threshold obtained may not

be robust for all parts of the image. The primary disadvantage is in this way of threshold

computation the algorithm needs to be worked two times on the image: First time it will

calculate the threshold and then again the control point will be shifted to the starting point

of the image and it starts nullify clutter based on the threshold. Hence this process can

hardly be made as real time. As SAR generation is itself a high time consuming process,

the objective of any target detection scheme based on SAR is to reduce further overhead.

To combat with this requirement, the scheme to generate threshold locally while

performing target detection is proposed. That means, one threshold is computed for every

segmented portion of the image. Each algorithm first computes the threshold from a small

mxm segmented portion and then removes clutter from that segmented portion based on

the threshold calculated before. The main advantage is each localized area has it’s own

threshold i.e. the threshold is determined based on localized clutter statistics, hence it

should be more effective compared to global threshold. Next, as each algorithm just

needs to scan the whole image once, a lot of computation time is saved and makes this

process more and more towards real time.

As CS-SVD and TF-SVD performs satisfactorily in offline train mode, the

eagerness was to see their performance in self train mode. Figure 4.14 displays that. As

expected, it can be seen, the local self train mode outperforms the global self train

algorithms. The performance of EN-SVD in local self train mode does not cost any extra

gain over global self-train mode and same as the case for CS-SVD, where their

performance are almost comparable. But the noticable improvement observed, if we see

the performance curves of TF-SVD in local self train mode comapred to that in global

131
self train mode. At probability of false alarm of 0.05, the probability of detection of TF-

SVD in global self train mode and local self train mode is 0.8 and 0.88 respectively. The

TF-SVD algorithm in local self training mode provides 100% target detection at a small

probability of false alarm of 0.205, whereas to acheive 100% target detection

performance the TF-SVD algorithm-global threshold has a probability of misclassify of

about 80.5%. Hence we propose to implement the TF-SVD with local threshold in SAR

target detection scheme operates in self-train mode.

Fig. 4.15: Performance comparison of EN-SVD, CS-SVD and TF-SVD in self-train mode (here x –
axis in logarithmic scale)

132
4.5.4.3: Performance Comparison of various techniques

Fig. 4.16: ROC Performance comparison of various techniques

133
Fig. 4.17: ROC Performance comparison of various techniques (x axis is in logarithmic scale)

In this section, we compare the performance of EN-SVD, CS-SVD and TF-SVD in

offline train mode and self train mode together in a single picture. Figure 4.16 displays

that. In addition to that there are median filtering techniques combined with CS-SVD and

TF-SVD cases i.e. CS-SVD and TF-SVD output images are meadian filtered before

scoring the targets and clutter to form ROC curves. Also only the median filter output is

134
also shown i.e. the test image is median filtered and then target detection algorithm ran

onto it to generate ROC curves. In addition, a different image segmented size m=5 is

chosen in CS-SVD offline case to verify that the performance of CS-SVD algorithm will

be optimum at m=15. It is to be noted that median filtering is not implemented with none

of the algorithms that are running in self training mode. This is intentionally done to keep

the overhead minimum as the self training mode runs in real time and median filtering

slows the system. Another reason for this is median filtering hybrid with one of the

proposed algorithm doesn’t significantly improve the target detection performance

compared to that algorithm alone. This can be verified if we observe the target detection

performance in CS-SVD offline case with that of CS-SVD + median offline case. Same is

true for TF-SVD also. It can be verified from the above plots that the proposed CS-SVD

and TF-SVD significantly improve the target detection performance at low false alarm

rate compared to median filtering used for target detection from SAR image.

4.5.4.4: Conclusion

In summary, this initial research in the area of “self-training” algorithms for UWB SAR

target detection in a severe clutter simulation environment indicates that there is potential

for modifying existing approaches to radar real-time adaptive processing by designing

pre-filtering stages that eliminate significant portions of hard clutter discretes associated

with many severe UWB radar environments. The simulation ROC curves and images

depict encouraging potential of self-learning algorithms if a prefiltering step is employed

before applying the baseline algorithm onto it. The Terrain Filtered SVD approach

provides significant performance improvement over contemporary approaches in self-

135
train mode. Possible scope for further research includes developing approaches that, for

example, combine time-frequency-based pre-filtering techniques with (in addition to

these SVD-based techniques) other existing adaptive processing techniques such as

eigenvector-based matched filtering, auto-regressive techniques, as well as other non-

linear filtering techniques.

136
5: SUMMARY AND FUTURE WORK

5.1: Summary

A new approach using High Range Resolution (HRR) profiles to recognize time critical

military targets is considered in the first part of the thesis. In the second part, a new time

recursive sequential multi-target updating procedure, which is helpful in joint tracking

with recognition, is proposed. In the last part, three new SAR target detection algorithms

are proposed, which also performs satisfactorily in self-train mode.

5.1.1: Hybrid ATR

In order to overcome the limitations of Synthetic Aperture Radar (SAR) image-based

ATR in recognition of moving targets, the HRR scheme, studied as part of this thesis, can

be used to obtain superior recognition rates. HRR-ATR also offers considerable

computational savings when compared with SAR-ATR scheme. Results described in this

thesis were based on MSTAR database.

This thesis includes details of how HRR profiles are created and pre-processed to

achieve superior ATR performance. For large number of target classes with significantly

large number of templates, considerable on-line processing may be necessary for

classification. As a trade-off, range profiles are divided into sectors.

137
Many commonly known estimation and detection algorithms for ATR work best

or are computationally simple for gaussian signals. HRR profiles being positive valued,

their density function tend to be Gamma or Rayleigh distributed. The power transform

(PT) operation is shown to transform the density function to more Gaussian that in turn is

shown to result in higher recognition rates for uncorrupted Observation data. But it was

observed that in case of noisy observation profiles, application of PT on detected noisy

HRR data leads to considerable deterioration in the ATR performance. Hence in this

thesis, the power transform operation is avoided while making HRR profiles

preprocessing. The HRR profiles are normalized prior to classification as normalization

was shown to be a pre-requisite for the use of Matched filter in the classification stage, to

provide consistently superior results.

In this thesis, a new ATR algorithm is proposed, which will combine the Eigen

Template based Matched Filtering (ETMF) technique with Hidden Markov Model

(HMM) and target classification is performed based on this hybrid technique. The

process of template formation and the classification technique in case of ETMF is

described in detail. The effectiveness of SVD for removal of clutter from a HRR profiles

are described next. It has been shown that Eigen filtering technique, i.e., reconstruction

after zeroing out small singular values, which correspond to noise or clutter, can be

highly effective removing clutter. Next, the target classification using HMM technique is

described in detail. Lastly, the theoretical approach is provided for the process of

combining between ETMF and HMM. The experimental results are provided in Forced

detection case as well as in unknown target scenario.

138
5.1.2: Time Recursive Sequential ATR

In the time recursive sequential ATR technique, target decision is not based on a single

profile. Instead the maximum matched correlations for each target are saved and

combined later with those for other profiles coming in sequence. The target decision is

based on the combined hypothesis score. The main characteristics for this sequential

multiple hypothesis technique is the hypothesis score for range profiles is always updated

in a time recursive manner, hence the target decision made at time t is dependent on all

the hypothesis score from initial upto time t. The theoretical approach to develop this

algorithm is presented along with simulation results comparing single profile hypothesis.

5.1.3: SAR target detection algorithms

An ultra-wideband (UWB) synthetic aperture radar (SAR) simulation technique

that employs physical and statistical models is developed and presented. The three

proposed SAR target detection algorithms are named as EN-SVD, CS-SVD and TF-SVD.

Firstly, the developed SAR target detection algorithms are trained in offline mode to

learn the clutter statistics. But as it is not always possible to learn beforehand, approaches

towards developing “self-training” algorithms for UWB radar target detection are also

investigated using the results of this simulation process. These adaptive approaches

employ some form of modified Singular Value Decomposition (SVD) algorithm where

small blocks of data in the neighborhood of a sliding test window are processed in real-

time to estimate clutter characteristics. These real-time clutter models are then used to

cancel clutter in the sliding test window. In this thesis, the theoretical approach to

139
develop the algorithm is described first and then the algorithm is analyzed with

simulation results.

5.2 Future Work

The algorithms proposed in this thesis are made with the objective to achieve high

ATR/Detection for moving targets using minimum processing time. Recent advances in

superior sensor technologies and sensor simulation tools that allow wider classes of target

scenarios available at the ATR developmental stage, higher resolution imaging based on

super-resolution techniques, increasingly faster and superior computing hardware, and

appropriate advanced ATR strategies are all expected to be beneficial for achieving

improved performance from the evolving ATR methedologies.

Further data compression using Wavelet Transform would lead to reduction in

processing time. Due to the transform it is possible that the energy of some targets may

be distributed among more than the largest singular value and more than one eigenvector

may be more appropriate for template formation for some targets to achieve reasonable

ATR results.

In target detection case, further research includes developing approaches that, for

example, combine time-frequency-based pre-filtering techniques with (in addition to

these SVD-based techniques) other existing adaptive processing techniques such as

eigenvector-based matched filtering, auto-regressive techniques, as well as other non-

linear filtering techniques.

140
BIBILIOGRAPHY

[1] Li, H.J., and Yang, S.H. (1993), “Using range profiles as feature vectors to identify

aerospace objects”, IEEE Transactions on Antennas and propagation, 41, 3 (1993), 261-

268.

[2] Hudson, S., and Psaltis, D. (1993), “Correlation Filters for aircraft identification from

radar range profiles”, IEEE Transactions on Aerospace and Electronic Systems, 29, 3

(1993), 741-748.

[3] Mitchell R.A. and Westerkamp J.J. (1999), “Robust Statistical feature based aircraft

identification”, IEEE Transactions on Aerospace and Electronic Systems, 35, 3 (1999),

1077-1094.

[4] Chiang, H-C., Moses, R.L. and Potter, L.C. (2000), “Model based classification of

radar images”, IEEE Transactions on Information Theory, 46, 5 (2000), 1842-1854.

[5] Novak, L.M., Owirka, G.J. and Weaver, A.L (1999), “Automatic Target Recognition

using enhanced resolution SAR data”, IEEE Transactions on Aerospace and Electronic

systems, 35(1), pp. 157-175, 1999.

141
[6] Mirkin M. and Hodges, B. (1997), Private conversation at MIT Lincoln Laboratory,

1997 (according to Nguyen D., Benitz G., Kay J., Orchard B. and Whiting R.).

[7] Nguyen, D., Benitz, G., Kay, J., Orchard, B. and Whiting R (2001), “Superresolution

HRR ATR with High definition Vector Imaging”, IEEE Transactions on Aerospace and

Electronic Systems, 37(4), 1267-1286, 2001.

[8] Liao, X., Runkle, P., and Carin, L. (2002), “Identification of Ground targets From

Sequential High Range Resolution Radar Signatures”, IEEE Transactions on Aerospace

and Electronic systems, 38(4), pp. 1230-1241, 2002.

[9] Mitchell, R.A. and Westerkamp, J.J. (1997), “A statistical Feature based Classifier

For Robust High range Resolution Radar target Identification”, Automatic Target

recognition Working Group, Huntsville, AL, Oct 1997.

[10] Westerkamp J. (1998), “A Framework for 1D HRR ATR evaluation”, AFRL

Technical Report (Draft), presented at the TRUMPETS Kick-Off, Aug. 1998.

[11] Shaw, A. K. and Bhatnagar, V. (1998), “Automatic Target Recognition using Eigen

Templates.” Synthetic Aperture radar Imagery V: Proceedings of SPIE, Orlando, FL,

448-459, April 1998.

142
[12] Shaw, A K. and Vashist, R. (2000), “HRR-ATR using Eigen-Templates and

Normalized Matched Filter as Classifier”, Synthetic Aperture radar Imagery V

Proceedings of SPIE, Orlando, FL, April 2000.

[13] Hawley, R.W. (1999), “Complex HRR signal generation for TRUMPETS”,

Trumpets Working Research Note, July 29, 1999.

[14] Augustyn, K. (1992), “A new approach to automatic target recognition”, IEEE

Transactions on Aerospace and Electronic Systems, 28(1): 105-114, January 1992.

[15] Dudgeon, D.W. and Lacoss, R.T. (1993), “An overview of automatic target

recognition”, The Lincoln Laboratory Journal, 6(1), 1993.

[16] Novak, L.M., Halversen, S.D., Owirka, G.J. and Heitt, M. (1995), “Effects of

polarization and resolution on the performance of a SAR automatic target recognition

system”, The Lincoln Laboratory Journal, 8(1), 1995.

[17] Verly, J.G., Delanoy R.L. (1996), “Model-based automatic target recognition (ATR)

systems for forward looking ground based and airborne imaging laser radars”

Proceedings of the IEEE, 84(2), February 1996.

[18] Curlander, J.C. and McDonough, R.N. (1991), “Synthetic Aperture Radar”, John

Wiley & Sons, Inc., 1991.

143
[19] Goj, W.W. (1993), “Synthetic Aperture Radar and Electronic Warfare”, Artec

House, MA, 1993.

[20] Kovaly, J.J. (1976), “Synthetic Aperture Radar”, Artech House, 1976.

[21] Scheler, D.C. (1991), “MTI and Pulsed Doppler Radar”, Artech House, MA, 1991.

[22] Wehner, D.R. (1987), “High Resolution Radar”, Artech House, MA, 1987.

[23] Scheler, D.C. (1980), “MTI Radar”, Artech House, INC., 1980.

[24] Scheler, D.C. (1991), “MTI and Pulsed Doppler Radar”, AIL Systems, INC., 1991.

[25] Povejsil, D., Raven, R. And Waterman, R. (1961), “Airborne Radar”, D. Van

Nostrand Princeton, NJ, 1961.

[26] Stimson, G. (1983), “Introduction to Airborne Radar”, Hughes Aircraft Company,

ElSegundo, CA, 1983.

[27] Kovaly, J.J. (1976), “Synthetic Aperture Radar”, Artech House, 1976.

[28] Hovanessian, S.A. (1980), “Introduction to Synthetic Array and Imaging Radars”.

Artech House, 1980.

144
[29] Anderson, J.A., Mohanty, N.C. and Bhattacharya, P.K. (1995), “Model for detection

& tracking of multiple targets”, SPIE PROCEEDINDS, 2984(4): 147-153, April 1995.

[30] Morgan, D.R. (1995), “Automatic target recognition via classical detection theory”,

SPIE PROCEEDINGS, 2484(4): 494-500, April 1995.

[31] Jacobs, S.P. and O'Sullivan, J.A. (1997), “High resolution radar models for joint

tracking and recognition”, PROCEEDINDS OF SPIE, 3069(7): 94-107, April 1997.

[32] Xun, Z., Zhaowen, Z. and Guirong, G. (1997), “Automatic HRR target recognition

based on matrix pencil method and multiresolution neural network. PROCEEDINGS OF

SPIE, 3069(7): 510-517, April 1997.

[33] Yiding, W., Xiangguo, L. and Yunhong, W. (1997), “Recognition of high resolution

ground targets”, PROCEEDINGS OF SPIE, 3069(7): 478-485, April 1997.

[34] Dudgeon, D.E., Grimson, W. and Tenney, R. (1995), “Issues in SAR model-based

target recognition”, SPIE PROCEEDINGS, 2484(4): 501-511, April 1995.

[35] Morgan, D.R. (1995), “Automatic target recognition via classical detection theory”,

SPIE PROCEEDINGS, 2484(4): 494-500, April 1995.

145
[36] Verly, J.G., Delanoy, R.L. and Lazott, C.H. (1993), “Principles and evaluation of an

automatic target recognition system for synthetic aperture radar imagery based on the use

of functional templates”, SPIE PROCEEDINGS, 1960(3): 57-73, April 1993.

[37] Daubechies, I. (1988), “Orthogonal bases of compactly supported wavelets”, Comm.

Pure and Applied Mathematics, 41: 909-996, 1988.

[38] Mahalanobis, A. (1995), “Automatic target recognition using correlation filters”,

SPIE Proceedings 2484(4): 546-557, April 1995.

[39] Chui, C.K. (1992), “An Introduction to Wavelets”, Academic Press, CA, 1992.

[40] Editor Chui, C.K. (1992), “Wavelets: A tutorial in Theory and Applications”,

Academic Press, CA, 1992.

[41] Daubechies, I. (1990), “The wavelet transform, time frequency localization and

signal analysis”, IEEE Transactions on Information Theory, 36: 991-1005, 1990.

[42] Antonini, M., Barlaud, M. and Mathieu, P. (1990) “Image coding using vector

quantization in the wavelet transform domain”, Proc ICASSP-90, Albuquerque, NM,

pages 2297-2300, 1990.

146
[43] Mallat, S. (1989), “Multifrequency channel decomposition of images and wavelet

models”, IEEE Transactions on Acoustics Speech and Signal Processing, 37: 2091-2110,

1989.

[44] Mallat, S. (1989), “A theory of multiresolution signal decomposition: The wavelet

representation”, IEEE Trans. on Pattern Anal. And Machine Intell, 11: 674-693, 1989.

[45] Guo, H., Odegard, J.E., Lang, M., Gopinath, R.A., Selesnick, I.W. and Burrus, C.S.

(1994), “Wavelet based speckle reduction with applications to SAR based ATR”, Proc.,

ICIP, Austin, TX, November 1994.

[46] Pederson, J.F., Saff, E.B. and Schweiker, K. (1993), “Wavelets for training a neural

network to recognize objects”, SPIE PROCEEDINGS, 1960(3): 244-254, April 1993.

[47] Tagliarini, G.A., Page, E.W., Kesden, G.M., Chiang, D. and McPartland P.J. (1997),

“Application of wavelet and neural processing to automatic target recognition”

PROCEEDINGS OF SPIE, 3069(7): 154-161, April 1997.

[48] Bottcher, C. and Strayer, M.R. (1993), “Classical scattering theory of waves from

the view point of an eigenvalue problem and application to target identification”, SPIE

PROCEEDIMGS, 1960(3): 425-435, April 1993.

147
[49] Paul, D.B. (1990), “Speech recognition using hidden markov models”, The Lincoln

Laboratory Journal, 3(1), 1990.

[50] Rabiner, L.R., Levinson, S.E. and Sondhi, M.M. (1983), “An Introduction to the

application of the theory of probabilistic functions of a markov process to automatic

speech recognition”, Bell Systems Technical Journal, 62: 1035, 1983.

[51] Flake, L., Krishnamurthy, A. and Ahalt, S. (1995), “Multi-aperture SAR target

detection using hidden markov model”, SPANN Technical Report TR-95-02, The Ohio

State University, 1995.

[52] Stroud, P.D. and Gordon, R.C. (1997), “Automated military unit identification in

battlefield simulation”, PROCEEDINGS OF SPIE, 3069(7): 375-386, April 1997.

[53] Rao, K. (1992), “Combinatronics reduction for target recognition in ATR

applications”, SPIE PROCEEDINGS, 17002(2): 285-295, April 1992.

[54] Jacobs, S.B., O'Sullivan, J.A. and Bekker, Z. (1997), “Recognition of ground targets

from high resolution radar range-profiles”, To Appear, 1997.

[55] Jacobs, S.P., O'Sullivan, J.A., Faisal, M. and Snyder, D.L. (1996), “Automatic target

recognition using high resolution radar”, SPIE Proceedings, The International Society of

Optical Engineering, 2845, August 1996.

148
[56] Worrell, S. (1996), “Air to ground automatic target recognition”, An unclassified

report, 1996.

[57] Stewart, C., Larson, V. and Halsey, J.D. (1992), “Comparison of classification

approaches for high-range resolution radar”, SPIE PROCEEDINGS, 1700(2): 146-155,

April 1992.

[58] Bhatnagar, V., Shaw, A.K. and Williams, R. (1998), “Improved Automatic Target

Recognition using Singular Value Decomposition”, IEEE International Conference on

Acoustics, Speech and Signal Processing, Seattle, WA, vol. 5, pp. 2717-2720, May 1998.

[59] Bhatnagar V. (1998), Automatic Target Recognition using High Range Resolution

Data, MS Thesis, Electrical Engineering Department, Wright State University, 1998.

[60] Forney, G.D., “The Viterbi algorithm”, Proc. IEEE, 1973, 61, 268-278.

[61] Linde, Y., Bujo, A. And Gray R., “An algorithm for vector quantizer design ”, IEEE

Transactions on Communication, COM-28, 1980, 84-98.

[62] Rabiner, L.R., “ A tutorial on hidden Markov models and selected applications in

speech recognition”, Proceedings of the IEEE, 1989. 77(2), 257-286.

149
[63] Do M.N. and Wagner M. (1998), “Speaker recognition with small training

requirements using a combination of VQ and DHMM”, Proc. of Speaker Recognition and

its Commercial and Forensic applications, Avignon, France, 1998, 169-171.

[64] Mitra, A. K., Lewis, T. L., Paul, A. S. and Shaw A. K. (2002), “Ultra-Wideband

Radar Data Models and Target Detection with Adaptive Rank Order Filters”, Synthetic

Aperture Radar Imagery V: Proceedings of SPIE, Orlando, FL, April 2002.

[65] Mitra, A.K. and Lewis, T.M. (2002), “Spectral Analysis of high-frequency SAR

data”, Synthetic Aperture Radar Imagery V: Proceedings of SPIE, Orlando, FL, April

2002.

[66] Chiang, H.C., Moses, R. L. and Potter, L.C. (2000): “Model-based Classification of

Radar Images,” IEEE Transactions on Information Theory, vol. 46, no. 5, pp. 1842 –

1854, August 2000.

[67] Delignon, Y. and Pieczynki, W. (2002) “Modeling Non-Rayleigh Speckle

Distribution in SAR Images,” IEEE Transactions on Geoscience and Remote Sensing,

vol. 40, no. 6, June 2002.

[68] Soumekh, M. (1999): “Synthetic Aperture Radar Signal Processing with MATLAB

Algorithms”, Wiley Interscience, 1999.

150
[69] Davis, M., Tomlinson, P. and Paul, M., “Technical Challenges in Ultra-Wideband

Radar Development for Target Detection and Terrain Mapping”, Proceedings of the

IEEE, National Radar Symposium, 1999.

[70] Melvin, W.L. (1998): “EigenBased Modeling of Nonhomogeneous Airborne Radar

Environments”, IEEE National Radar Conference, 1998, Dallas, TX, 12-13 May 1998.

151

You might also like