You are on page 1of 136

Announcing a new Special Issue:

MICROWAVES IN CLIMATE CHANGE


Initial Manuscript Submission Deadline: July 1st, 2024
Based on feedback from our special series article, “Making Waves: Microwaves in Climate Change,” the IEEE
Journal of Microwave HYPERLINK "https://mtt.org/publications/journal-of-microwaves/"s is putting
together a full special issue on this extremely prescient topic area. Microwave devices, instruments, systems,
measurements, applications, and data analysis provide enabling technology and science retrieval in areas related
to climate tracking, atmospheric chemistry and evolution, alternative energy development, efficient generation
and usage of non-fossil-based fuels, waste conversion, electrification, transportation management, and every
sector of the society relying on communications.

We are soliciting articles on the following topics as well as entertaining alternative suggestions from
interested contributors:

• Active and Passive Microwave Remote Sensing


• Microwave Heating
• Microwave Generation of Alternative Fuels and Catalysts
• Microwave Power Beaming
• Microwave Energy Harvesting
• Microwaves in Fusion and Energy Generation
• Low-Loss Microwave Transmission
• Microwaves in Waste Management
• Microwave Assisted Chemistry
• Microwaves in Geophysics
• Microwave Data Analysis for Climate Research
• Microwave Tracking for Habitat Assessment and Animal Science
• Microwave Resource Monitoring
• Other topics related to climate science and relevant resource monitoring
If you have a particular idea or topic you would like to contribute, please contact Peter Siegel, JMW Editor-in-
Chief (phs@caltech.edu).

The deadline for submission of initial manuscripts is Monday, July 1st, 2024. Final upload of accepted paper
proofs is September 1st, 2024. All production ready manuscripts will be posted on IEEE Xplore Early Access and
will appear in final paginated sequence in the Special Issue – scheduled for final release in October 2024.
Contributed papers should be targeted at ten pages but review and special invited papers can be longer. All
submissions will be reviewed in accordance with the normal procedures of the journal. Please tag uploaded
papers as “Special Issue” through our Author Portal.

We hope you will consider contributing to this Special Issue of IEEE Journal of Microwaves and continue to
support the journal through your regular research submissions.

Digital Object Identifier 10.1109/MGRS.2023.3306031


SEPTEMBER 2023
VOLUME 11, NUMBER 3
WWW.GRSS-IEEE.ORG

FEATURES

8 The Orbital X-Band Real-Aperture


Side-Looking Radar of Cosmos-1500
by Ganna B. Veselovska-Maiboroda, Sergey A. Velichko,
and Alexander I. Nosich


21 Airborne Lidar Data Artifacts
by Wai Yeung Yan

OPENER CREDIT: LAVA AND SMOKE BLANKET FAGRADALSFJALL IN ICELAND,


JULY 2023 © NASA
46 Interferometric Phase Linking
by Dinh Ho Tong Minh and Stefano Tebaldini

63 There Are No Data Like More Data


by Michael Schmitt, Seyed Ali Ahmadi, Yonghao Xu,
Gülşen Taşkin, Ujjwal Verma, Francescopaolo Sica,
ON THE COVER:
and Ronny Hänsch
Fig. S1 from Schmitt et al., page 63 in this issue, showing a schematic
illustration of the size measure used to characterize Earth Observa-
tion datasets.
BACKGROUND IMAGE LICENSED BY INGRAM PUBLISHING

SCOPE
IEEE Geoscience and Remote Sensing Magazine (GRSM) will
inform readers of activities in the IEEE Geoscience and
Remote Sensing Society, its technical committees,
and chapters. GRSM will also inform and educate
readers via technical papers, provide information on
international remote sensing activities and new satellite
missions, publish contributions on education activities,
industrial and university profiles, conference news, book
reviews, and a calendar of important events.

Digital Object Identifier 10.1109/MGRS.2023.3304525

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 1


IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE
EDITORIAL BOARD Vice President of Professional Activities
Dr. Lorenzo Bruzzone
Dr. Paolo Gamba University of Trento, Italy
COLUMNS & University of Pavia
Department of Electrical, Biomedical, Vice President of Meetings and Symposia
DEPARTMENTS and Computer Engineering
Pavia, Italy
Sidharth Misra
NASA-JPL, USA
paolo.gamba@unipv.it
Vice President of Technical Activities
4 FROM THE EDITOR Dr. Fabio Pacifici
Subit Chakrabarti (Cloud to Street, USA) Maxar, USA
6 PRESIDENT’S MESSAGE Gong Cheng (Northwestern Polytechnical Secretary
University, P.R. China) Dr. Steven C. Reising
98 SOFTWARE AND DATA SETS Colorado State University, USA
Michael Inggs (University of Cape Town,
Chief Financial Officer
114 CONFERENCE REPORTS South Africa) Dr. John Kerekes
Rochester Institute of Technology, USA
126 TECHNICAL COMMITTEES George Komar (NASA retired, USA)
IEEE PUBLISHING OPERATIONS
Joseé Levesque (Defence Research and Journals Production Manager
Development, Canada) Sara T. Scudder

Andrea Marinoni (UiT, Artic University Senior Manager, Production


Katie Sullivan
of Norway, Norway)
Senior Art Director
Fabio Pacifici (Maxar, USA) Janet Dudar
Associate Art Director
Mario Parente (University of Massachusetts, USA) Gail A. Schnitzer
Nirav N. Patel (Defence Innovation Unit, USA) Production Coordinator
Theresa L. Smith
Michael Schmitt (Universität der Bundeswehr, Director, Business Development–
Germany) Media & Advertising
Mark David
Vicky Vanthof (Univ. of Waterloo, Canada) +1 732 465 6473
m.david@ieee.org
Fax: +1 732 981 1855
Hanwen Yu (University of Electronic Science
and Technology of China, P.R. China) Advertising Production Manager
Felicia Spagnoli
GRS OFFICERS Production Director
Peter M. Tuohy
President
Mariko Sofie Burgin Director, Editorial Services
NASA Jet Propulsion Laboratory, USA Kevin Lisankie
Executive Vice President Senior Director, Publishing Operations
MISSION STATEMENT Saibun Tjuatja Dawn M. Melley
The IEEE Geoscience and Remote Sensing Soci- The University of Texas at Arlington, USA
ety of the IEEE seeks to advance science and Vice President of Publications
technology in geoscience, remote sensing and Alejandro C Frery
Victoria University of Wellington, NZ
related fields using conferences, education, and
other resources. Vice President of Information Resources
Keely L Roth
Salt Lake Cty UT, USA

IEEE Geoscience and Remote Sensing Magazine (ISSN 2473-2397) is published provided the per-copy fee indicated in the code is paid through the Copyright Clear-
quarterly by The Institute of Electrical and Electronics Engineers, Inc., IEEE ance Center, 222 Rosewood Drive, Danvers, MA 01923 USA; 2) pre-1978 articles
Headquarters: 3 Park Ave., 17th Floor, New York, NY 10016-5997, +1 212 419 without fee. For all other copying, reprint, or republication information, write to:
7900. Responsibility for the contents rests upon the authors and not upon the Copyrights and Permission Department, IEEE Publishing Services, 445 Hoes Lane,
IEEE, the Society, or its members. IEEE Service Center (for orders, subscrip- Piscataway, NJ 08854 USA. Copyright © 2023 by the Institute of Electrical and Elec-
tions, address changes): 445 Hoes Lane, Piscataway, NJ 08854, +1 732 981 tronics Engineers, Inc. All rights reserved. Application to Mail at Periodicals Postage
0060. Individual copies: IEEE members US$20.00 (first copy only), nonmem- Prices is pending at New York, New York, and at additional mailing offices. Cana-
bers US$110.00 per copy. Subscription rates: included in Society fee for each dian GST #125634188. Canada Post Corporation (Canadian distribution) publica-
member of the IEEE Geoscience and Remote Sensing Society. Nonmember tions mail agreement number 40013885. Return undeliverable Canadian addresses
subscription prices available on request. Copyright and Reprint Permissions: to PO Box 122, Niagara Falls, ON L2E 6S8 Canada. Printed in USA.
Abstracting is permitted with credit to the source. Libraries are permitted to
photocopy beyond the limits of U.S. Copyright Law for private use of patrons: IEEE prohibits discrimination, harassment, and bullying. For more information,
1) those post-1977 articles that carry a code at the bottom of the first page, visit http://www.ieee.org/web/aboutus/whatis/policies/p9-26.html.

Digital Object Identifier 10.1109/MGRS.2023.3301259

2 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Harness the
publishing power
of IEEE Access. ®

IEEE Access is a multidisciplinary open


access journal offering high-quality peer
review, with an expedited, binary review
process of 4 to 6 weeks. As a journal
published by IEEE, IEEE Access offers
a trusted solution for authors like you
to gain maximum exposure for your
important research.

Explore the many benefits of IEEE Access:


• Receive high-quality, rigorous peer review • Establish yourself as an industry pioneer
in only 4 to 6 weeks by contributing to trending, interdisciplinary
topics in one of the many topical sections
• Reach millions of global users through IEEE Access hosts
the IEEE Xplore® digital library by publishing
open access • Present your research to the world quickly
since technological advancement is ever-changing
• Submit multidisciplinary articles that may not
fit in narrowly focused journals • Take advantage of features such as multimedia
integration, usage and citation tracking, and more
• Obtain detailed feedback on your research
from highly experienced editors • Publish without a page limit for $1,750
per article

Learn more at ieeeaccess.ieee.org


FROM THE EDITOR
BY PAOLO GAMBA

Issues and Special Issues

I n line with what I did for the June issue, I will use my
IEEE Geoscience and Remote Sensing Magazine (GRSM)
editorial on one hand to summarize the content of the
operating phase. Still considering radar systems but
focusing on Interferometric Synthetic Aperture Radar
(InSAR), [A2] reviews the phase linking algorithms that
current issue and, on the other hand, to introduce a fea- have been proposed with respect to their efficiency, pre-
ture of this magazine that may not be well known to cision, and usefulness with respect to final applications.
(or understood by) all our readers. Specifically, I will de- As a preprocessing tool for various SAR applications,
scribe the possibility to publish special issues in GRSM. these algorithms well match the topic of the following
As mentioned in my first editorial this year, GRSM ac- article [A3], which focuses on the many different types
cepts proposals for special issues. In June I spent some of artifacts that affect light detection and ranging (lidar)
time to explain why articles should be submitted in a datasets. The comprehensive analysis of these artifacts
white paper format; here I will go a little bit into explain- and the most common techniques to reduce them is a
ing how white papers and special issues go together and very good introduction of preprocessing techniques ap-
what types of special issues are welcome in GRSM. plied to lidar data. Finally, the last technical article in
the issue is devoted to datasets coming from multiple
IEEE GEOSCIENCE AND REMOTE SENSING sources and their relationship with deep learning tech-
MAGAZINE ISSUE CONTENT niques [A4]. Indeed, Earth observation (EO) datasets
Let us first start, however, by describing the contents of are very different from what is usually considered for
this issue, which is full of valuable technical articles, algorithm development in computer vision, and a better
as well as interesting and timely columns providing in- knowledge of the datasets that are available and their
sights about what happens in our Society. Indeed, this features is a way to strengthen the link between the
issue includes four technical articles and five columns. GRSS and the computer vision community without los-
As you have seen, I am pushing to have more columns ing the link with the physics at the very basis of EO data.
than in the past years because columns are meant to The columns are actually well connected to the
provide information about the activities of the Society fourth technical article by means of the first two of
and are very valuable for IEEE Geoscience and Remote them because they are reports about new datasets that
Sensing Society (GRSS) members. Columns have both are available, one for deep learning and the other one for
an informative and an engaging purpose: readers are classification purposes. The first column [A5] provides
informed about what happens in the various GRSS details about a large-scale, global, multimodal, and mul-
committees and working groups and find out how they tiseasonal corpus of satellite imagery from multispectral
could be involved, if interested. and radar sensors. The second column [A6] introduces
The technical content of this issue encompasses an open dataset that tackles the burned area delineation
hardware and software and refers to multiple sensors problem using prefire and postfire Sentinel-2 acquisi-
and systems. The first article [A1] introduces the struc- tions of California forest fires that took place starting in
ture, history, and performance of the Cosmos-1500 real 2015. The following two columns refer to activities per-
aperture radar, with an interesting description of the formed in the framework of the recent 2023 IEEE Inter-
challenges faced by its developers in the designing and national Geoscience and Remote Sensing Symposium.
In [A7], the interested reader will find the highlights of
Digital Object Identifier 10.1109/MGRS.2023.3304503
the awards and opening sessions during the first day of
Date of current version: 19 September 2023 the conference. The next column [A8] summarizes the

4 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


activities related to diversity, equity, and inclusion that style of IEEE publications is available. In the last portion of
have been promoted and organized by the GRSS Inspire, the proposal, the guest editors should provide a tentative
Develop, Empower, Advance Committee right before and schedule for all the major deadlines involved in the spe-
during the conference. Finally, the last column [A9] reports cial issue because according to GRSM policy, submissions
about a session organized to explore the topic of data sci- will be articulated in a two-step procedure: First, the guest
ence at scale, which was part of the summer school, “High editors should solicit a short white paper from the invited
Performance and Disruptive Computing in Remote Sens- authors (four to five pages in double-column format). The
ing,” hosted at the University of Iceland last May. white paper will summarize the foreseen objectives of the
article and discuss the importance of the addressed topic,
IEEE GEOSCIENCE AND REMOTE SENSING the impact of the contribution, and the authors’ expertise
MAGAZINE AND SPECIAL ISSUES and past activities on the topic. Based on the white papers,
In this second part of my editorial, I will introduce GRSM spe- the guest editors will select the contributions to be sub-
cial issues, which are collections of technical articles (neither mitted as full articles. The full articles will then be peer-
reviews nor tutorials) that are published in GRSM after a rig- reviewed by international experts, and those accepted for
orous selection and peer review evaluation procedure based publication will be included in the special issue.
on a specific topic that has been suggested by a few guest edi-
tors and that has been accepted by the GRSM editorial board. APPENDIX: RELATED ARTICLES
To match the scope of a magazine like GRSM, each of [A1] G. B. Veselovska-Maiboroda, S. A. Velichko, and A. I. Nosich,
such collections cannot be composed by articles showing “The orbital X-band real-aperture side-looking radar of Cos-
specific technical results in the area of the special issue. In- mos-1500: A Ukrainian IEEE Milestone candidate,” IEEE Geos-
stead, special issue articles should be technical articles sum- ci. Remote Sens. Mag., vol. 11, no. 3, pp. 8–20, Sep. 2023, doi:
marizing the achievements, the statuses, and the challenges 10.1109/MGRS.2023.3294708.
of the technical activities connected with one of the aspects [A2] D. Ho Tong Minh and S. Tebaldini, “Interferometric phase
of the aforementioned topic. In the last four years, GRSM linking: Algorithm, application, and perspective,” IEEE Geosci.
has published one special issue only, on the topic of hyper- Remote Sens. Mag., vol. 11, no. 3, pp. 46–62, Sep. 2023, doi:
spectral imaging, and the interested reader is referred to that 10.1109/MGRS.2023.3300974.
issue to better understand the right blend of technical nov- [A3] W. Y. Yan, “Airborne lidar data artifacts: What we know thus
elty, deep coverage, and clear explanation that the articles far,” IEEE Geosci. Remote Sens. Mag., vol. 11, no. 3, pp. 21–45,
in that issue provided. GRSM has not published any special Sep. 2023, doi: 10.1109/MGRS.2023.3285261.
issues since then for the past five and a half years, and we [A4] M. Schmitt et al., “There are no data like more data: Datasets
do not expect to be able to publish one in 2023, but we are for deep learning in Earth observation,” IEEE Geosci. Remote
already planning for one or two special issues in 2024. Sens. Mag., vol. 11, no. 3, pp. 63–97, Sep. 2023, doi: 10.1109/
Keeping the previous considerations in mind, interested MGRS.2023.3293459.
guest editors who think they have a valuable proposal for [A5] Y. Wang, N. A. A. Braham, Z. Xiong, C. Liu, C. M. Albrecht,
a special issue of GRSM are welcome to submit a proposal and X. X. Zhu, “SSL4EO-S12: A large-scale multimodal, multi-
using the template, which is available at https://www.grss- temporal dataset for self-supervised learning in Earth observa-
ieee.org/wp-content/uploads/2023/06/GRSM-TEMPLATE- tion,” IEEE Geosci. Remote Sens. Mag., vol. 11, no. 3, pp. 98–106,
FOR-SPECIAL-ISSUE-PROPOSALS.pdf. Such a proposal Sep. 2023, doi: 10.1109/MGRS.2023.3281651.
should introduce the overall topic of the special issue and [A6] D. R. Cambrin, L. Colomba, and P. Garza, “CaBuAr: Califor-
articulate the intended content, applications, and style of nia burned areas dataset for delineation,” IEEE Geosci. Remote
the contributions. The guest editors should also describe Sens. Mag., vol. 11, no. 3, pp. 106–113, Sep. 2023, doi: 10.1109/
potential subdivisions of the contributions into separate MGRS.2023.3292467.
sessions. Subsequently, the main motivation of the pro- [A7] A. Moreira, F. Bovolo, D. Long, and A. Plaza, “IGARSS 2023 in
posed special issue should be described in terms of the Pasadena, California: Impressions of the first days,” IEEE Geos-
timeliness and scientific/technical relevance of the pro- ci. Remote Sens. Mag., vol. 11, no. 3, pp. 114–125, Sep. 2023, doi:
posed topic and its expected interest for GRSM readers (i.e., 10.1109/MGRS.2023.3303685.
the whole GRSS membership). A special issue proposal [A8] V. Vanthof, H. Mcnairn, S. Tumampos, and M. Burgin, “Rein-
submission should also include a clear description of the forcing our commitment: Why DEI matters for the IEEE GRSS,”
relevance of the proposal to geoscience and remote sensing IEEE Geosci. Remote Sens. Mag., vol. 11, no. 3, pp. 126–129, Sep.
and to the scope of GRSM. In agreement with the approach 2023, doi: 10.1109/MGRS.2023.3303874.
and style of the magazine, the guest editor should strive to [A9] M. Maskey et al., “A summer school session on mastering geo-
create a balanced mix between ensuring scientific depth spatial artificial intelligence: From data production to artificial
and dissemination to a wide public, which would encom- intelligence foundation model development and downstream
pass remote sensing scientists, practitioners, and students. applications,” IEEE Geosci. Remote Sens. Mag., vol. 11, no. 3, pp.
Finally, the names of the proposed guest editor team must 129–132, Sep. 2023, doi: 10.1109/MGRS.2023.3302813.
be included, so that for each member, a biosketch in the GRS

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 5


PRESIDENT’S MESSAGE
BY MARIKO BURGIN

Letter From the President

H ello again! My name is Mariko Burgin, and I am


the IEEE Geoscience and Remote Sensing Society
(GRSS) president. You can reach me at president@grss
can be organized within local chapters or confer-
ences. The Sister Societies agreed to jointly work to
recruit new members and encourage current mem-
-ieee.org and @GRSS_President on X, formally known bers to become members of both Sister Societies. The
as Twitter. Sister Societies will explore the possibility of a joint
I am delighted to report that on 17 June 2023, the distinguished lecturer program and jointly sponsored
GRSS signed two separate memorandums of under- special issues programs and also have agreed to pro-
standing (MoUs) with two of its IEEE Sister Societies. mote each other’s community of conferences, meet-
I signed an MoU with Dr. Stefano Maci, president of ings, and trade shows. This will enhance visibility
the IEEE Antennas and Propagation Society (AP-S) within the relevant fields and allow members access
(see Figure 1), and another MoU with Dr. Nuno Borg- to events exclusively organized by the signatory Sister
es Carvalho, president of the IEEE Microwave Theory Society. The Sister Societies will be able to offer an
and Technology Society (MTT-S) (see Figure 2). array of promising prospects for each Society’s mem-
Through these MoUs, the IEEE Societies can facili- bers. Aside from the wealth of expertise, resources,
tate knowledge exchange and enable the dissemina- and networks that will be shared through these col-
tion of best practices, research findings, and industry laborative agreements, I am optimistic that new ideas
trends through joint technical or social meetings that and intersociety activities will flourish. These joint
partnerships will be a positive step toward synergistic
Digital Object Identifier 10.1109/MGRS.2023.3304090
growth and enrich each Society’s operations contrib-
Date of current version: 19 September 2023 uting to the overall success and impact of the GRSS

FIGURE 1. On 17 June 2023, the GRSS signed an MoU with FIGURE 2. On 17 June 2023, the GRSS signed an MoU with its
its IEEE Sister Society, the AP-S. On the left, Dr. Mariko Burgin, IEEE Sister Society, the MTT-S. On the left, Dr. Mariko Burgin,
president of the GRSS. On the right, Dr. Stefano Maci, president president of the GRSS. On the right, Dr. Nuno Borges Carvalho,
of the AP-S. president of the MTT-S.

6 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


together with the AP-S and the MTT-S. I am delighted to
renew our commitment to collaboration together with
the presidents and presidents-elect of the AP-S and the
MTT-S (see Figure 3).
I look forward to similar conversations around partner-
ship and collaboration with the IEEE Aerospace and Elec-
tronic Systems Society and the IEEE Oceanic Engineering
Society, as well as other Societies and institutions in the
remainder of 2023.
Would you like to be part of the next step on this jour-
ney, to turn these MoUs into actions? Reach out! We are
looking for enthusiastic volunteers! We need YOU to shape
FIGURE 3. Group photo of the Society presidents and presidents-elect your community!
of the GRSS, the MTT-S, and the AP-S on 18 June 2023, in Chicago, IL. Warmly,
From left to right, Dr. Saibun Tjuatja (GRSS president-elect), Dr. Mariko Mariko Burgin
Burgin (GRSS president), Dr. Nuno Borges Carvalho (MTT-S president), IEEE GRSS President
Dr. Stefano Maci (AP-S president), Dr. Maurizio Bozzi (MTT-S president-
elect), and Dr. Branislav Notaros (AP-S president-elect). GRS

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 7


The Orbital X-Band
Real-Aperture Side-Looking
Radar of Cosmos-1500
A Ukrainian IEEE Milestone candidate

GANNA B. VESELOVSKA-MAIBORODA, SERGEY A. VELICHKO ,


AND ALEXANDER I. NOSICH

W e revisit the development and operation of


the orbital X-band real-aperture side-looking
radar (R A-SLR) onboard the USSR satellite Cos-
mos-1500 in the historical context. This radar was
conceived, designed, and tested in the early 1980s
and then supervised, in orbit, by a team of Ukrai-
nian scientists and engineers led by Prof. Anatoly I.
Kalmykov (1936–1996) at the O. Y. Usikov Institute
of Radiophysics and Electronics (IRE) of the Nation-
al Academy of Sciences of Ukraine (NASU). It had a
magnetron source, a 12-m deployable slotted-wave-
guide antenna, and an onboard signal processing
unit. Instead of preplanned meticulous experi-
ments, only five days after placement into the polar
Earth orbit in the autumn of 1983, the SLR of Cos-
mos-1500 rendered truly outstanding service. It pro-
vided a stream of microwave images of the polar sea
ice conditions that enabled the rescue of freighters
in the Arctic Ocean. Two years later, similar imag-
ery was equally important in the rescue of a motor
vessel (MV) in the Antarctic. However, the way to
success was far from smooth. Besides the technical
problems, Kalmykov had to overcome the jealousy
and hostility of his home institute administration,
colleagues from Moscow research laboratories, and
high-level USSR bureaucracy. Later, Kalmykov’s ra-
dar was released to the industry and became the
main instrument of the USSR and Russian series of
remote sensing satellites Okean and Ukrainian satel-
lites Sich-1 and Sich-1M. We believe that the RA-SLR
of Cosmos-1500 is a good candidate for the status of
an IEEE Milestone in Ukraine.

Digital Object Identifier 10.1109/MGRS.2023.3294708


Date of current version: 25 July 2023

8 2473-2397/23©2023IEEE IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


INTRODUCTION radiometer, and transponders for collecting data from
In NASA’s Space Science Coordinated Archive, there is a ice and buoy transmitters.”
page devoted to the Earth satellite Cosmos-1500 launched Although the organizations and, in part, the people who
40 years ago in a country that does not exist anymore, conceived, designed, and built the main remote sensing in-
the USSR [1]. It communicates brief information on strument onboard Cosmos-1500, an X-band RA-SLR, are still
that mission. alive, the time is merciless, and the memory tends to turn
“The Cosmos 1500 spacecraft was a precursor to the humans’ experience into a legend. We would like to intro-
operational Russian Okean series of oceanographic re- duce the readers to the history of the creation and operation
mote sensing missions. The Cosmos 1500 tested new of Cosmos-1500. Many interesting details of that story can be
sensors and methods of data collection and processing. found in reviews [2], [3], [4], [5] and a book [6]. However,
Cosmos 1500 had the capability of overlapping and most of them have never been translated into English and
processing images from its sensors. Data from Cosmos remain unknown to international readers. Besides, the years
1500 were sent directly to ships or automated data re- that have passed since 1983 and the experience of the post-
ceiving stations and applied in navigation in northern USSR developments enable us to reveal important details
oceans. The instrument complement was highlighted that escaped earlier publications, ensure proper positioning
by an all-weather Side-Looking Real Aperture Ra- of that achievement, and add a “human dimension” to the
dar operating at 9.5 GHz. Other instruments included whole story. This article builds upon the preceding short con-
a multispectral ­scanner, a scanning high-frequency ference paper [7], which has been considerably extended.

THE THREE-HEADED DRAGON


OF USSR SCIENCE
In the USSR, science was a state-
owned dragon of three heads, tight-
ly controlled by the Communist
Party (CP), whose goals—tech-
nological efficiency and political
control—had always contradicted
each other [8]. The first head was
the research and development
(R&D) establishments of the min-
istries, each ministry being a “state
inside the state” in the USSR,
where no companies existed. Many
of these establishments, in engi-
neering sciences, were called design
bureaus (DBs) [9]. This head, the
richest, was responsible for the ap-
plied research and designing and
testing of prototypes. To facilitate
technology transfer to the industry,
every DB was associated with some
plant. Of the ministries, the most
powerful were those of defense,
the nuclear industry, the space in-
dustry, the radio industry, commu-
nications, the aircraft industry, the
shipbuilding industry, the mari-
time fleet, and some others.
IMAGE LICENSED BY INGRAM PUBLISHING

The second head was a network


of large laboratories called R&D in-
stitutes of the Academy of Sciences
(AS) of the USSR (in reality, this
was the AS of Russia) and similar
academies of sciences of the union
republics. The AS of the Ukrainian
SSR [now NASU)] was the largest of

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 9


the latter, hosting around 25% of all AS research laborato- Pivdenmash) Industry, which was, since the mid-1950s,
ries and manpower [8]. This head, officially, was respon- one of three major rocket, missile, and spacecraft indus-
sible for fundamental research using direct state funding; trial complexes in the USSR [9]. Of course, Pivdenmash
however, it was allowed to compete for the projects funded belonged to the extremely powerful USSR “Ministry of
by the ministries. The third head, the poorest, represented General Machine Building,” an Orwell-style cover for the
the university science where professors were encouraged to Ministry of Space Industry. For instance, the famous SS-18
take projects funded again by the ministries. This activity Satan heavy intercontinental ballistic missile (ICBM) and
was concentrated exclusively in large cities. some of the military satellites were developed and manu-
Since Stalin’s times, the research patterns of all acade- factured here until 1992.
mies of sciences and universities were heavily biased toward
technical and engineering sciences with either military or PREHISTORY, NAMES, AND DATES
double-purpose applications in mind. The CP and gov- Since 1976, IMH in Sebastopol and DBP in Dnipro were
ernment priorities were crystal clear: 1) nuclear weapons, involved in the design of experimental USSR satellites Cos-
2) missiles to deliver nuclear weapons, and 3) radars to aim mos-1076 and Cosmos-1151, equipped with low-resolution
and guide nuclear weapons. From the 1950s to the end of radar-like sensors called scatterometers [1], [2], [3], [6]. Their
the USSR in 1992, military-flavored research projects con- task was determining the parameters of the sea waves, in
tributed sizable funds to the budgets of all AS institutes that line with a secret decree of the CC CPSU on the develop-
related to what we can call, for brevity, the IEEE scope of inter- ment of the general-purpose orbital remote sensing system
ests. Still, there existed an important difference between the “Resurs.” By that time, IMH had already enjoyed collabora-
AS R&D institutes in Russia and outside of Russia; the latter tion around sea wave research, using coastal and airborne
could not have more than 25% of the total budgets coming sensors, with the radar group of Kalmykov at IRE NASU in
from the ministries and industry, while the former were al- Kharkiv [10], [11], [12], [13], [14], [15] (Figure 1).
lowed to exceed this limit. However, the “scatterometers” of the late 1970s, which
Such a limitation had, obviously, political origins and were a sort of radar prototype device, had failed to satisfy
reflected the distrust of “union republics.” It was estab- the customers, who were from various state services and
lished by the Science Department of the Central Commit- organizations, including polar navigation, maritime and
tee of the Communist Party of the Soviet Union (CC CPSU) port services, meteorology, etc. This proved the necessity of
as the supreme supervising and controlling body over all more concentrated efforts aimed at the development of ac-
ministries, academies, and universities. tive microwave sensors, i.e., radar.
Of some 50 R&D institutes of NASU, the second-largest Thanks to the fact that the work on the whole subsys-
cluster, after Kyiv, was and still is in Kharkiv. In particu- tem “Resurs-O” (i.e., oceanic survey satellites) was super-
lar, IRE (now IRE NASU) used to be the national research vised from DBP (see the “Obstacles to Overcome: Not Only
center of physics and technology of microwaves and mil- Technical” section), Kalmykov could expect to be in the
limeter waves. The IRE is the focus of our story, together center of the associated design and testing. However, he
with the Institute of Marine Hydrophysics (IMH NASU) lacked both equipment and R&D manpower. Part of the
in Sebastopol (currently occupied by Russia). Another problem was the extreme hostility of the then-IRE admin-
R&D establishment that played a crucial role was the DB istration [6].
“Yuznoye” [now DB Pivdenne (DBP)] in Dnepropetrovsk According to insiders [16], by the summer of 1979,
(now Dnipro). This is an engineering laboratory, now in- Kalmykov had given up and decided to move to IMH in
dependent and then associated with the Yuzhmash (now Sebastopol. As stated by the same source, it was the IMH
director who persuaded the top bosses of the extremely
powerful USSR Ministry of Space Industry to intervene and
rescue Kalmykov’s team at IRE. The then-director of IRE, V.
P. Shestopalov, received a phone call from Moscow, suggest-
ing that he urgently organize, at IRE, a research unit deal-
ing with space radio oceanography and sea ice sensing. The
ministry also promised to allocate IRE significant funds
dedicated to such research. As a result, a 20-strong Depart-
ment of Earth Remote Sensing Techniques was created at
IRE on 1 September 1979, headed by Kalmykov.
Immediately, the department initiated the R&D of a nov-
el all-weather active orbital sensor, specifically designed to
study the sea surface and ice covers. This was an X-band RA-
SLR. One group was designing a 100-kW pulse power mag-
netron source, another group designed a slotted-waveguide
FIGURE 1. Anatoly Kalmykov in his office at IRE NASU around 1990. antenna, and still another group was responsible for the

10 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


signal processing. A prototype airborne system allowing Besides carefully selected radar parameters, the high
in-flight testing was also designed, and systematic flights efficiency of the SLR Cosmos-1500 system was expected
onboard a dedicated MI-8 helicopter were organized. Be- due to the simultaneous acquisition of overlapping images
sides, it was decided to add another passive sensor to the from two other sensors so that three different wavelength
SLR, working in the millimeter-wave range—a Ka-band ra- bands were involved. This could enable the improved in-
diometer, also developed at IRE. Moreover, to produce the terpretation of images and the elimination of errors in
images, an onboard electronic data processing block was retrieved parameters. Further, the onboard preliminary
developed at DBP and IMH and added both to the airborne processing of the radar data and the transmission of the
prototype and to the orbital system.
The very first airborne experiments soon confirmed the
high efficiency of the designed instruments for studying the
water and ice surfaces [17], [18], [19], [20]. The joint use of 8
microwave images obtained from the X-band SLR and Ka-
band radiometer offered, in principle, more efficient study 7
of the state of the sea and ice than using the data from each
individual sensor. However, initial tests had also shown
that obtaining reliable information on the water-surface 2
waving needed a much deeper level of data processing than
available at that time. In contrast, quite reliable data were 4 1
3
obtained in a simpler way in the helicopter observations
of ice. The results of airborne studies convinced Kalmykov 2
of the favorable prospects for radar observations of sea ice
from space. 3
Still, attempts to interpret the ice-sounding data be- 3 5
yond simple discrimination between thin and thick ice
did not lead, unfortunately, to the creation of an adequate 3
model. The phenomenon of the scattering from the ice 6
turned out to be much more complicated than the scat-
tering from the water surface. Still, other possible applica-
tions emerged, such as wind measurements and oil slicks FIGURE 2. Cosmos-1500 and its microwave remote sensing instru-
detection [18], [20]. ments: (1) bus, (2) solar panels, (3) rotatable instrument panels,
The Cosmos-1500 satellite (Figure 2) was launched on (4) SLR antenna, (5) radiometer, (6) optical sensors, (7) telescopic
28 September 1983 from Plesetsk by the Tsyklon-3 rocket mast, and (8) gravitational stabilizer [3].
vehicle (a derivative of a heavy ICBM SS-18) and placed
into low-altitude near-circular polar orbit. It remained
operational until 16 July 1986. This was the first ever civil TABLE 1. PARAMETERS OF THE SLR OF COSMOS-1500.
satellite to carry an X-band RA-SLR working at the wave- Wavelength 3.1 cm
length of 3.16 cm with vertical polarization; the swath Polarization VV
width was about 460 km, and the spatial resolution was Viewing angle range 20–46°
2.4–3.2 km in the flight direction and 1.3-0.6 km in the Antenna pattern width
normal direction [4], [22], depending on the incidence In the azimuthal plane 0.2°
In the elevation plane 42°
angle (see Table 1).
Spatial resolution
The antenna system was based on the 12-m-long slot- Along the flight direction 2.4–3.2 km
ted waveguide, which was kept folded at the launch and Transverse to the flight direction 1.3–0.6 km
then automatically unfolded in orbit. This radar was sup- Average resolution in the swath, provided via APT
In the UHF band 0.8 × 2.5 km
plemented with a 37-GHz horizontally polarized side-
In the VHF band 2 × 2.5 km
looking passive radiometer, designed at IRE NASU, and
Receiver sensitivity −140 dB/W
a four-channel visible range imaging system from the In- Transmitter power 100 kW
stitute of Radio-Engineering and Electronics (IRE RAS) in Pulse duration 3 µs
Moscow. The polar orbit was selected to provide data on the Pulse repetition frequency 100 Hz
ice conditions in the Arctic in the hope to be useful for the Orbit altitude 650 km
navigation of ships in the northern latitudes, which were Orbital inclination 82.6°
not visible from geostationary satellites. The chosen RA- Swath 450 km
SLR parameters were considered as optimal for all-weather
Reproduced from [4] after correction of typos and translation mistakes.
studies of the polar sea ice covers and the dynamics of ice VV: vertical transmit-vertical receive.
formation, migration, and melting.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 11


synthesized images, using the simple 137.4-MHz Auto- Lenin was sent to the De Long Strait. However, soon, she
matic Picture Transmission (APT) channel, to hundreds got one propeller crushed by the ice and the other dam-
of users, including a central site in Moscow and autono- aged. Her brand-new nuclear-powered sister ship Brezhnev
mous points in Kharkiv and Sebastopol, was also a very (named after the recently deceased general secretary of
big step ahead. the CC CPSU) had also failed to crush the pack ice. In
Here, Kalmykov had to fight with Moscow colleagues the polar night season, air surveillance was pointless, and
from IRE RAS, who wanted to have a full monopoly on the SLR of Cosmos-1500 became the only available source
satellite imagery. Many details of the SLR of Cosmos-1500 of trusted sea-ice information (as already mentioned, the
design and operation can be found in reviews [2], [3], [4], other all-weather instrument, the onboard Ka-band radi-
[5], [6], [21], [22], [23]. Being a general-purpose instrument, ometer, failed to work) in and around the De Long Strait.
it outperformed greatly the preceding USSR military orbital Already, the first radar images of the disaster area
SLR “Chaika” aimed at the search of massive surface tar- ­(Figure 4) showed that the situation was not hopeless. In-
gets such as the U.S. Navy air carriers [3] (see the “Other deed, 100 km north of the caravan, near Wrangel Island,
Contemporary Orbital Radar Systems: A Monster in the an extensive polynya (a sea area where the ice is either
Shadow” section). absent or very thin) could be seen, together with a strip
of wide cracks and crevasses in heavy multiyear ice along
WHEN LENIN WAS HELPLESS: RESCUE MISSIONS OF which it was possible to direct the caravan to the polynya.
THE COSMOS-1500 SIDE-LOOKING RADAR Although the ad hoc staff of the rescue operation was re-
The work program of the new spacecraft envisaged many luctant to trust the microwave imagery, in the total ab-
weeks of meticulous tests and experiments; however, this sence of alternatives, it took up the risks and ordered the
had to be greatly revised at the very beginning as the on- icebreaker to go north. On reaching the polynya, the ice-
board Ka-band radiometer failed to operate. Moreover, only breaker and the freighters turned southwest and, in a few
five days after placement into the polar Earth orbit in the days, sailed in safe waters.
autumn of 1983, the SLR of Cosmos-1500 obtained a new
task, which was absolutely unexpected. By the time of the
launch of the spacecraft, a true drama had developed on
the Northern Maritime Route, which runs all the way from
the Atlantic to the Pacific along the Arctic coasts of the Rus-
sian Federation. That September, extremely strong north-
west winds pushed the heavy multiyear ice to the De Long
Strait near Wrangel Island, where a caravan of 22 freight-
ers (perhaps several caravans as sometimes 40 and even
57 ships plus five icebreakers are mentioned) got blocked.
The ships were loaded with cargo worth some US$8 billion
[23], which was carried as winter supplies to the Arctic re-
gions of the USSR. The MV Nina Sagaidak was soon crushed
by the ice and sank (Figure 3), and there was a real threat of (a)
further losses, especially as the polar night was approaching.
Wrangel Island
The authorities created ad hoc interservices staff to
monitor and guide the caravans. Besides conventional
icebreakers, the first USSR nuclear-powered icebreaker

Pen. Chukotka

(b)

FIGURE 4. The rescue mission of the USSR freighter caravan in the


De Long Strait, October 1983. (a) A radar image and (b) a topical
map demonstrating the ships location and the route of their escape
from the heavy ice area ( thin young ice, one-year ice, thick
FIGURE 3. The MV Nina Sagaidak sinking in the Arctic Ocean in perennial ice; rescue route is the yellow line. Pen.: peninsula.
September 1983. (Source: Reproduced from [4].)

12 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Amazingly, within its 33-month lifetime, the SLR of that of Figure 6(c) was obtained 14 h later. These images
Cosmos-1500 was destined to fulfill another rescue mission, allowed the correct estimation of the velocity of the cy-
this time in the Southern Hemisphere [2], [3], [4], [5], [6], clone translocation, around 7 km/h, and its total power,
[23]. The research MV Mikhail Somov, sent in 1985 to the about 1.2 × 108 MW.
Antarctic to bring a rotation crew to a USSR polar station,
was blocked in the 5-m-thick ice. To rescue her, a USSR OBSTACLES TO OVERCOME: NOT ONLY TECHNICAL
icebreaker was sent all the way from the Northern Hemi- When creating the RA-SLR of Cosmos-1500, Kalmykov had
sphere; this was the conventional vessel Vladivostok as New to solve many problems of organizational, technical, and
Zealand, where the final replenishment had to take place, human nature. Until 1972, Kalmykov closely collaborated
had already prohibited nuclear-powered ships from using with IRE’s theoreticians and enjoyed support and encour-
its harbors. What is important for our story is that onboard agement from the first IRE director, O. Y. Usikov. All that
the Vladivostok, a satellite information reception point was changed when the latter was replaced by V. P. Shestopalov
deployed to ensure the quick reception of microwave radar by the decision of the CPSU Committee of the Kharkiv Re-
images from Cosmos-1500. gion. The new director was a relative newcomer in the R&D
The high polar orbit of Cosmos-1500 was well suited for institute as, until the end of the 1960s, he was just an as-
such a mission. Its images (see Figure 5) enabled daily cor- sociate professor in physics at second-level universities in
rections of the icebreaker route in the ice fields both on the Kharkiv. His career sky-rocketed when a cousin of his wife
way to the blocked ship and back to the clean waters. became a secretary of the CPSU Committee of the Kharkiv
At the crucial phase of the operation, radar images re- Region. It is no surprise that (as explained in [3], [6]) he
vealed a wide polynya in heavy ice, stretching toward the was looking at space radar research, which promised both
drifting ship. Thanks to this, instead of using a helicopter heavy responsibility and tight control from Moscow, as a
to evacuate the crew of the drifting ship, which would high-risk activity that should be avoided. As mentioned, by
have had to be abandoned, the icebreaker rushed toward 1979, Kalmykov had gotten so desperate that he decided on
Mikhail Somov, freed her from the trap, and led her out of moving to IMH. In Sebastopol, his friend (and the head of
the ice [23].
When preparing this publication, we discovered that in
today’s Russia, the role of Cosmos-1500 and its RA-SLR in
the maritime rescue missions of 1983 and 1985 is subject
to total oblivion. In several “documentary” films made in
the 2010s, and even in the Russian-language Wikipedia,
the existence of Cosmos-1500 is not mentioned at all, and
instead, it is the “intuition” of icebreaker captains that is
highly credited. Okay, intuition can be a powerful thing,
especially when it is supported by microwave radar images
received twice a day. A comprehensive 675-page Russian
monograph [24], published in 2010 by the leading staff of
the USSR space synthetic aperture radar (SAR) works at the (a)
Vega State Co., mentions Cosmos-1500 and its SLR. How-
ever, it does not mention its polar seas missions; instead of
IRE NASU, the development of this radar is linked to the
Kharkiv Institute of Radio Electronics, which was a techni-
cal university.
It should also be noted that within the 19-month pe-
riod between two polar rescue missions and after the
second of them, the RA-SLR of Cosmos-1500 was engaged
in its main operational tasks: research into the remote
sensing of the mesoscale phenomena caused by the in-
teraction between the ocean and the atmosphere. This Arctic Continent
related, first of all, to the detection and tracing of cy- (b)
clones, typhoons, and hurricanes; however, less pow-
erful formations, such as quasi-regular convective cell FIGURE 5. The rescue mission of the USSR MV Mikhail Somov
structures, cloud fronts, and vortices were also studied in the Antarctic, July 1985. (a) A radar image and (b) a topical
[2], [6], [23]. Figure 6 shows optical ­[Figure 6(a)] and mi- map demonstrating the ship location and the escape route of the
crowave [Figure 6(b) and (c)] images of the tropical cy- icebreaker Vladivostok. (Reproduced from [4]; the color legends
clone Diana dated 11–12 September 1984. The images of are the same as in Figure 4. Note a misprint: the continent edge is
Figure 6(a) and (b) were obtained at the same time, and Antarctica.)

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 13


the collaborating group) was V.V. Pustovoytenko, who had Moscow and Leningrad. Konyukhov was a rare exception.
his own troubles with his administration but was support- Perhaps this was because of the success of ICBM SS-18 de-
ed both by Moscow and Dnipro. velopment at DBP. Besides, it was his predecessor at DBP, V.
As admitted in [3], [6], the resistance of V. P. Shestopalov M. Kovtunenko, who initiated, in 1974, the development
was partially overcome only due to the extraordinary per- of equipment for the study of oceans from orbit [3], [5],
sonal efforts of S. N. Konyukhov, the head of the rocket di- leading to the previously mentioned secret decree of the
vision of DBP in Dnipro, who was made responsible for the CC CPSU and the government (1977) about the creation
whole remote sensing payload for the “Resurs-O” program. of the system “Resurs.” Still, it was the success of SAR work
Here, it should be explained that, in the USSR, R&D centers onboard the U.S. Seasat in 1978 and its huge effect on the
located not in Russia but in other republics seldom coordi- USSR political and military leaders that caused a decision
nated the state programs. Usually, this was entrusted either to speed up the work.
to the industry bosses, the military, or the R&D centers in Such was the background for the phone call to IRE’s
director from the USSR Ministry of Space Industry about
establishing a space radar R&D unit headed by Kalmykov.
However, even after obtaining his own department at IRE,
with rich funding from the ministry, Kalmykov’s working
conditions remained far from perfect.
Thus, Konyukhov coordinated the work of all three
Ukrainian R&D centers, one ministerial (DBP) and two
academic (IRE and IMH). Still, he was supervised by his
ministry in Moscow, where the other powerful organiza-
tions, such as IRE RAS and the almighty ministerial Cen-
tral R&D Institute of Device Building, were developing
the optical sensors and, in part, the information storage,
processing, and transmission to customers’ equipment for
Cosmos-1500.
As recalled in [3], a mutually beneficial collaboration
between the Ukrainian teams was established quickly;
however, a similar level of synergy was never reached with
(a) the central organizations. In 1983, the conflict culminated
in a series of heated discussions where the directors and
leading experts of several powerful Moscow R&D centers
attacked Kalmykov, Pustovoytenko, and B. Y. Khmyrov
(Konyukhov’s successor at DBP). They demanded aban-
doning SLR in favor of SAR and, therefore, transferring
the radar development to their laboratories. To rationalize
their demands, which were fed by professional jealousy,
they referred to the success of the U.S. Seasat and used a
vague accusation of the allegedly insufficient “informa-
tion potential” of Kalmykov’s SLR data. Still, Moscow SAR
designs existed only in drawings and needed at least one
more year of intensive development and testing (in reality,
the first successful USSR SAR was placed into orbit only
in 1991), while Kalmykov had an unbeatable argument—
the successful operation of the airborne analog of his SLR.
Thanks to this circumstance, Khmyrov expressed full sup-
port to Kalmykov, the attacks of the Moscow colleagues
were rebuked, the IRE team released the radar, and Cos-
mos-1500 was assembled at DBP and launched according
to the schedule.
When Kalmykov’s SLR got successfully into its orbit, the
(b) (c) feud between the developers faded off, at least for a while.
However, suddenly, new powerful opponents emerged. As
FIGURE 6. Tropical cyclone Diana in the Atlantic Ocean, 11 Septem- mentioned, already before the placement of the SLR into
ber 1984. (a) An optical image and (b) and (c) microwave images orbit, a caravan of USSR freighters got blocked by heavy ice
obtained at 6:30 p.m. (b) and at 8:30 a.m. the next day (c). in the Eastern Arctic. The situation was made public by the

14 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


authorities, although with a delay, as usual for the USSR me- radar programs were focused mainly on monitoring land
dia. When Kalmykov learned about it, he tried to approach parameters [25], [26], [27].
the ad hoc committee put in charge of the rescue mission to Still, in deep secrecy, there existed another USSR or-
propose his aid. This happened to be nearly impossible. A bital RA-SLR system for oceanic observations, perhaps
legend tells that Kalmykov wrestled his way to the commit- a hundred times more expensive than Cosmos-1500 and
tee meeting room, showing the satellite imagery printouts all its derivatives. It was initiated as early as 1960, first
to the KGB guards; however, more probably, he had found placed into low-Earth orbit in 1975, and finally, closed
someone who played the role of mediator. Still, this was in 1988 after at least 39 launches. This radar was called
not the end. Unexpectedly, the top administration of the Chaika, and it and the spacecraft equipped with it were
USSR Chief Directorate of the Northern Maritime Route, part of the top-secret naval reconnaissance and target-
the dominant service in the rescue committee, displayed a ing system “Legenda” [24]. In the West, they were known
huge distrust of the satellite data, which suggested a non- as Radar Ocean Reconnaissance Satellites (RORSATs) [28],
trivial escape route—to the north of the disaster site. At the [29], [30]. Each RORSAT had two magnetrons (principal
crucial moment, Kalmykov had to voice a threat to file a and backup); one, or after 1985, two RA-SLRs working
complaint to the superpowerful authority: the CC CPSU. at the frequency of 8.2 GHz; and one or two 10-m-long
This worked out, and a nuclear-powered icebreaker was slotted-waveguide antennas to provide left-side and right-
­ordered to move north. side swaths, each 450 km wide (Figure 7). These satellites
were designed to find and track U.S. Navy air carriers,
OTHER CONTEMPORARY ORBITAL RADAR SYSTEMS: first of all, in the North Atlantic and North Pacific, and
A MONSTER IN THE SHADOW release the targeting data to the USSR assault triad: Navy
The first space-based microwave Earth imaging experiment bombers of the Tu-22M3 type, superheavy cruise missile
using the L-band SAR of the U.S. Seasat-A satellite was con- submarines of the Oskar-I and Oskar-II (Kursk) types,
ducted in 1978. That radar worked for three months at the and heavy guided missile cruisers of the Pyotr Velikiy and
wavelength of 23 cm with a swath of 100 km and provided Moskva types.
a spatial resolution of 25 m [25], [26]. The results of this Each component of the triad had to launch many dozens
experiment exceeded all expectations and showed the high of cruise missiles with conventional and nuclear warheads.
capabilities of orbital systems. However, the radar images For instance, according to a comprehensive description of a
were synthesized not onboard but on the ground, with retired USSR Navy officer [31], to attack one U.S. air carrier
great delay, which prevented their use in time-sensitive ap- from the air, as many as three full regiments of Tu-22M3
plications. Essentially the same test SARs operated onboard medium-range strategic bombers (i.e., 100 aircraft) were as-
the Space Shuttle Columbia in 1981 (five days) and 1984 (sev- signed. Some of the bombers and all dedicated submarines
en days) [25], [26], [27]. and cruisers were equipped with receivers of the “Legenda”
It should be noted that, in parallel to Kalmykov’s SLR, system (Figure 8). Through the network of communication
the USSR SAR systems were also developed: in Moscow. satellites known as Parus, RORSAT data information was
Test SAR “Travers” was installed onboard the spacecraft passed on to these assets and a dedicated USSR Navy con-
Resurs-O-1 launched as Cosmos-1689 in 1985 [2], [3] trol center in Noginsk near Moscow [24]. The task of the
and, later, on the Priroda module of the orbital station system was not just to locate and identify naval ships but to
(OS) Mir. The other SAR required full OS power; it was provide the targeting data that could, allegedly, be fed di-
launched in 1987 onboard Cosmos-1870 and in 1991 on rectly into antiship missiles, such as the 6-ton X-22 liquid-
OS Almaz-1 [6], [24]. propellant ones carried by the Tu-22M3.
Despite an order of magnitude lower resolution than The reason for such a tremendous concern was that in
SAR, SLR was attractive due to higher radiometric accu- the 1970s and early 1980s, the USSR submarine-launched
racy and an order wider swath. It could use an available ballistic missiles (SLBMs) had limited range and accuracy, so
simple magnetron source, which had less stable character- to fire them, the submarines had to come nearer to the U.S.
istics than needed for SAR; additionally, onboard image East and West Coasts. Therefore, the U.S. Navy air carriers
processing, lower cost, and much quicker delivery were were viewed as an extremely dangerous force, able to block
also very important. The orbital system of Cosmos-1500 or destroy the USSR submarine fleet in their home bases
had no contemporary analogs in the day-to-day practical at the Cola and Kamchatka Peninsulas. Still, traditionally,
monitoring of the ocean and ice. It was true that the Seasat according to the USSR and Russian military doctrine and
and shuttle SAR experiments (and later ERS-1, RADAR- ethos, all bombers, submarines, and cruisers taking part in
SAT, and other SAR systems) were primarily designed to a raid on a U.S. air carrier were viewed as expendables. As for
serve oceanography and generally met and even exceeded the bombers, probable losses were estimated at 50%; “Leg-
expectations. However, they turned out to be even more enda” was not trusted by the pilots and the air staff, and a
useful for land applications, where a several-day delay in suicide raid of two dedicated Tu-16 reconnaissance aircraft
signal processing was not as critical as in maritime naviga- was always envisaged to make visual contact with air car-
tion. As a result, the practical components of their space riers [31]. Similarly, the Navy staff always sent a destroyer

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 15


or even a trawler to follow every U.S. air carrier task force. the submarine reflector antenna of the “Legenda” receiver
Both submarines and cruisers shared the same nickname was a huge retractable structure christened Punch Bowl by
of “single-shot assets” as the reloading of their missiles was the North Atlantic Treaty Organization (NATO) fleets (Fig-
not available. According to [31], they expected no more than ure 9). To use it, the submarines had to stay at periscope
30 min of life after firing their first and the last salvos of 24 depth for long hours preceding their attack, which should
or 20 (or 16) “Granit” cruise missiles, respectively. Moreover, have added to the kamikaze spirit of their crews.

Conceptual Configuration of US-A Spacecraft


Based on Sketches in (24) Image From russianspaceweb.com
Ejection of Reactor Core

NaK Droplets Deployment of Side-Looking


Radar Antenna

Ejected
Reactor Reactor
Core Container
5.8 m Attitude Thruster Pod
3,495

1,135 1,300 Orbit Insertion Propulsion System


Shield 4E18 Propulsion System

892
1,075 257
Reactor Radiator Propulsion
System Main Spacecraft Side-Looking Radar Antenna,
to Raise Including Radar Payload One on Each Side of the Spacecraft f = 8.2 GHz
19.542-MHz Transmitter in this Part Reactor 166-MHz Transmitter 0 5
in this Part (m)
(a)

(b)

FIGURE 7. (a) RORSAT concept and configuration. (Source: [28].) (b) A demonstration copy at the DB “Arsenal” in St. Petersburg showing
the nuclear reactor at the forward end and two unfolded SLR antennas in the rear. (Source: [24].)

16 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


(a) (b)

FIGURE 8. (a) “Legenda” satellite receiver antennas on the Pyotr Velikiy nuclear-powered USSR cruiser (a)—white radome on a vertical
structure at the side of the tower—and (b) on the sunken Moskva cruiser (b)—similar light-grey radome just above the rear launch tube.
To eliminate the shadowing of antennas, the same equipment was placed at the portside.

The whole USSR orbital naval reconnaissance and reactor into a high “graveyard” orbit and instead crashed
targeting system “Legenda” carried the same stigma of over Canada’s Great Slave Lake, contaminating a wide area.
hopeless gigantomania and kamikaze spirit. The USSR The debris was examined by U.S. Lawrence Livermore Na-
electronics of that time were quite backward and unreli- tional Laboratory scientists, which enabled them to get a
able; additionally, the signals backscattered from ships better understanding of the design and mission of ROR-
are accompanied by intensive clutter due to the sea SATs. This catastrophe led to a two-year break in RORSAT
waving and precipitation effects. To compensate for in- launches, used for improvements in their design. Still, an-
sufficient sensitivity and poor signal processing, USSR other similar accident happened at the beginning of 1983
developers used an extremely monstrous approach. First, with Cosmos-1402 when separate parts of the reactor fell
the RORSATs used small fast neutron nuclear reactors into the Indian and Atlantic Oceans.
(“Buk”) to provide the 3-kW power needed to feed the ra-
dar; second, they always flew at low orbits of 250–270 km
with a 65° inclination that made their lifetimes short,
less than two months on average. Even more—to enable
determining the direction and speed of the sea target
with such an incoherent sensor as SLR—a primitive but
efficient solution was found—nuclear RORSATs were
launched in pairs and placed into identical orbits with
a half-hour separation [28].
The combination of a low orbit and a nuclear power
source introduced a serious risk of accident or uncontrolled
reentry [29].
“To counter the problem, each RORSAT consisted of (a)
three major components: the payload and propul-
sion section, the reactor, and a disposal stage used to
lift the reactor into a higher orbit, with an altitude of
900 km, at the end of the mission.”
Each of at least 33 reactors launched in 1975–1988 con-
tained more than 30 kg of weapon-grade (enriched to 90%)
uranium-235, besides the sodium-potassium coolant. This
means that presently about 940 kg of highly enriched ura-
nium and a further 15 tons, mostly shaped as tens of thou-
sands of radioactive coolant droplets, 0.6–2 cm in diameter, (b)
orbit Earth [29].
There were several accidents of the malfunctioning FIGURE 9. (a) and (b) “Legenda” receiver antenna in a radome on
of RORSATs that triggered public attention to the dan- the top of the coning tower of the USSR Oskar type submarines
ger they presented. On 24 January 1978, five years before (one of them was the ill-fated Kursk, which exploded in 2000)
Kalmykov’s success, Cosmos-954 failed to throw its nuclear (a) in the harbor and (b) at sea. (Sources: [32], [33].)

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 17


According to the U.S. Central Intelligence Agency as- his IMH colleagues that the development of the signal pro-
sessment, RORSATs were able to track U.S. air carriers in cessing for SLR “Chaika” was facilitated by the Black Sea
good sea and weather conditions; however, they became experiments with its airborne analog on a dedicated turbo-
useless otherwise. These spacecraft were so tremendous- prop aircraft [24].
ly expensive and slow in production that their launches However, given the USSR spy mania and intensive and
were usually tied to the massive naval drills of the U.S. even brutal rivalry between the rocket DBs, Kalmykov
and NATO fleets, thus leaving lengthy gaps in air carrier could not know anything except general terms. Note
tracking. By the end of the 1980s, the USSR SLBMs had the difference in the satellite composition: vertical for
improved their range and accuracy to the extent that the Cosmos-1500 versus longitudinal for RORSATs (see Fig-
submarines could stay on patrol just in the Sea of Ok- ures 2 and 7). This was apparently connected to the flight
hotsk, which obtained the name of “a submarine aquar- altitude; low-orbit RORSATs had to have a more “aero-
ium.” Therefore, the RORSAT program was terminated in dynamic” shape to reduce the effect of the atmosphere,
1988, although the other part of the “Legenda” system, while Cosmos-1500 could instead neglect it. It is also im-
higher-orbit electronic intelligence satellites (EORSATs), portant that his SLR was developed 10 years later than
survived, and their derivatives are still in operational use “Chaika” and integrated Kalmykov’s multiyear collabo-
by the Russian Navy [9]. ration with IRE’s theoreticians around the sea and ice
For our story, it is interesting to admit that the frequency backscattering research. In terms of onboard signal pro-
of operation, the type of microwave source and antenna, cessing, Cosmos-1500 implemented two relatively new, at
and the swath width of Kalmykov’s RA-SLR and the “Chai- least for the USSR space programs, operations: incoher-
ka” SLR were rather similar to each other. Kalmykov had ent integration of eight successive images along the swath
security clearance and should have known about the ex- and compression of the image intensity dynamic range
istence of naval RORSATs designed by the rocket and mis- across the swath using the automatic gain control. Addi-
sile DB “Prikladnaya Mekhanika” (Applied Mechanics) in tionally, telemetric data on the current parameters of the
Moscow (now Khrunichev State Co.). He could even know SLR units were also transmitted to the ground stations.
about their design principles because EORSAT satellites Therefore, the SLR of Cosmos-1500 can be safely consid-
were designed at DBP and produced serially at the Yuzh- ered as an original instrument.
mash Industry in Dnipro. Moreover, he should know from
POST-HISTORY: FROM USSR AWARDS TO
A POSSIBLE IEEE MILESTONE
Two polar sea rescue missions of Cosmos-1500 gave a rare
chance to the USSR authorities to present the USSR space
program as a peaceful, truly useful, and efficient activity,
while in reality, it was heavily biased to military use, poorly
balanced, and plagued by the fierce rivalries of different
players, and it suffered from numerous accidents and catas-
trophes [9]. These missions were broadly highlighted in the
USSR newspapers and on TV. Already in 1985, a full-size
copy of Cosmos-1500 was displayed in Moscow at a perma-
nent exhibition (Figure 10).
This publicity and attention helped obtain fair recogni-
tion at the national level. In 1987, Kalmykov and his nine
colleagues were awarded the National Prize of Ukraine in
Science and Technology with the citation, “For the devel-
opment and implementation of radar methods of Earth
remote sensing from aerospace platforms.” The recogni-
tion at the USSR level was restricted to several state orders,
the highest of which, Lenin’s Order, was given to the same
IRE director who nearly pushed Kalmykov out in 1979. For
comparison, the developers of the secret naval SLR “Chai-
ka” and the whole system “Legenda” were awarded, despite
its low efficiency and RORSAT disasters, a secret Lenin’s
Prize of the USSR.
As mentioned previously, the design and development
FIGURE 10. The spacecraft Cosmos-1500 with microwave equip- of the X-band orbital RA-SLR of Cosmos-1500 led to the suc-
ment for remote sensing of Earth at the permanent USSR Exhibition cessful overcoming of a wide range of scientific and techni-
of Achievements of National Economy in 1985. (Source: [4].) cal problems. This enabled the technology transfer to the

18 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


R&D Institute of Radio Measurements in Kharkiv (now the ACKNOWLEDGMENT
RADMIR Institute), which had a small-series production We thank the Physics Benevolent Fund of the Institute
line. In all, essentially the same SLR was exploited on six of Physics, U.K., for one-off emergency support in the
remote sensing satellites of the USSR/Russia Space Opera- context of solidarity with Ukraine. Alexander I. Nos-
tive System Okean in 1986–2004 and two Ukrainian satel- ich acknowledges the support of the European Federa-
lites named Sich in 1997 and 2004. It was successfully used tion of Academies of Sciences and Humanities via a
to detect and monitor many critical situations and natural research grant from the European Fund for Displaced
phenomena on a global scale [2], [3], [4], [5], [6], [19], [20], Scientists and the hospitality of the Institute of Elec-
[21], [22], [23]. tronics and Numerical Technologies of the University
When Ukraine got its independence in 1992, its space of Rennes, France.
industry, centered around DBP and Pivdenmash, hoped to
keep working, although Ukraine had no launch sites and AUTHOR INFORMATION
only a few tracking and control facilities. Indeed, Russia Ganna B. Veselovska-Maiboroda (veselovska3@ukr.net)
was dependent on DBP for the maintenance of its major is with the Department of Physical Foundations of Radar,
nuclear ICBM force of several hundred silo-based SS-18 and O.Y. Usikov Institute of Radiophysics and Electronics, Na-
for the supply of Tsiklon boosters and EORSAT satellites. tional Academy of Sciences of Ukraine, 61085 Kharkiv,
As a part of the bargain, two Ukrainian radar remote sens- Ukraine.
ing satellites named Sich were launched from the Russian Sergey A. Velichko (sergey_velichko@yahoo.com) is
launch sites. The first, Sich-1, was fully operational for three with the Department of Earth Remote Sensing, O.Y. Usikov
years: from 1997 to 2000. However, the second of them, Institute of Radiophysics and Electronics, National Acad-
Sich-1M, was placed into the wrong orbit and failed to deliv- emy of Sciences of Ukraine, 61085 Kharkiv, Ukraine.
er the expected data. Given that Russia’s president in 2004 Alexander I. Nosich (anosich@yahoo.com) is with the
was the same as today and in view of the Russian invasion Laboratory of Micro and Nano Optics, O.Y. Usikov Institute
of Ukraine, one can guess that the “wrong orbit” of Sich-1M of Radiophysics and Electronics, National Academy of Sci-
was one more secret-service operation, aimed at denying ences of Ukraine, 61085 Kharkiv, Ukraine. He is a Fellow
Ukraine sensitive information and spoiling its reputation of IEEE.
as a reliable spacecraft developer.
After 1992, Kalmykov had to restrict his work to the air- REFERENCES
borne analog of his SLR. Such a system, called MARS, was [1] “NASA space science coordinated archive, cosmos 1500,”
developed and used by various Ukrainian ministries and NASA, Washington, DC, USA. Accessed: Jun. 30, 2023. [On-
services [34], [35], [36], [37], [38], [39]. However, the unex- line]. Available: https://nssdc.gsfc.nasa.gov/nmc/spacecraft/
pected death of Kalmykov in 1996 from kidney trouble left display.action?id=1983-099A
a void that was hard to fill. [2] A. I. Kalmykov, A. S. Kurekin, and V. N. Tsymbal, “Radiophysi-
Broadly speaking, the development and operation of cal researh of the Earth’s natural environment from aerospace
the RA-SLR of Cosmos-1500 initiated, 40 years ago, a new platforms,” Telecommun. Radio Eng., vol. 52, no. 3, pp. 41–52,
research area and discipline in the Ukrainian microwave 1998, doi: 10.1615/TelecomRadEng.v52.i3.100.
community—remote sensing of Earth from aerospace [3] G. K. Korotayev et al., “Thirty years of domestic space oceanol-
platforms. Therefore, at the national level, its impact is ogy,” Space Sci. Technol., vol. 13, no. 4, pp. 28–43, 2007.
truly huge. In the context of the USSR science and tech- [4] V. K. Ivanov and S. Y. Yatsevich, “Development of the Earth
nology, it played the role of a prototype for the family of remote sensing methods at IRE NAS of Ukraine,” Telecommun.
sea ice-monitoring satellites that provided safe navigation Radio Eng., vol. 68, no. 16, pp. 1439–1459, 2009, doi: 10.1615/
in the Arctic from 1986 to 2004. Besides, it was used by TelecomRadEng.v68.i16.40.
Russian researchers to study the formation and dynamics [5] V. V. Pustovoytenko et al., “Space pilot of the nuclear-powered
of the ice covers of the Sea of Okhotsk [40] (it is quite pos- vessels,” in Proc. Int. Crimean Conf. Microw. Telecommun. Technol.
sible that this research was initiated by the USSR military (CriMiCo), 2013, pp. 19–22.
as that distant sea, bounded by the scarcely inhabited Rus- [6] A. G. Boyev et al., Aerospace Radar Monitoring of Natural Disas-
sian Far East territories and Kuril Islands, became a sanc- ters and Critical Situations. Kharkiv, Ukraine: Rozhko Publica-
tuary for the USSR fleet of nuclear-powered Typhoon-type tion, 2017.
SLBM submarines in the late 1980s). At the global level, it [7] G. Veselovska-Maiboroda, S. A. Velichko, and A. I. Nosich,
was one of the cornerstones of what was christened ocean- “Orbital X-band side-looking radar of Cosmos-1500: Potential
ography from space [25] and had initiated the systematic use IEEE Milestone candidate,” in Proc. Int. Conf. Ukrainian Microw.
of radar images for safe polar navigation. Week (UKRMW), Kharkiv, Ukraine, 2022, pp. 670–673.
As we believe, all of the previously presented infor- [8] P. Kneen, Soviet Scientists and the State. Albany, NY, USA: State
mation suggests that the SLR of Cosmos-1500 satisfies the Univ. New York Press, 1984.
requirements of the IEEE History Committee to be nomi- [9] B. Harvey, The Rebirth of the Russian Space Program. Chichester,
nated as an IEEE Milestone in Ukraine. U.K.: Springer-Verlag, 2007.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 19


[10] A. I. Kalmykov, I. E. Ostrovskii, A. D. Rozenberg, and I. M. [26] C. Elachi et al., “Spaceborne synthetic-aperture imaging ra-
Fuchs, “Influence of the state of the sea surface upon the dars: Applications, techniques, and technology,” Proc. IEEE,
spatial characteristics of scattered radio signals,” Sov. Radio- vol. 70, no. 10, pp. 1174–1209, Oct. 1982, doi: 10.1109/PROC.
physics, vol. 8, no. 6, pp. 804–810, Nov. 1965, doi: 10.1007/ 1982.12448.
BF01038278. [27] J. Cimino, C. Elachi, and M. Settle, “SIR-B-the second Shut-
[11] A. D. Rosenberg, I. E. Ostrovskii, and A. I. Kalmykov, “Frequen- tle imaging radar experiment,” IEEE Trans. Geosci. Remote
cy shift of radiation scattered from a rough sea surface,” Sov. Sens., vol. GE-24, no. 4, pp. 445–462, Jul. 1986, doi: 10.1109/
­Radiophysics, vol. 9, no. 2, pp. 161–164, Mar. 1966, doi: 10.1007/ TGRS.1986.289658.
BF01038952. [28] A. Siddiqi, “Staring at the sea: The Soviet RORSAT and ­EORSAT
[12] F. G. Bass, I. M. Fuks, A. I. Kalmykov, I. E. Ostrovsky, and A. programmes,” J. Brit. Interplanetary Soc., vol. 52, nos. 11–12,
D. Rosenberg, “Very high frequency radiowave scattering by pp. 397–416, 1999.
a disturbed sea surface part I: Scattering from a slightly dis- [29] S. Grahn, “The US-A program (Radar Ocean Reconnaissance
turbed boundary,” IEEE Trans. Antennas Propag., vol. 16, no. 5, satellites – RORSAT) and radio observations thereof.” Accessed:
pp. 554–559, Sep. 1968, doi: 10.1109/TAP.1968.1139243. Jun. 30, 2023. [Online]. Available: http://www.svengrahn.
[13] F. G. Bass et al., “Radar methods for the study of ocean waves,” pp.se/trackind/RORSAT/RORSAT.html
Sov. Phys. Uspekhi, vol. 18, no. 8, pp. 641–642, 1975, doi: [30] R. Kopets and S. Skrobinska, “Russia’s space program is dead-
10.1070/PU1975v018n08ABEH004920. locked: A space naval reconnaissance incident,” Int. Relations,
[14] F. G. Bass et al., “Radiophysical investigations of sea roughness Public Commun. Regional Stud, vol. 2, no. 6, pp. 28–38, 2019.
(radiooceanography) at the Ukrainian Academy of Sciences,” [31] M. Y. Tokarev, “Kamikazes: The Soviet legacy,” Nav. War College
IEEE Trans. Antennas Propag., vol. 25, no. 1, pp. 43–52, Jan. 1977, Rev., vol. 67, no. 1, 2014, Art. no 7.
doi: 10.1109/JOE.1977.1145324. [32] “Soviet submarine fleet 1945-1990 part 3–11.” (in Russian),
[15] Y. M. Galaev et al., “Radar detection of oil slicks on a sea Moremhod. Accessed: Jun. 30, 2023. [Online]. Available: http://
surface,” Izvestia SSSR Fizika Atmos. Okeana, vol. 13, no. 4, moremhod.info/index.php/library-menu/16-morskaya-tematika/
pp. 406–414, 1977. 188-pf7?start=10
[16] V. D. Yeryomka, private communication, 2018. [33] V. V. Bychkov and V. G. Cherkashin, “Strategic concept of de-
[17] A. I. Kalmykov and A. P. Pichugin, “Special features of the detec- velopment of naval reconnaissance and targeting system,” Nat.
tion of sea surface inhomogeneities by the radar methods,” Izves- Security Strategic Planning, vol. 2021, no. 2, pp. 30–37, 2021, doi:
tia SSSR Fizika Atmos. Okeana, vol. 17, no. 7, pp. 754–761, 1981. 10.37468/2307-1400-2021-2-30-37.
[18] A. I. Kalmykov, A. P. Pichugin, Y. A. Sinitsyn, and V. P. Shesto- [34] A. Kalmykov et al., “Radar observations of strong subsurface
palov, “Some features of radar monitoring of the oceanic sur- scatterers. A model of backscattering,” in Proc. Int. Geosci. Re-
face from aerospace platforms,” Int. J. Remote Sens., vol. 3, no. 3, mote Sens. Symp. (IGARSS), 1995, vol. 3, pp. 1702–1704, doi:
pp. 311–325, 1982, doi: 10.1080/01431168208948402. 10.1109/IGARSS.1995.524001.
[19] V. B. Efimov et al., “Study of ice covers by radiophysical means [35] A. I. Kalmykov et al., “The two-frequency multi-polarisation
from aerospace platforms,” Izvestiya SSSR Fizika Atmos. Okeana, L/VHF airborne SAR for subsurface sensing,” AEU Int. Electron.
vol. 21, no. 5, pp. 512–520, 1985. Commun., vol. 50, no. 2, pp. 145–149, 1996.
[20] S. A. Velichko, A. I. Kalmykov, Y. A. Sinitsyn, and V. N. Tsymbal, [36] S. A. Velichko, A. I. Kalmykov, and V. N. Tsymbal, “Possi-
“Influence of wind waves on radar reflection by the sea sur- bilities of hurricanes investigations by real aperture radars
face,” Radiophysics Quantum Electron, vol. 30, no. 7, pp. 620–631, Cosmos-1500/Okean type,” Turkish J. Phys., vol. 20, no. 4,
Jul. 1987, doi: 10.1007/BF01036296. pp. 305–307, 1996, doi: 10.55730/1300-0101.2567.
[21] A. I. Kalmykov et al., “Information content of radar remote sens- [37] E. N. Belov et al., “Application of ground-based and air/­
ing systems of earth from space,” Radiophysics Quantum Electron, spaceborne radars for oil spill detection in sea areas,” Telecom-
vol. 32, no. 9, pp. 779–785, 1989, doi: 10.1007/BF01038802. mun. Radio Eng., vol. 51, no. 1, pp. 1–8, 1997, doi: 10.1615/
[22] A. I. Kalmykov et al., “Kosmos-1500 satellite side-looking ra- TelecomRadEng.v51.i1.10.
dar,” Sov. J. Remote Sens., vol. 5, no. 3, pp. 471–485, 1989. [38] A. I. Kalmykov et al., “Multipurpose airborne radar system
[23] A. I. Kalmykov, S. A. Velichko, V. N. Tsymbal, Y. A. Kuleshov, “MARS” for remote sensing of the Earth,” Telecommun. Ra-
J. A. Weinman, and I. Jurkevich, “Observations of the marine dio Eng., vol. 53, nos. 9–10, pp. 120–130, 1999, doi: 10.1615/
environment from spaceborne side-looking real aperture ra- TelecomRadEng.v53.i9-10.150.
dars,” Remote Sens. Environ., vol. 45, no. 2, pp. 193–208, Aug. [39] M. V. Belobrova et al., “Experimental studies of the spatial ir-
1993, doi: 10.1016/0034-4257(93)90042-V. regularities of radio-wave scattering in the Gulf Stream zone,”
[24] V. S. Verba, Ed., Spaceborne Earth Surveillance Radar Systems. Radiophysics Quantum Electron., vol. 44, no. 12, pp. 949–955,
Moscow, Russia: Radiotekhnika Publications, 2010. Dec. 2001, doi: 10.1023/A:1014873927923.
[25] W. S. Wilson et al., “A history of oceanography from space,” in [40] L. M. Mitnik and A. I. Kalmykov, “Structure and dynamics of
Remote Sensing of the Environment, Manual of Remote Sensing, vol. the Sea of Okhotsk marginal ice zone from “Ocean” satellite ra-
6. Baton Rouge, LA, USA: American Society for Photogrammetry dar sensing data,” J. Geophys. Res, vol. 97, no. C5, pp. 7429–7445,
and Remote Sensing, 2005, pp. 1–31. 1992, doi: 10.1029/91JC01596. GRS

20 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Airborne Lidar
Data Artifacts What we know thus far

IMAGE LICENSED BY INGRAM PUBLISHING

WAI YEUNG YAN

D ata artifacts are a common occurrence in airborne lidar


point clouds and their derivatives [e.g., intensity im-
ages and digital elevation models (DEMs)]. Defects, such as
INTRODUCTION
A number of recent surveys published between 2020 and
2022 project that the lidar market will likely reach US$2.9–
voids, holes, gaps, speckles, noise, and stripes, not only de- 4.71 billion by 2026 to 2030 [1], [2], [3]. Among all the mul-
grade lidar visual quality but also compromise subsequent timodal platforms, airborne lidar is projected to capture
data-driven analyses. Despite significant progress in under- the majority of the market share because of its wide area
standing these defects, end users of lidar data confronted coverage and enhanced data collection capability. Differ-
with artifacts are stymied by the scarcities of both resources ent advanced airborne lidar systems have been rolled out
for the dissemination of topical advances and analytic soft- lately, including dual/triple laser channel systems operated
ware tools. The situation is exacerbated by the wide-ranging with green and infrared laser wavelengths (e.g., Optech Ti-
array of potential internal and external factors, with exam- tan and RIEGL VQ-1560i-DW), dual-wavelength airborne
ples including weather/atmospheric/Earth surface condi- lidar bathymetry (ALB) (e.g., Optech CZMIL and Leica
tions, system settings, and laser receiver–transmitter axial Chiroptera-5), dual-channel topographic lidar (e.g., RIEGL
alignment, that underlie most data artifact issues. In this LMS-Q1560 and Optech Galaxy T-2000+G2), and single-
article, we provide a unified overview of artifacts common- photon lidar (e.g., Leica SPL100). Despite these high-end
ly found in airborne lidar point clouds and their derivatives commercial products, the data analytical capability of lidar
and survey the existing literature for solutions to resolve is unable to keep pace with hardware development. These
these issues. The presentation is from an end-user perspec- market surveys also emphasize that having insufficient
tive to facilitate rapid diagnoses of issues and efficient re- data processing software tools sets a barrier to further ex-
ferrals to more specialized resources during data collection pansion of the lidar market. As a result, raw lidar data likely
and processing stages. We hope that the article can also grow at a rate that exceeds our processing capacity with cur-
serve to promote coalescence of the scientific community, rent software tools.
software developers, and system manufacturers for the on- Traditional satellite remote sensing and aerial pho-
going development of a comprehensive airborne lidar point togrammetric imagery data ride on the advancement of
cloud processing bundle. Achieving this goal would further computer vision, machine learning, and image process-
empower end users and move the field forward. ing algorithms. Therefore, commercial software bundles
handling these geospatial imagery data are widespread in
Digital Object Identifier 10.1109/MGRS.2023.3285261
the market and with ever-advancing capability and ap-
Date of current version: 5 July 2023 plicability. In contrast, airborne lidar data encompass a

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 2473-2397/23©2023IEEE 21


wealth of topographic information represented by a set handling data artifacts. The inability to resolve these arti-
of georeferenced 3D point clouds together with back- facts at the beginning stage may render the extracted infor-
scattered laser signal strengths. Although such an active mation unreliable.
remote sensing approach overcomes the limitations of This review summarizes the artifacts commonly found
traditional imaging techniques [4], the unstructured and in airborne lidar data and their derivatives, the internal and
bulky 3D point clouds are challenging for end users to external factors underlying these defects, and the current
handle. The deficiency of airborne lidar data processing status of research handling and resolving these issues. It
software on the market can also be ascribed to the end us- should be noted, however, that this review centers on ar-
ers’ primitive understanding of the physical mechanism tifacts that have significant visual and analytical impacts,
behind how a lidar system works. Most end users sim- rather than on defects or noise reported in backscattered
ply conduct visual examination of the collected 3D point lidar waveform signals [12]. In addition, although the term
clouds through generating intensity images and DEMs/ “airborne lidar system” in this article mainly refers to topo-
digital surface models (DSMs). As a result, users may be graphic airborne lidar, it also covers ALB, unmanned air-
unaware or dismissive of the potential causes of data arti- craft system (UAS), and airborne photon-counting lidar.
facts. These underlying issues, if unmanaged, could affect Specifically, the focus of this concise review is on three
the overall reliability of any subsequent analysis based on types of data artifacts (see Table 1 and Figure 1):
the compromised data. 1) Data voids and holes appear as blank regions that fre-
System manufacturers and service providers make every quently occur in most topographic airborne lidar point
effort to collect and produce the highest quality of airborne clouds. Data voids appear because of occlusions (i.e.,
lidar data. While the lidar community puts strong emphasis shadows caused by elevated objects nearby), laser drop-
on the quality assurance (QA) and quality check (QC) of outs found on water surfaces, ground objects with an
data [5], [6], most of the QA and QC efforts are in modeling extremely low reflectance (i.e., dark objects), swath gaps
system biases, which, in turn, rectifies only geometric accu- between data strips, or ground filtering. The presence of
racy. Collected lidar data have unavoidable artifacts in the data voids may cause unpleasant triangular meshes in
form of speckle and/or random noise, anomalies, defects, the resulting DEMs or intensity images, which, in turn,
systematic biases, spikes, pitfalls, stairsteppings, periodic affects the visual quality, topographic analysis, semantic
stripings, corduroys, and so on, especially if users examine labeling, and 3D model reconstruction.
the data at a sufficiently fine scale. Improvements in pulse 2) Stripe artifacts are commonly found in the resulting
repetition frequency (PRF) and a finer laser pulsewidth [7] intensity image. If stripes appear on an individual lidar
further drive a significant increase of point cloud density data strip, this is mainly attributed to misalignment
[8] and ultimately magnify these artifacts. These defects between the axes of the laser transmitter and receiver,
not only result in unpleasant visual effects but also interfere where this scenario is characterized as intensity band-
with the subsequent analyses. Commercial remote sensing ing. Due to signal attenuation at swath edges, i.e., long
and photogrammetric software tools usually bundle with range and a large scan angle, intensity striping may ap-
data preprocessing modules for radiometric/geometric pear when combining two overlapping lidar data strips.
correction, noise removal, image filtering, and spatial en- Similarly, striping may also appear in the overlapping
hancement. However, software tools for the airborne lidar region of the resulting DEM if the respective lidar data
community are comparatively more rudimentary, with the strips are not properly georeferenced and/or calibrated.
exception of a few prototypes created by researchers, e.g., 3) Random/speckle noise appears both in intensity images
LAStools [9], OPALS [10], and lidR [11]. Most commercial and DEMs. Noisy random data points can be caused by
software tools for lidar lack preprocessing functionalities backscattered returns from unwanted floating objects

TABLE 1. AIRBORNE LIDAR DATA ARTIFACTS AND THEIR EFFECTS ON THE DERIVATIVES.

LIDAR DATA DATA VOIDS AND HOLES STRIPES SPECKLE/RANDOM NOISE


Point clouds • Occlusions/shadows • Imperfect instrument calibration • Floating objects (e.g., clouds and birds)
• Laser dropouts and alignment • Bragg scattering and specular reflection
• Swath gaps • Banding (i.e., axial displacement (e.g., water surfaces)
• Objects with low reflectance ­between e­ mitter and receiver) • Solar background noise (e.g., photon
• Dark objects/weak targets • Signal attenuation at large incident counting)
• Ground filtering angles and long ranges • Range walk error/timing jitter
Intensity image • Huge triangular facets • Periodic striping in individual strips • Heterogeneous intensity on
• Missing intensity metrics • Striping found at swath edges in ­homogeneous surfaces
• Inaccurate understory information ­overlap region • Intensity pits
DEM/DSM • Triangular facets • Periodic striping in the • Craters/divots
• Discontinuous slope overlapping region • Blunders/spikes
• Inaccurate estimation of normal • Pits/spikes in canopy height models

22 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


(e.g., aerosols, rain, clouds, or birds), timing jitters, or OCCLUSIONS/SHADOWS
range walk errors (RWEs). Moreover, photon-counting An occlusion is a hindrance caused by a shadow cast by a
lidar is sensitive to solar background noise, resulting in nearby elevated object or terrain undulation. The line of
noisy returns. Forest scientists often suffer from data sight from the lidar system to the occluded region is thus
pits appearing in lidar-derived canopy height models blocked, with no illumination of the laser footprints, lead-
(CHMs). All these conditions ultimately lead to the ap- ing to data sparsity and voids
pearance of spikes and blunders in DEMs. Bragg scatter- [see Figure 2(a)]. Common
ing of lasers on water surfaces also causes speckle noise occlusions can be visualized
DATA VOIDS, HOLES, AND
in the resulting intensity images. The noise-induced in- as a shading effect caused by
tensity inhomogeneities pose serious challenges to us- nearby elevated objects. An- GAPS ARE AMONG THE
ing intensity data for mapping water surfaces. other example of occlusion ARTIFACTS THAT ARE
is shown in Figure 3, where a COMMONLY FOUND IN
VOIDS, HOLES, AND GAPS number of occluded regions AIRBORNE LIDAR POINT
Data voids, holes, and gaps are among the artifacts that on the terrain can be found CLOUDS AND ADVERSELY
are commonly found in airborne lidar point clouds and with shadows cast from atmo- IMPACT DATA DERIVATIVES
adversely impact data derivatives in general. These defects, spheric smoke and clouds. On
IN GENERAL.
apart from those present in the original data, may also be the other hand, occluded data
induced by ground filtering of aboveground features for ex- voids can also be found with
tracting the bare Earth. The presence of data holes results in sizes as small as the height of
undesired triangular facets after conducting a surface inter- a typical curbstone [15]. The shadow region s blocked by an
polation, regardless of a DEM or an intensity image. These elevated object with a known height hb can be estimated
unfavorable effects are particularly prominent in the vicin- with respect to the lidar system’s flying altitude H and scan-
ity of data discontinuities. For instance, the surface normal ning angle [16], using the following equation:
being estimated at the edge of a hole may tilt into a wrong
h #d
direction. Therefore, such data loss or missing information s = Hb - h b (1)
b
is considered a major obstacle to subsequent analyses and
modeling. Various airborne lidar specifications, includ- where db refers to the horizontal distance between the na-
ing the United States Geological Survey (USGS) lidar base dir point and the object (see Figure 2). If prior knowledge
specification, also stress that “a data void is considered to of the study environment is available, better mission plans
be any area greater than or equal to (4 # Aggregate Nomi- could be devised in advance to reduce shadows by having
nal Pulse Spacing (ANPS))2, which is measured using first multiple overlapping strips with parallel and crossing flight
returns only” [13], [14]. In general, there are four possible paths. In some special cases, the appearance of occluded
causes of these data voids, holes, and gaps. These include data voids causes discontinuities of outlines of objects and
1) occlusions/shadows, 2) laser dropouts on water surfaces, thus hinders the automation of detection and 3D recon-
3) swath gaps, and 4) weak targets and objects with a low struction processes.
reflectance. Each of these is discussed in the following sec- Various studies address such a challenge under differ-
tions, covering the causes, impacts, and possible solutions ent scenarios, especially in urban infrastructure, for in-
currently available. stance, mapping curbstones and roadsides where there are

Adverse Potential
Effects Causes
Airborne Lidar

Random Stripes/ Speckle Void/ Cloud/Rain Ground System Moving


Noise Banding Noise Hole /Smoke Condition Settings Object

FIGURE 1. Airborne lidar data artifacts’ causes and adverse effects.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 23


Power
Occluded
Lidar Flight Direction
Region

Time
H
First Return

hb Second Return
Nadir s

Front View Top View Third Return


db s
(a) (b)

FIGURE 2. (a) The occluded region, s, is blocked by an elevated object. Here, H is the flying altitude, db is the horizontal distance between
the nadir point and the object, and hb is the height of the object. (b) A laser beam penetrates the tree crown, resulting in multiple returns.
The final return may cause an appearance of a data void due to occlusion caused by the overstory and/or a data pit in a CHM.

occlusions caused by on-road parked


vehicles [15], automatic rooftop ex-
traction with very close canopies
situated nearby [17], [18], road seg-
ments represented by segregated li-
dar data points due to the blockage
of objects found right next to roads
[19], and reconstruction of a lower-
level structure occluded by the upper
bridge in a multilayer-interchange
highway [20] (see Figure 4 for an ex-
ample). Most of these studies focus
on object detection and 3D recon-
(a) (b) struction, and thus, they respectively
propose corresponding workflows
and algorithms to restore the missing
FIGURE 3. (a) Smoke and clouds appear during airborne lidar data collection, resulting in
parts of the structure, based on prior
occlusions on the terrain. (b) Unnatural artifacts (dark patches) also appear in the resulting
knowledge of the objects’ geometric
intensity image.
characteristics. The occluded regions
in these scenarios all have notable
visual effects of the missing regions.
Aside from urban infrastructure,
occlusion caused by tree canopies,
though inducing less visual salience,
may impose a burden for retrieving
understory vegetation information.
If a laser footprint has a projected
area larger than the foliage, the la-
ser energy, at one point, is backscat-
tered from the first instance, and
the rest of it propagates down the
canopy. The footprint may subse-
quently reach the twigs, branches,
(a) (b)
and trunk and finally land on the
ground, resulting in a number of
FIGURE 4. (a) Occlusions found in the lower structure/part of (b) a multilayer highway. backscattered laser echoes [see

24 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Figure 2(b)]. Therefore, the overstory reduces the possibil- rotatingprismsoroscillatingmirrors(seeFigure5).Althoughthe
ity of laser penetration to the understory, and such an is- USGS mentions such unavoidable data voids in water re-
sue appears to be particularly serious on closed canopies. gions in the lidar base specification [13], these missing re-
As a result of the occlusion, the ground vegetation receives gions, if left aside, may adversely impact visualization and
considerably fewer laser pulses compared to the upper subsequent analyses.
canopy layer, inducing data voids and holes. This causes Depending on the laser wavelength and instantaneous
a number of adverse impacts. First, the generation of bare water conditions, a laser pulse usually undergoes (Fresnel)
ground in dense forest regions becomes stimulating. Even reflection, refraction, and (Bragg) scattering when it inter-
if a lidar end user applies advanced ground filtering and acts with a water body. A near-
interpolation algorithms, the resulting digital terrain infrared (NIR) laser, operated
model may be inaccurate and suffer from a loss of rugged at 1,064 nm, may suffer from
EVEN IF A LIDAR END USER
details [21], [22]. The situation gets even worse on moun- pulse absorption if the wa-
APPLIES ADVANCED
tainous terrain with steep slopes [22], [23]. In addition, ter is clean and the surface
GROUND FILTERING AND
forest scientists may encounter difficulties in computing is calm. On the other hand,
lidar metrics and predicting the substantial numbers of NIR laser undergoes Bragg INTERPOLATION
understory vegetation elements when there is a sparsity of scattering with instantaneous ALGORITHMS, THE
data points [24], [25]. water waves or turbulence RESULTING DIGITAL
Apart from these notable effects, the radiometric mea- [37] or water with a high de- TERRAIN MODEL MAY BE
surement of the lower vegetation layer may suffer from the gree of suspended sediments. INACCURATE AND SUFFER
transmission loss of laser energy due to canopy occlusion. Green laser (e.g., operated at FROM A LOSS OF RUGGED
The intensities of understory vegetation are underpresent- 532 nm) backscatters from
DETAILS.
ed since the laser signals have been intercepted due to ear- the water surface, refracts and
lier backscattering with the overstory. Indeed, it is mostly penetrates along the water
impossible to correct or normalize the intensity values of column, and finally reaches
the second to the last returns, especially for understory the seabed (depending on the water depth). That is the rea-
vegetation, based on the discrete-return or multiecho li- son why ALB systems, such as Optech SHOALS or CZMIL,
dar system [26]. Unless the recorded waveform echoes are are usually equipped with dual lasers with a NIR channel
available for radiometric correction [27], an accurate esti- capturing the water surface and a green channel mapping
mation of spectral reflectance of understory vegetation can the water bottom.
hardly be achieved. Laser dropouts usually occur at large incident angles
Unlike the scenarios found in urban areas that may be [35], [36] and can be explained by a number of reasons:
resolved by multiple overlapping data strips, the occlu- 1) the recorded laser energy is usually low at a large in-
sions caused by forest canopies are impossible to avoid. In- cident angle, as attributed to the laser backscattering
creasing the PRF and restricting the
swath with a narrow off-nadir angle
may allow better laser penetration 4096 4096
deeper into canopies [26], [28] but
Intensity
Intensity

only to a limited extent. Possible so-


lutions are to either conduct an air-
borne lidar measurement during the
0 Laser Dropouts 0
leaf-off season [24], [25], [29], [30] or 50 at Swath Edge 50
Elevation (m)

Elevation (m)

compensate the understory coverage


with multiple ground measurements
using mobile or terrestrial lidar sys-
tems [25], [31], [32]. 0 0
–20 0 20 –20 0 20
Scan Angle Scan Angle
LASER DROPOUTS ON Land
Water
WATER SURFACES
Topographic airborne lidar may record
Scan Line Profile Scan Line Profile
null returns when scanning over water
bodies, resulting in holes appearing Swath Width
in point clouds [33], [34]. These are
commonly known as laser dropouts
[35], which are found to be evident in
data collected by linear array scanners, FIGURE 5. NIR lasers usually experience laser dropouts at swath edges, with a high variation
i.e., lidar systems operated with either of intensity values backscattered on the water surface along each scan line [36].

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 25


geometry [38]; 2) laser pulses experience Fresnel reflec- To further take advantage of laser dropouts found at
tance, which can be as low as 2% [39]; and 3) atmospher- large incident angles, (2) can be further revised as
ic turbulence and high near-shore wind speeds further v N
SLIER = v zI # cos i # n ss (3)
attenuate the laser energy, leading to the failure of the
laser signal to reach a threshold to be processed [39]. where i refers to the scan angle, N s equals the maximum
With the combination of these effects, null returns may number of points found among the scan lines in L, and n s re-
be found at swath edges and occasionally appear in the fers to the number of data points found in the current scan line
close-to-nadir region (see Figure 5 for an illustration). s. If the swath of s completely covers a water region, the laser
Apart from these, the backscattered laser pulses may dropouts cause fewer data points, leading to a small n s value.
have a high variance of signal strength in these water re- This can further boost the SLIER, causing higher values to be
gions, resulting in some form of speckle noise in the in- computed on the water surface and lower SLIER values associ-
tensity images (discussed in the “Speckle Intensity Noise ated to the land. Sample water surface data points can be locat-
on Water Surfaces” section). ed with higher SLIER values (e.g., the top 10% within L). Then,
The presence of data voids in water regions causes un- a virtual water surface can be defined based on the elevation
natural triangular facets when generating DEMs. Indeed, of these sample water data points [36]. These data points can
the USGS has emphasized replacing these defects in the further serve as training datasets for classifiers to distinguish
resulting DEM with a flat virtual water surface [13]. This between land and water regions [43]. Figure 6 provides an ex-
process is called hydroflattening. Existing solutions mainly ample of the SLIER and the result of water mapping.
rely on either manual digitization or the incorporation of With the SLIER aiding in water identification, those data
existing break lines along coastal regions and river banks voids caused by laser dropouts can be located and compen-
[40], [41], depth sounding data [34], and instantaneous sated. Recently, [44] proposed two scan line void-filling algo-
water surface measurements via boat surveys [42] to locate rithms to add artificial data points in the water gaps in the
water regions, followed by hydroflattening. close-to-nadir region and swath edges. The algorithm han-
The shortcomings of relying on manual intervention or dling the close-to-nadir region aims to add these artificial
a semiautomatic process have been addressed lately. Water data points if the distance between two consecutive points
data points can be first extracted based on an airborne lidar along the scan line is larger than the mean point spacing (or
ratio index, the scan line intensity–elevation ratio (SLIER) ANPS) via interpolation. The second algorithm first esti-
[36]. The SLIER reaps the benefits of how a laser interacts mates the maximum swath width w within L and then adds
with the water body. Along each scan line s, the water sur- artificial data points to the swath until the length of s reaches
face usually has a relatively lower variance in terms of eleva- w. Assuming that the elevation of these artificial data points
tion z but a higher degree of fluctuation in terms of intensity equals the mean value of the water data points, the entire L
I for a given lidar dataset L. With 6s ! L, the SLIER can be and these artificial data points are combined to generate a
defined as DEM. The resulting process leads to a hydroflattened water
v surface, and those unpleasant triangular facets disappear in
SLIER = v zI . (2) the resulting DEM accordingly (see Figure 7 for an example).

1,550 1,064 532 1,550 1,064 532 1,550 1,064 532 1,550 1,064 532
(nm) (nm) (nm) (nm)
Elevation (m) Intensity SLIER Land–Water Classification
138 146 0 4096 Low High Water Land
(a) (b) (c) (d)

FIGURE 6. (a) Laser dropouts found in a water region, particularly at the swath edges. (b) A high variance of the intensity values on the
water surface. (c) High SLIER values computed on the water region, particularly on laser channels of 1,064 and 1,550 nm. (d) The use of the
SLIER for land–water classification [36].

26 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Edge
Nadir

(d)

378
Height (m)

0 0.2 km
243

(a) (b) (c) (e)

FIGURE 7. (a) Laser dropouts cause data voids along an inland river. (b) The implementation of a scan line void-filling algorithm implement-
ed at the close-to-nadir region and swath edges. (c) The resulting hydroflattened DEM. (d) The corresponding 3D view of the original DEM
with unpleasant triangular facets along the river. (e) The hydroflattened DEM [44].

SWATH GAPS not be the case for systems equipped with rotating prisms.
Another explanation of void appearance can be attributed The change of swath width also happens if the flight alti-
to the gaps found between swaths. In common practice, a tude changes during the flight mission. This may also lead
15%–50% overlap between adjacent data strips is a rule of to the formation of a gap between two swaths. Second, data
thumb during mission planning [13], [14], [45]. Combin- gaps may be found between two data strips, which can be
ing a number of partially overlapping data strips can lead ascribed to the accuracy of the GNSS and inertial measure-
to a seamless mapping of the study region. Putting aside ment unit (IMU). Nevertheless, data gaps may be particu-
the possibility of poor flight planning, adjacent flight lines larly obvious along the vertical direction in the overlapping
may have gaps due to various internal and external factors. region. Such a case also would lead to stripe artifacts in the
Instantaneous flight conditions, such as air turbulence or a resulting DEM. Strip adjustment should therefore be ap-
change of heading direction, likely cause undesired move- plied to remove system errors and biases so as to improve
ments of the aircraft, resulting in a crooked flight trajec- the geometric accuracy. Details can be found in the “Stripe
tory. It is a common problem not only in an airborne lidar Artifacts in Digital Elevation Models” section.
system but also in a UAS platform. As demonstrated in Fig-
ure 8, a portion of the airborne lidar data strip is twisted
with a curvy swath due to possible turbulence. A notable
void is thus found between the swaths, a scenario consid-
ered an unintentional cause of data voids. On the other
hand, a change of system settings or system error/bias may
also lead to the presence of swath gaps. This situation can
be further broken down into the following two scenarios.
First, a significant change of terrain height, particularly
on rugged mountainous terrain, may alert the swath of the
airborne lidar if the flying altitude remains unchanged.
Combining two adjacent strips under this circumstance
may lead to the appearance of a swath gap. Although
topographic airborne lidar systems operated with oscillat-
ing mirrors are capable of maintaining a consistent swath FIGURE 8. A swath gap with an annulus-shaped sector due to a
width via adjusting the degree of angular rotation, this may crooked flight trajectory.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 27


DARK OBJECTS/WEAK TARGETS sand and gravel samples that were thoroughly soaked. Wa-
Objects with a low reflectance, which can be categorized as ter standing on rooftops may cause a lack of data points
dark or weak targets, may result in less backscattered laser that further hinders the subsequent automatic 3D recon-
energy traveling to the receiver. Energy absorption by dark struction [49]. Such an influence has a greater overall con-
objects is not limited to water bodies. In some cases, the trol on hard objects, while the impacts of moisture have a
atmospheric absorption and scattering may further lead to limited effect on vegetative covers [50].
significant energy attenuation, resulting in a null return re- Indeed, a lidar system is usually configured to expect
corded by the system. As a re- both weak returns and retroreflections. Nevertheless, ex-
sult, the cause of such a data treme high-energy returns may saturate the laser receiver,
void depends on two issues: which also looks for weak returns from dark objects at the
OBJECTS WITH A LOW
instantaneous ground con- same time. Some of the airborne lidar systems are equipped
REFLECTANCE, WHICH CAN ditions and system settings with automatic gain control (AGC), which can leverage the
BE CATEGORIZED AS DARK (i.e., gain control and emitted varying strengths of backscattered laser returns and fly-
OR WEAK TARGETS, MAY laser power). ing altitudes. The airborne lidar ramps down the detector’s
RESULT IN LESS Various studies report null AGC when surveying retroreflective targets, while weak
BACKSCATTERED LASER returns or a comparative spar- targets or dark objects drive up the receiver’s gain [51]. De-
ENERGY TRAVELING sity of point density in ob- spite that, the mechanism of AGC, though it is commonly
jects with a low reflectance or known as a linear system, is considered a “black box” to the
TO THE RECEIVER.
made with highly reflective end users. There have been several attempts to utilize the
materials. Transparent struc- AGC mechanism and the recommended intensity correc-
tures, such as glass, may also tions to retrieve the (psuedo)spectral reflectance for a Leica
cause a sparsity of point clouds found on rooftops since ALS50 with the AGC setting [51], [52].
laser beams likely penetrate through the glass and back- It is commonly known that a lidar system increases the
scatter from the objects underneath. This makes automatic PRF to look for an improved data density. The increase, in
building or rooftop detection and segmentation challeng- turn, causes a decrease in emitted laser pulse energy [53]
ing [17]. Asphalt roads or tar-coated rooftops, such as in the as a side effect and further reduces the possibility of weak
example in Figure 9, absorb laser pulses and thus lead to target detection. Therefore, some airborne lidar systems
null returns and data voids. may produce a varying laser signal strength to leverage
Objects having a high degree of moisture and wetness the PRF. For instance, the Optech Titan is equipped with
(i.e., after a rainfall) also likely hinder laser backscattering. three laser channels with a linear detector response. As a
Indeed, a negative correlation exists between surface mois- result, the recorded backscattered intensities are linear
ture (or wetness) and backscattered laser energy, and this across the entire scale. However, this particular system has
phenomenon is particularly obvious in NIR lasers. Surface an intensity transfer function [36], which can change the
moisture further causes a reduction of the intensity dif- power ratio of the green (532-nm) and short-wave infrared
ference among surface objects, for instance, sand rich and (SWIR) (1,550-nm) channels with respect to the infrared
shale rich, as reported in [46], which makes object segmen- (1,064 nm). The power ratio of the green channel can be
tation challenging. The moisture may reduce the backscat- twice as much as that of the infrared channel, regardless of
tered reflectance (after conducting radiometric correction) the PRF settings. The SWIR channel has a varying power
by 7%–27%, comparing wet and dry stretches of road sur- ratio, depending on the PRF. The SWIR channel can gener-
faces [47]. In another experiment [48], the backscattered ate a higher laser power ratio of up to three times with the
reflectance was further reduced by 30%–50% for different reduction of the PRF [36]. Although an increased power ra-
tio can improve the capability of target detection and mini-
mize the presence of data voids, it may also induce a high
saturation of signal returns that appear like “speckle noise”
in the resulting intensity image. More details are covered
in the “Speckle Intensity Noise on Water Surfaces” section.

SOLUTIONS AND FUTURE DIRECTIONS


With respect to the previously mentioned causes, detection
of data voids becomes critical and subsequent void-filling
strategies should be implemented to resolve these defects.
Indeed, good mission planning can alleviate the effects
of occlusions and swath gaps. Various lidar specifications
recommend a 15%–50% overlap between adjacent flight
FIGURE 9. An image of concrete rooftops and asphalt roads that strips [13], [14], [45]. Also, seamless mapping should in-
has a sparsity of data points and holes. clude multiple scans of the study area, with different flight

28 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


directions so as to minimize the effects of occlusions and on the ground or not [57]. Despite a recent study on terres-
swath gaps. Very often, end users remove the data points trial lidar that classified occlusions and laser dropouts on
with large incident angles (215°) since these off-nadir terrains [58], research in point cloud void classification re-
data points suffer from significant laser energy attenuation, mains in an early stage. A specific filling mechanism should
resulting in stripe artifacts in the intensity image. If these be designed for gradient changes of water slopes, such as at
data points are incorporated in subsequent analyses, radio- waterfalls or downstream locations [59]. For those artificial
metric correction should be implemented. More details can objects, point cloud completion or generative adversarial
be found in the “Intensity Striping at Swath Edges” section. networks (GANs) can be introduced to repair the missing
Detection of data voids within airborne lidar point regions if intensive training data are available [60], [61],
clouds can be made based on certain assumptions or [62], [63]. Persistent homology, which discerns topological
threshold techniques. Zhou and Vosselman [15] identified features from the point cloud through algebraic theories,
the gaps of curb lines caused by the occlusions of on-road can also be introduced to determine significant holes from
vehicles, and only those gaps with a predefined length the topological noise [64]. All in all, a comprehensive work-
(based on trial and error) were bridged. Yan [44] used the flow for automatic void detection, classification, filling/in-
time stamp and distance between two consecutive points painting, and reconstruction is desired.
along the scan line as a criterion to look for laser dropouts
on water surfaces. Feng et al. [18] fitted a minimum bound- STRIPE ARTIFACTS
ing rectangle to estimate whether a rooftop suffers from Stripe artifacts are also referred to as rippling, periodic band-
partial occlusion caused by adjacent trees. Elberink and ing, stairstepping, or corduroy in the literature. Such a defect
Vosselman [49] adopted the rooftop area, orthogonal dis- may not be spotted out until one interpolates the airborne
tance, and shortest distance with corresponding predefined lidar point clouds to an intensity image or a DEM. An in-
thresholds to assess whether there exists a lack of data dividual lidar data strip may suffer from intensity banding
points or gaps. Apart from making use of the raw lidar data due to the misalignment between the field of view (FOV) of
properties, various computational geometry and computer the receiver and emitter. When combing multiple overlap-
vision methods are proposed via detecting and filling the ping lidar data strips, intensity striping may appear at swath
holes in triangular meshes [54], [55]. edges. This can be explained by the significant attenuation
After gap/void detection, the design of a corresponding of laser beams with longer ranges and larger incident angles.
filling mechanism highly depends on the type of voids/ Geometric misalignment between two overlapping swaths
gaps being found. A flattened water surface with a specific may also lead to a rippling effect in the resulting DEM.
elevation value can be assigned after compensating the la-
ser dropouts in the water region [44]. Studying the saliency INTENSITY BANDING
features, such as normal and curvature, can determine the If an airborne lidar data strip has a notable systematic strip-
corresponding way to close the gaps found on rooftops ing pattern found in the interpolated intensity image, the
with a sparsity of data points [17]. On the other hand, a lidar data highly likely suffer from intensity banding issues
sigmoid function along a spline can be used to reconstruct (see Figure 10). Intensity banding occurs only if the lidar data
the distorted curb lines caused by occlusion [15]. There- strip is collected by an airborne lidar system operated with
fore, the filling and inpainting processes are object and an oscillating mirror that has an incorrect optical alignment
shape dependent. [65], [66]. Therefore, the intensity banding effects caused by
Most of the time, data voids with-
in an airborne lidar dataset can be
detected and located. Nevertheless,
it is hard to decide the corresponding
filling mechanism, especially if the
dataset contains various voids found
in water regions, terrains, and objects
with low reflectance as well as at
steep slopes [56]. As a result, a future
direction should focus on a machine
learning approach to determine and
classify the types of voids being
found. Classification of data voids/
gaps is critical since it may affect the
inference of change detection. Data (a) (b)
gaps caused by building occlusions
should be considered since doing so FIGURE 10. (a) Stripe artifacts found in an intensity image generated by an individual lidar
may infer whether there is a change data strip. (b) The intensity image after the removal of the banding effect.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 29


stripe artifacts are always along the across-track direction. low-pass filters, it also oversmoothens the intensity image,
Also, one should note that the term “intensity banding” causing a blurry effect. On the other hand, [65] proposed to
here does not share the same meaning as “banding” in lidar adopt histogram matching to normalize the intensity histo-
point sampling, where a condensed or stretched point distri- grams collected by the two opposite scans.
bution is found along the data strip, due to a change of the Recently, lidar scan line correction (LSLC) was proposed
aircraft’s speed during the survey [67]. in [66] to adjust the radiometric misalignment found in the
Intensity banding is mainly attributed to the misalign- two opposite scans within a lidar data strip. The LSLC is,
ment between the axes of the laser transmitter’s and in essence, a polynomial function, which normalizes all
receiver’s FOV. Under normal circumstances, the two the scan lines of a particular direction having a lower mean
axes should be properly aligned and parallel to each intensity (i.e., I l) with reference to the opposite, which has
other (see Figure 11). In this case, the backscattered laser a higher mean intensity value (i.e., I h). The model param-
beam should be completely captured within the receiver’s eters, including both the intensity and scan angle i, can
FOV. Due to improper calibration, mechanical failure, or be estimated by first pairing up the nearest points of the
a collision during installation or delivery, the oscillating two scans and solving the polynomial function by using an
mirror may cause the laser beam’s axis to lean away from iteratively reweighted least-squares adjustment, as follows:
or bend toward the receiver’s axis. As a result, the back-
scattered laser beams, which are collected in the scanning I h = a 0 + a 1 I l + a 2 i + a 3 I l i + a 4 I l2 + a 5 i 2 . (4)
direction, with the laser beam’s axis leaning away from
the receiver’s axis, may partially fall outside the receiver’s All the scan lines suffering from the banding are cor-
FOV. It thus causes a consistent attenuation with the col- rected using the LSLC together with the estimated param-
lected laser signal strength in this particular scanning di- eters. The resulting intensity image after the LSLC should
rection and ultimately leads to the stripe artifacts found be striping-free. The authors of [66] reported a reduction
in the lidar dataset. Interested readers are referred to [66]. of the coefficient of variation (cv) by 13%–80% on three
Currently, a number of advances have been made to re- monochromatic lidar datasets suffering from light, mild,
solve intensity banding. Image filters are one of the com- and extreme banding effects after implementing the LSLC.
mon ways to handle this issue since the stripe artifacts are The work in [69] further applied the LSLC on the two laser
being treated as systematic intensity noise. Nobrega et al. channels (1,064 nm and 1,550 nm) of the multispectral air-
[68] proposed an anisotropic diffusion filter, which is capa- borne lidar data collected by an Optech Titan to remove the
ble of smoothing the intraregional areas preferentially over intensity banding, with a reduction of cv by 4.8%–66.9%
interregional areas. This not only removes the striping but and 0.2%–23.1%, respectively.
also preserves the edges of objects. However, although the If point cloud density is not a major concern, some of
filter can successfully achieve denoising and outperform the existing studies simply abandon the lidar scan lines of a

Transmitter Left-to- Right-to-


Receiver Right Scan Left Scan Mirror Rotates
System’s Side View

Mirror Rotates in a Constant at a Lower Speed


Speed Within this Zone Within this Zone
Receiver’s Laser
Axis Beam
Mirror Decelerates

Transmitter’s Laser
Transmitter’s Axis is Not Parallel
Receiver’s FOV Beam
Axis to the Receiver’s axis
Falls Receiver’s FOV Mirror Accelerates
Cross Section of Outside
Receiver's FOV Receiver’s
Cross-Section View

Laser
FOV
Beam

In normal case, A portion of laser The entire laser Flight


the laser beam beam falls outside beam falls within Direction
completely falls the FOV in case the FOV in case
within the the transmitter’s the transmitter’s
receiver’s FOV. axis leans away axis leans toward Swath Edge Nadir Swath Edge
from the receiver’s the receiver’s axis.
axis.
(a) (b)

FIGURE 11. The imperfect alignment of a laser beam’s axis and receiver’s FOV in a specific direction, causing intensity banding [66].

30 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


particular scanning direction that have a consistently lower physical parameters and/or overlap functions. This can lead
intensity value than the opposite ones. This can be achieved to a more robust solution to resolve intensity banding.
by computing the mean intensity value of the respective two
scanning directions by using the “scan flag” field in the LAS INTENSITY STRIPING AT SWATH EDGES
file. For instance, Okhrimenko and Hopkinson [70] omitted Stripe artifacts at swath edges can be explained by laser
half of the lidar dataset (i.e., those scan lines suffering from attenuation (see Figure 12). When a laser beam travels
intensity banding) to compute the spectral vegetation indi- through the atmosphere and finally reaches the ground,
ces when using an Optech Titan. The resulting scan line pat- the laser beam undergoes aerosol/molecular absorption
tern would be similar to those collected by a rotating prism. and scattering. As a result, the longer the range, the great-
In [71], the authors adopted a similar approach in their radio- er the attenuation of the emitted and backscattered laser
metric analysis but retained the entire dataset for geometric/ echoes. A laser beam reaching the swath edge implies a
spatial analysis of canopies. Nevertheless, Goodbody et al. large incident angle, which also significantly reduces the
[72] also argued that the reduced point cloud density, par- backscattered laser power. Combining data points with la-
ticularly in the green channel (532 nm) of an Optech Titan, ser attenuation occurring at the swath edges with the data
affected their voxelization results in computing the vegeta- points from the nadir of another strip thus causes the stripe
tion indices for modeling forest inventory attributes. artifacts notable in the overlapping region of the intensity
Although the intensity banding issue has been ad- image. To physically model the laser attenuation, the radar
dressed in the literature lately, all the previously mentioned (range) equation [38] is used to described the received laser
methods have their own drawbacks. Specifically, the band- power with respect to various parameters:
ing issue is prominent at the close-to-nadir region but more
subdued at the swath edges [66]. This can be explained by Pt G t v r
Pr = h atm h sys (5)
the speed of the oscillating mirror, which decreases when 4rR 2 4rR 2 D 2
approaching the swath edges, causing a high overlap fac-
tor between the laser beam and the receiver’s FOV. On the where Pr refers to the recorded laser power that is typically
other hand, the oscillating mirror is rotating at a high speed scaled to an 8- to 16-b intensity value, Pt is the transmitted
when approaching the nadir. As a result, the overlap factor laser power, R is the laser range, G t refers to the antenna’s
between the laser beam and the receiver’s FOV decreases. gain factor, D is the aperture diameter, and h sys is a trans-
Therefore, image filtering and histogram equalization [65] mission factor that is system dependent. The atmospheric
may not be able to appropriately handle different levels attenuation factor, h atm, can be modeled based on the Beer–
of banding in these regions. Although the LSLC considers Lambert law:
the use of the scan angle as one of the correction param-
eters, the method relies on global parameter estimation h atm = e -2aR (6)
lacking in robustness and flexibility [73]. In the future, it
is recommended to propose a local parameter estimation where a is the extinction coefficient that can be determined
method within, say, a pair of scan lines or an individual re- by summing up the aerosol (Mie) scattering, Rayleigh scat-
gion along the across-track direction and to consider other tering, and aerosol/molecular absorption, all of which are

Top View Change of System


Overlapping Scan Setting
I = 28 (e.g., Flying Height)
I = 50 First Scan
Scan
Range I = 45 First Second First
Angle
I = Intensity Scan Scan Scan
Flying ~ 600 m ~ 600 m
Height ~ 600 m
Second
Scan
~ 300 m

Asphalt
Nadir Surface
Half of
Swath Width Footprint Striping at Overlapping
Striping Found in
Swath Edges Region
Swath Edge the Entire
Overlapping Region
(a) (b) (c)

FIGURE 12. (a) Laser attenuation found at long ranges and large incident angles, resulting in smaller intensity values at the swath edges
compared to those at the nadir region. (b) Striping found at the swath edges when combining two overlapping data strips. (c) Striping found
in the entire overlapping region, due to different flight altitudes.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 31


wavelength-dependent parameters [74]. The laser cross sec- ^ I n h being directly proportional to the R to the power of a
tion v in (5) can be formulated based on the assumption of factor f; i.e.,
Lambertian reflectance:
I n = I·a R k (9)
R f
r
v = 4rtA cos (i r)(7)
where R r refers to a reference range value that is commonly
where t is the (pseudo)spectral reflectance of the target ob- set to be the minimum or median range value within the
ject, A is the area of the projected laser footprint, and i r is collected lidar data strips and the value of f is commonly
the laser incident angle. two, according to the radar (range) equation [38]. Neverthe-
Indeed, radiometric calibration or intensity correction less, such a setting handles only simple land cover scenarios
models are mainly built upon the previously mentioned by monochromatic lidar systems with stable settings. Dif-
radar (range) equation, which describes the relationship ferent research studies have attempted to look for an opti-
between the emitted and recorded laser energy with re- mal f value through using a cross-validation approach with
spect to the range, incident angle, surface condition, at- a range of predefined values, e.g., an f between zero and
mospheric attenuation, and other system factors [75], [76]. four [52], [80], [81]. Indeed, the value of f varies with re-
Theoretically, the collected intensity data represent the spect to the study area (or ground objects to be surveyed)
amplitude of the laser backscattered signal strength after and specific lidar systems. It is highly unlikely to have a
linearization, and thus, the purpose of radiometric calibra- fixed f value to be applied to a certain lidar system or study
tion/correction aims to convert the intensity value into a area. Also, the use of the cross-validation approach (i.e., set-
relative (pseudo)spectral reflectance after considering the ting a range of f values to yield the lowest cv) to look for
preceding attenuation factors [74], [75], [77] through refor- an optimal f is not practical at all. Therefore, an automatic
mulating (5) as follows: approach, one that can estimate the correction parameters
from the overlapping lidar data strips, is desired, especially
1
I ? t· ·cos i·e -2cR . (8) for the increased number of laser channels of multispectral
R2
lidar systems.
Range normalization is also widely adopted to correct Recently, Yan et al. [69] reported two-stage correction
the intensity data (I) [78], [79], where it simply considers for multispectral airborne lidar data by first removing the
the effects of the range, without taking into account the rest intensity banding, using the LSLC in each of the individ-
of the parameters. This results in the normalized intensity ual data strips [refer to (4)], followed by an overlap-driven

(a)

Swath Edge

(b)

0 40 m

(c)

FIGURE 13. Two overlapping swaths of a multispectral airborne lidar dataset, one with a large scan angle at the bottom and the other with a
small scan angle (near the nadir region) at the top. The (a) original intensity, (b) original intensity + LSLC (removal of the intensity banding),
and (c) original intensity + LSLC + overlap intensity correction (further removal of the striping at the swath edge).

32 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


intensity correction (see Figure 13 for the removal of the coordinates ^ X G h of each range measurement ^ r h with re-
striping caused by the laser attenuation found at the swath spect to a known coordinate system ^ X 0 h, as follows [87]:
edges [44]). The overlap intensity correction is able to es-
timate the correction parameters of range, angle, and at- R 0 V
S W
mospheric attenuation from the overlapping data strips X G = X 0 + R ~, {, l PG + R ~, {, l R T~, T{, Tl R Si i S 0 W . (10)
S- ^r + Tr hW
through an iteratively reweighted least squares with a Hu-
T X
ber estimator to resist outliers. The removal of striping can,
in turn, improve not only the lidar data quality but also All the components (i.e., laser ranger, POS, and oscil-
subsequent analyses. lating mirror) in an airborne lidar system certainly have
Given that the lidar community is moving toward in the their own error sources [88], and neither is it the case that
use of intensity and/or waveform features for surface classi- these components can perfectly align to a physical common
fication and object extraction, a number of studies reap the origin (see Figure 14). The laser ranger may be intrinsically
benefits of radiometric calibration, range normalization, subject to systematic errors (i.e., Tr) in each of the range
and intensity correction, resulting in improved intensity measurements (r). The oscillating mirror, which distributes
homogeneity and classification accuracy [74], [75], [77]. On laser beams on the ground, with scan angle i, is embed-
the other hand, radiometric treatment seems to provide a ded with a scale factor S i, as shown in (10). In terms of the
marginal impact on lidar intensity-derived forest metrics physical alignment, the spatial offset and orientation dif-
[69], [82], [83]. Despite these well-developed techniques, ference between the laser ranger and the POS respectively
apparently, a stratified approach to intensity correction, correspond to the issues of the lever arm ^PG h and boresight
similar to topographic correction in satellite remote sens- angle ^R T~, T{, Tlh . Besides, the sampling rate of each com-
ing, is desired so that heterogeneous land covers receive ponent is different and difficult to precisely synchronize.
different levels of corrections for the laser range, incident These uncertainties and errors thus accumulate in each of
angle, and atmospheric attenuation [84]. the lidar data strips, leading to swath mismatch and striping
in the resulting DEM (see Figure 15). Striping in DEMs can
STRIPE ARTIFACTS IN DIGITAL
ELEVATION MODELS
Aside from the unpleasant triangu- Laser POS
lar facets found in the lidar-derived Scan Angle Ranger ω
DEM (refer to the “Laser Dropouts P⃗G
on Water Surfaces” section), the ex- ϕ
isting literature also reports the ap- Height Mirror k Uncalibrated
pearance of stripes in the resulting r⃗ Lidar Strips
X⃗0
DEMs, especially those collected by
small-footprint airborne lidar sys- Z0
tems. These stripes appear notable X⃗G Y0
on rooftops and flat ground, dis- With Strip
X0 Adjustment
rupting terrain analyses (e.g., the
computation of slope and aspect).
Stripe artifacts in DEMs intrinsically FIGURE 14. The system calibration of airborne lidar data strips.
imply an imperfect calibration and
improper alignment of the overlap-
ping data strips. Causes of these can
be explained by GNSS and IMU mea-
surement errors.
An airborne lidar system comes
with a positioning and orientation
system (POS), including the GNSS
and an IMU for direct georeferenc-
ing [85], [86]. The GNSS determines
the instantaneous 3D coordinates
of the aircraft, while the IMU col-
lects the angular motion or orienta-
tion, i.e., R ~, {, l, of the aircraft. For (a) (b)
a linear lidar system, the lidar equa-
tion thus provides a direct georefer- FIGURE 15. (a) Striping on a DEM. (b) A DEM generated by geometrically calibrated lidar data
encing solution to estimate the 3D strips based on a quasi-rigorous calibration procedure [87].

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 33


be resolved in two major ways depending on whether the solutions to resolve the striping in lidar-derived DEMs,
original point clouds or resulting DEMs are available. which ultimately improves the DEM quality for various ap-
If the original lidar data strips are available, strip adjust- plications, such as volcanic analyses [96].
ment offers a way to improve the geometric data quality
via minimizing the discrepancies found in the overlapping SPECKLE/RANDOM NOISE
region of the adjacent strips [89]. In general, the process Similar to stripe artifacts, noisy returns may appear to be
of strip adjustment begins via identification of common scattered in the original point clouds as well as their data
shapes (such as rooftops or planar patches) extracted from derivatives, such as intensity images and CHMs. 3D point
the point clouds or the derived DEM in the overlapping re- cloud noise can be removed by postprocessing algorithms
gion. Surface matching then determines and quantifies the based on spatial filtering, differences in elevations, or local
discrepancies between the two overlapping data strips. densities. Speckle intensity noise found on water surfaces,
A number of options are available for surface matching; although appearing to be difficult to handle, sheds light on
these include point-to-point analysis (e.g., the iterative clos- certain information for automatic water mapping.
est point), surface or linear feature matching [90], and so
on. Apart from using specific common shapes/objects for SPECKLE INTENSITY NOISE ON WATER SURFACES
surface matching, lidar intensity values can also serve as When an airborne lidar system surveys a water region, the re-
another constraint for surface matching and subsequent sulting intensity image may be corrupted by speckle noise in
adjustment, and this is found predominantly in reducing addition to the effects of laser dropouts (i.e., data voids), as
the horizontal discrepancy [91]. Finally, the strip discrep- previously mentioned (refer to the “Laser Dropouts on Water
ancy between two strips can be minimized by applying a Surfaces” section). Unlike passive optical remote sensing im-
3D geometric transformation of the strips (such as a 3D af- ages capturing a homogeneous reflectance over a water region,
fine transformation) or a correction to the previously men- the emitted laser beams undergo a series of scattering, reflec-
tioned sensor parameters in (10) (i.e., a system calibration tion, and refraction processes. When the incident angle is
approach). Interested readers are referred to [89]. almost perpendicular, specular reflection or even retroreflec-
If end users are unable to obtain the original lidar data tion occurs, causing a strong reading of the echo amplitude
strips, there are also several solutions to resolve the stripe [97]. Since the instantaneous conditions of water bodies keep
artifacts in the resulting DEMs. Albani and Klinkenberg changing, all these processes cause different levels of energy
[92] proposed a cross-smoothing algorithm with a prob- transfer, leading to a variation of the backscattered laser pulse
ability replacement function to constrain the elevation of strength [36]. In addition, if the gain control keeps changing
the striping in the DEM. Gallant [93] adopted a multiscale (as mentioned in the “Dark Objects/Weak Targets” section),
adaptive smoothing filter to aid in DEM noise removal with the collected intensity also varies even as the surveyed objects
varying characteristics. Image filters based on wavelet and/ remain the same [51]. All these effects result in the presence of
or Fourier transforms can help eliminate the stripe arti- speckle noise in the intensity image (see Figure 16). The draw-
facts in different orientations [94], [95]. All these provide backs of speckle noise can be summarized in various ways.

(a) (b) (c) (d)

FIGURE 16. Speckle intensity noise appears in different types of water regions, such as (a) inland rivers, (b) natural shores, (c) inland lakes,
and (d) rocky shores.

34 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


In optical remote sensing, multispectral image bands range can then be computed by considering the round-trip
can aid in deriving biophysical indices not only for the travel time and the speed of light. As a result, the accuracy
retrieval of vegetation parameters but also assessing water of range detection highly depends on the backscattered
conditions, e.g., the Normalized Difference Water Index echo’s amplitude. If a certain degree of high-frequency
(NDWI). With the invention of dual- and tri-channel li- noise, caused by fluctuations in air currents, for instance, is
dar systems, green and NIR laser channels can conceptu- superimposed on the backscattered laser echo, it may cause
ally serve such a purpose. However, Yan et al. [36] have an upward noisy artifact that preempts the actual signal.
proved that speckle intensity noise found on water sur- The TOF mechanism thus records a return earlier than it
faces degrades the quality of the derived NDWI used for should and recognizes it as a data point that is closer to the
delineating water bodies from land. Such an issue is par- system than its intended position. Such a scenario is named
ticularly serious when dealing with high waves, turbu- timing jitter [101] [see Figure 17(a)]
lence, and white caps because laser pulse returns tend to Depending on the target range, size, and reflectivity,
interact with the water surface via Bragg scattering [37]. the backscattered laser echoes may have similar shapes but
Apart from the inappropriateness of deriving the water with different amplitudes. As shown in Figure 17(b), al-
index, direct use of lidar intensity in a machine learning though the echoes’ peaks are identical to each other in the
classifier to extract water regions may suffer from the is- time domain, the echo with the higher peak amplitude has
sues of between-class spectral confusion and within-class an upward shift of the signal passing the detection thresh-
spectral variation [43]. This may result in salt-and-pepper old earlier than the one with the lower peak amplitude.
noise found in the classification results. Therefore, proper Similar to the scenario of timing jitter, the TOF mechanism
treatment of the intensity should be instituted prior to in- may recognize a return earlier than it should, resulting in
formation extraction. a noisy point hovering on the point cloud. Such a scenario
A number of attempts are reported to mitigate speckle is commonly named RWE [101]. The timing jitter problem
intensity noise through image filtering; these include the is indeed hard to solve since it is a random error. The RWE,
median filter [98] and mean filter [99]. However, these fil- on the other hand, can be treated as a systematic error that
ters may oversmoothen the entire dataset, in which most can be adjusted by considering the time offset between the
regions are, in fact, free of speckle noise contamination. pulse rise time of the transmitted and recorded echoes.
Incekara et al. [100] proposed to implement a mean-shift Adjusting the threshold over time can also account for the
segmentation on the intensity image for image smoothen- variation of target properties and atmospheric attributes.
ing since it maintains the edge of a water lake while elim- Apart from the aforementioned system-induced noisy
inating the noise. To the best knowledge of the author, a returns, airborne lidar systems are sensitive to instanta-
comprehensive solution that completely removes the inten- neous environmental and atmospheric conditions, result-
sity speckle noise on water surfaces is yet to be available. ing in unwanted returns recorded in the point clouds. Fire
Nevertheless, since speckle noise can already be used to dis- smoke and cumulus clouds appear as various clustered
tinguish water regions from other topographic features (the points located on top of the terrain surface (see the ex-
use of the SLIER by [36], as described in the “Laser Drop- amples in Figure 18). Although end users can adjust the
outs on Water Surfaces” section), other potentially useful point clouds by removing these obvious anomalies manu-
information that aids downstream analyses could, in prin- ally, the corresponding occlusion still causes the problem
ciple, also be harnessed. Our deepened understanding of
the nature of the noise from these efforts will eventually
establish the foundations for a more comprehensive solu- Floating
Point
tion of noise removal.
Threshold
3D POINT CLOUD NOISE
Power

End users of airborne lidar data may notice noisy returns


Power

hovering on the 3D point clouds. The drawbacks of these


anomalies, if not properly resolved, may cause strong con- t1 t2 Time
Emitted Backscattered
vexities in the data, leading to spikes in DEMs or heteroge-
Pulse Pulse
neous intensity in the projected images. The causes of these (a) (b)
defects can be ascribed to system and environmentally in-
duced factors, such as random/systematic errors of range FIGURE 17. (a) Timing jitter: a certain degree of high-frequency
detection and instantaneous atmospheric conditions, e.g., noise superimposed on the backscattered laser echo causes an up-
rain, snow, the solar background, and so on. ward noisy signal that advances the time, exceeding the threshold.
A time-of-flight (TOF) lidar system determines the (b) RWE: the leading edge detection of a backscattered laser echo
range of an on-ground object by measuring the arrival time influenced by the pulse amplitude and width that may result in an
of backscattered laser echoes if the amplitude of its lead- earlier record (t1) of the backscattered laser pulse. Regardless of
ing edge exceeds a specific detection threshold. The target (a) and (b), floating points may hover on the point cloud.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 35


(a) (b) (c) (d)

FIGURE 18. Noisy 3D returns caused by (a) smoke, (b) unknown flying objects, (c) cloud, and (d) solar background noise or rain.

of data voids (refer to the “Occlusions/Shadows” section). one may simply investigate the distance between any pair
Instantaneous rainfall or snow storms during airborne li- of closest points or construct a voxel grid to assess the lo-
dar surveys generate random scattered returns, which are cal density [7] so as to look for the outliers. If end users
challenging to remove manually. Similar scenarios have are unable to have the original point clouds on hand, a
been reported in photon-counting lidar systems, such as despiking algorithm, such as the despike virtual defores-
the Leica SPL-100, where the system’s detector is sensitive tation algorithm [106] or spike-free algorithm [107], can
to solar background photons [7]. As a result, the solar back- be applied to the triangular area network (TIN) for re-
ground noise contaminates the collected point clouds. Un- moving the spiky features and negative blunders as well
like the previously mentioned unwanted returns caused by as smoothening the terrain.
clouds or smoke, random noisy returns are challenging to
remove manually. PITS AND SPIKES IN CANOPY HEIGHT MODELS
There are two ways to remove the outliers when The penetrating property of lasers and the ability of li-
one faces the preceding situations. Users can perform de- dar systems to probe and reconstruct environments in
noising on the point cloud, based on different 3D spa- 3D have allowed scientists to adopt airborne lidar data
tial filters [102]; clustering approaches, such as principal to retrieve/understand forest structures and properties.
component analysis [103] and density-based spatial clus- Although airborne lidar systems have been proved to be
tering of applications with noise [104]; or deep learning superior to optical imaging sensors in fine-scale forestry
approaches [105]. Unlike those point clouds collected by studies [108], [109], the collected datasets have certain
low-cost sensors, the density of outliers, in most cases, is imperfections disrupting the retrieval of forest inventory
significantly lower than the point cloud itself. Therefore, parameters. Aside from the occlusion effects caused by
foliage and upper canopies, as mentioned in the “Occlu-
sions/Shadows” section, the retrieved CHMs often suffer
from randomly distributed data pits and spikes [see Fig-
ure 19(a)]. These unnatural pits cause a high fluctuation
of elevations in the CHMs, which may influence the es-
timation of tree heights, crown boundaries, basal areas,
and stand volumes or induce errors in treetop detection
[107], [110], [111].
Due to the natural variation of canopies and irregu-
larities in tree heights, the emitted laser beams usually
produce a number of returns from the upper canopies,
(a) (b) tree trunks, branches, and foliage before reaching the
understory or even hitting the ground [see Figure 2(b)].
FIGURE 19. (a) A CHM suffering from data pits. (b) The CHM after Occasionally, laser beams may not interact with any
the data pits are removed. overstory and simply penetrate deeply into the crown,

36 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


resulting in the first return being backscattered from the OTHER ARTIFACTS
lower canopy branches, foliage, or ground. These first re-
turns represent a significantly lower elevation than their UNWANTED POINTS PRESENT BETWEEN
immediate surroundings, and thus, they appear like TWO NEARBY OBJECTS
dark holes or negative blunders in the resulting CHMs. Occasionally, end users may notice the existence of unde-
Combining multiple overlapping data strips may further sired data points between two objects or surfaces that are lo-
worsen the issue of data pits since laser beams may “peek cated close to each other in discrete-return or full-waveform
under” different locations from multiple f light lines lidar data. As described in Figure 20, a set of points appears
[107]. The appearance of data pits can also be ascribed between the stiffener and plating located on a stainless steel
to the interpolation of CHMs for filling missing/empty T-bar. This set of data points forms a line, which is believed
regions [110]. to be the lidar’s line of sight. Nevertheless, these data returns
Forest modelers and scientists desperately look for are unlikely to be backscattered from the T-bar itself since
pit-free CHMs prior to the retrieval of forest parameters. the geometry of the bar does not have an inclined surface
Traditional image filters, such as mean or median filters, or plate supporting the stiffener and plating. Indeed, the
can work to some extent. Nevertheless, the kernel size is appearance of such defects can be explained by the limited
certainly a critical factor since data pits often appear in range resolution and pulsewidth of the lidar systems used.
various sizes. Also, such an image filtering approach may Common lidar systems adopt a laser pulsewidth rang-
oversmoothen all the elevations (i.e., pixel values) in the ing from 2 to 5 ns. This range converts to a distance of ap-
CHMs [112]. Among all the existing methods, the pit-free proximately 0.6 to 1.5 m (considering the speed of light
algorithm proposed by [111] is considered to be the most . 3 # 10 8 ms-1). A lidar system is unlikely to be able to dis-
favorable method to generate a pit-free CHM. Such an ap- tinguish two returns, or, technically, two objects/surfaces,
proach relies on a cascading workflow by first generating if they have an offset distance smaller than the previously
CHMs at different elevations (e.g., at 2 m, 5 m, and so on) mentioned range. This is mainly because the two backscat-
using all first returns and subsequently stacking all the re- tered echoes are closely mixed with each other, where the
sulting CHMs while taking the maximum height values signal decomposition method, such as the Gaussian mix-
[see the example in Figure 19(b)]. ture model, may return a solution with a single return that
Another way to tackle data pits involves pit detection results between these two objects/surfaces (see Figure 20).
and filling. It usually begins with ground filtering to delin- As a result, in the example of a T-bar, instantaneous data
eate the ground and nonground points. Then, the process returns are found within the T-bar, where the laser beam
carries on by ignoring the data pits and spikes to generate first partially illuminates the plating and the final return
the CHMs. A common approach simply looks around each comes from the stiffener. The mixed echoes backscattered
data point with a spherical kernel and assesses whether from these two surfaces result in a return between. Such a
the elevation of the surrounding nonground points ex- scenario also causes ALB challenges to mapping the seafloor
ceeds a certain threshold so as to conduct an interpola- of very shallow water regions [115]. Since it is nearly impos-
tion for CHM generation [113]. A similar mechanism is sible to change the pulsewidth of the emitted laser echoes or
found in [107] by studying whether
the length of a triangular edge in
the resulting TIN is within a freeze
distance, where the value depends
Power

on the ANPS. All these approaches Plating


require the determination of a spe- Lidar ti tj
(b) i
cific threshold, and a parameter-free
Power

approach is often desired. Recently,


k
cloth simulation has been intro- Unwanted
duced to remove data pits without ti tj Points
(c) j
the need of undergoing the pit de- i
Power

tection and filling processes [114]. k


All the previously mentioned data j Stiffener
tk
pits appear in monochromatic lidar- (a) (d)
derived CHMs. Indeed, similar data
pits are also noted in the forest covers FIGURE 20. (a) Laser pulses are backscattered from the plating of a T-bar at i and stiffener at
of multispectral airborne lidar inten- j. (b) If the distance between the plating (i ) and stiffener ( j ) is larger than a laser pulsewidth,
sity data. However, there is a lack of two discrete returns can be identified. (c) If the distance between the plating (i ) and stiffener
studies thus far addressing these in- ( j ) is small, two laser returns from i and j mix, and (d) waveform decomposition, such as
tensity pits, providing opportunities Gaussian decomposition, may treat this as a single return with the echo peak located between
for further investigations. i and j, i.e., k, which is the location of unwanted points.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 37


increase the sampling rate of the waveform digitizer, these As detailed in Figure 21, a combination of shearing
artifacts can be removed only in the postprocessing stage. and stretching effects can be found in different on-road
moving vehicles. One should note that if the vehicle
MOTION ARTIFACTS is moving in the opposite direction to the aircraft [i.e.,
Modern terrestrial lidar systems, suc h as t he L eica i v = 180° in (11)], it results in a compressing effect [see
R TC360, are equipped with fast double scans to re- Figure 21(b) and compare the stationary vehicles parked
move abnormal point clouds caused by rain, water waves, on the roadside]. On the other hand, if the vehicle is
walking pedestrians, and so on. However, the function heading in the same direction as the aircraft, the moving
is considered suitable only for stationary measurements. vehicle suffers from a stretching effect [see Figure 21(c)].
Although an airborne lidar survey captures the instan- Either scenario may also cause a certain degree of shear-
taneous conditions of topography, motion artifacts are ing since the heading vectors of aircraft and moving ve-
notable when moving objects, especially on-road vehicles, hicles are unlikely to be exactly the same as or opposite
appear on the scene. As a result, the shapes of moving ve- to one another.
hicles are likely to be distorted. Despite being a type of visible artifact, the sheared or
Motion artifacts in airborne lidar datasets appear stretched stretched vehicles in lidar point clouds can nevertheless
and/or sheared depending on whether an along-track mo- be shown to reveal their motion information based on
tion or an across-track motion occurs. To model the geom- their distortions. Yao et al. [116] proposed a workflow
etry of a distorted vehicle, the following equation can be to first detect and extract the data points of vehicles, pa-
adopted [116]: rameterize the shapes of vehicles, classify their motions,
and, finally, estimate their speeds. Such an approach can
lv # va
v a - v v cos ^i v h
ls = (11) aid in traffic analysis to a large spatial extent, which can,
in turn, contribute to the macroscopic fundamental dia-
where l s and l v refer to the sheared length and original gram that requires information of both traffic speed and
length of the vehicle, respectively; v a and v v refer to the density [117].
speed of the aircraft and the vehicles, respectively; and i v
refers to an angle of intersection between the heading of the HARDWARE MALFUNCTIONS
aircraft and vehicle. The sheared angle ^i sh of the distorted System manufacturers usually conduct pref light mission
vehicle can be modeled using the following formula: inspection and calibration. Also, system malfunctions
should be reported through flight reporting and logs, as
^ h
i s = arctan d + 90° n . (12)
v v sin i v mentioned in existing specifications and guidelines [14],
v a - v v cos ^i v h [118]. Still, it is inevitable to experience hardware failure
before or during flight missions. As
mentioned in the “Intensity Banding”
section, if the lidar system is jarred
during transportation or installation,
Scanning the misalignment between the axes
Direction of the laser emitter and receiver re-
sults in intensity banding. A mechan-
ical solution to such a problem can be
found in [119], where the incorpora-
Vehicle Flight tion of a field lens in the telescope’s
Direction Direction
focal plane is proved to improve the
overlap factor between the receiver’s
FOV and laser beam. Stripe artifacts
in DEMs can be ascribed to systemat-
ic and random errors in the POS and
lidar systems [89].
Figure 22 presents an airborne
lidar dataset suffering from a me-
chanical problem. The laser scan-
(a) (b) (c) ning process literally starts when
the transmitter emits a laser pulse
FIGURE 21. (a) A vehicle moving in an across-track direction, resulting in a sheared effect. to the folding and scan mirrors and
(b) A vehicle moving in the opposite direction to the aircraft, leading to a compressed effect. passes through the sensor window;
(c) A vehicle moving in the same direction as the aircraft, causing a stretching effect. Aside the equipment needs to flip the sen-
from the motion artifacts, the DEM suffers from the effect of divots on the ground. sor window so as to avoid too much

38 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


backscattering or retroref lection. However, due to an the geometric accuracy can be measured by assessing the
unknown reason, it turns out that at a certain range of root-mean-square errors in both the horizontal and ver-
scan angles (213° and 1 23), the backscattered light tical directions. This can be achieved through comparing
finds a way to travel back to the detector. The light is too sets of selected target points from the ground survey of
strong and overwhelming, leading the recorded intensity the 3D point clouds or the resulting DSM/DEM against
to drop to one. The resulting intensity image thus shows its known coordinates. Nevertheless, there is a lack of
a specific direction of scans having a significant energy objective quality metrics that are capable of inferring the
loss in addition to the banding issue, as mentioned in the radiometric quality of lidar data. In view of the improved
“Intensity Banding” section. Postprocessing is unable to spectral information col-
restore the intensity in this scenario, and thus, the hard- lected by airborne lidar sys-
ware needs to be sent back to the manufacturer for sub- tems, there should be a way
DESPITE THE AVAILABILITY
sequent repair. to quantify the quality of air-
borne lidar point clouds. OF IMAGE, AUDIO, AND
THE WAY FORWARD A few initial attempts have VIDEO TRAINING DATASETS,
been reported lately in the THERE APPEARS TO BE A
ALGORITHMIC DEVELOPMENT computer science commu- LACK OF OPEN WELL-
AND TRAINING DATABASE nity for red–green–blue point LABELED DATASETS FOR
Regarding algorithmic development, both geometric and clouds collected by low-cost AIRBORNE LIDAR POINT
radiometric preprocessing (including calibration, correc- sensors. Yang et al. [122] pro-
CLOUDS.
tion, or normalization), ground filtering, and object ex- posed a process of resampling
traction (such as buildings) have indeed reached a mature the key points, local graph
stage. Nevertheless, some of the issues and artifacts cov- construction, aggregation of
ered in this article still require further research efforts to color gradients, and similarity derivation. The objective
improve robustness and accuracy. The latest developments of the workflow aims to assess the graph similarity of key
on this front include consideration of the development of a points extracted at high spatial frequency from the point
stratified strategy for the intensity correction of multispec- cloud and compare them against a reference point cloud
tral airborne lidar data, in that different land cover features without noise. Projecting a 3D point cloud onto a 2D image
should undergo different levels of range, angle, and atmo- for image quality assessment seems to be a viable approach
spheric attenuation correction [84], and using persistent [123], especially for airborne lidar point clouds. Through
homology to create a threshold-free approach for void de- comparing the image entropy difference (e.g., Boltzmann
tection [64] and classification coupled with a correspond- entropy) and the level of information extracted (e.g., clas-
ing point cloud inpainting strategy [120]. sification accuracy), both before and after radiometric
Despite the availability of image, audio, and video train- correction (refer to the “Intensity Banding” and “Intensity
ing datasets, there appears to be a lack of open well-labeled Striping at Swath Edges” sections), the radiometric quality
datasets for airborne lidar point clouds. Indeed, the lidar of airborne lidar data can be quantified.
community should contribute to working on open bench-
mark datasets that can serve different applications and
purposes. GANs can be trained and adopted to generate ar-
tificial data points to cover up voids for specific types of ob-
jects. In view of the improved mapping scale of regional and
global land cover maps, a large number of labeled airborne
lidar point clouds are also needed for training deep neural
networks for large-area land cover classification, semantic
segmentation, and point cloud completion. Also, a subjec-
tive point cloud quality assessment database should be cre-
ated so that it can act as a reference for assessing point cloud
quality. This can be achieved by assessing the geometric and
radiometric discrepancy between the reference points in the
database and the collected samples, such as the Point-Cen-
tered Quarter Method [121] and GraphSIM [122]. Further
discussion can be found in the following section.

INFERRING THE POINT CLOUD QUALITY


Airborne lidar flight surveys usually go along with ground FIGURE 22. The mechanical failure of a sensor window, causing
surveys or adopt existing ground controls to assess the incorrect intensity records at certain scan angles in addition to
geometric accuracy of the collected point clouds. Thus, intensity banding.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 39


ADVANCEMENTS IN OPTICS AND PHOTONICS functions handling the aforementioned artifacts. Cur-
Aside from algorithmic development, new findings and rent software tools may include functions for data in-
developments in optics and photonics can certainly im- teroperability, data compression, terrain generation,
prove the signal-to-noise ratio of lidar systems. The re- ground filtering, object extraction, segmentation, and
cent invention of the avalanche photodiode drives the classification. Nevertheless, preprocessing functions,
next-generation lidar systems. Jones et al. [124] adopted including void detection, classification and reconstruc-
the use of molecular beam epitaxy to grow a digital al- tion, geometric correction, intensity correction, data
loy, which has long-wavelength sensitivity and low excess noise and pit removal, and so on, are needed to improve
noise. Although further research efforts are needed to im- the data quality prior to information extraction. Also,
prove the sensitivity to the single-photon level and extend near-future software development certainly rides on the
the operating wavelength to the infrared region [125], wave of multicore CPU and GPU processing, on which
the reported study has made a significant breakthrough stand-alone and cloud-based platforms are expected to
to maintain eye safety while keeping high laser power, be based. In five to 10 years, data processing for large-
which is considered a bottleneck in existing avalanche area fine-resolution multimodal lidar point clouds may
photodiode technologies. On the other hand, quantum arrive when the development of quantum computing
parametric mode sorting (QPMS) appears to be a promis- reaches a mature stage [131].
ing new method for single-photon noise reduction [126].
The QPMS imprints certain quantum properties on the CONCLUSIONS
emitted laser beams and filters the backscattered laser Information extraction from airborne lidar point clouds
pulses in such a way that only those photons with match- has gone viral in the remote sensing community lately.
ing quantum properties are adopted and recorded. This Nevertheless, most of the studies, riding on the wave of
can aid in cleaning up unwanted noisy returns from am- artificial intelligence and machine learning, do not pay
bient light sources, e.g., sunlight, where existing solutions much attention to lidar data artifacts. Ensuring data
mainly rely on offline noise filtering algorithms to im- quality, i.e., completeness, accuracy, and reliability, is
prove the data quality of photon-counting lidar systems always crucial prior to data analysis in any application
[7]. An integrated antifogging and antireflective coating domain. Reducing these lidar data defects and artifacts
on lidar optical surfaces can facilitate water permeability can, in turn, maximize the potential of point cloud classi-
while maintaining low reflectivity [127]. This can boost fication, semantic labeling, the inference of topographic
the performance of lidar systems, avoid unwanted re- properties, and so on. Therefore, this article provided a
flections, and ultimately improve lidar data quality. The comprehensive review of three common types of airborne
white laser invention allows a full span of visible spec- lidar data artifacts, i.e., voids, stripes, and noise, by going
trum, which further improves lidar’s radiometric quality through their causes, potential impacts, current solutions,
if the laser output stability and eye safety issue can be and future directions. Most of the artifacts can be attrib-
guaranteed for long-range measurements [128]. Tech- uted to the geometric and radiometric discrepancies of
niques based on the principle of entanglement, such as lidar data with respect to instantaneous atmospheric and
two-photon interference lidar [129] or two-photon dual- ground conditions and hardware deficiency. One should
comb lidar [130], can facilitate improved depth resolution bear in mind that these artifacts, though they are usu-
to capture finer details of objects through the mechanism ally treated as undesired defects, sometimes shed light on
of wave interference. A significant enhancement of the certain semantic information. A rectangular data void in
detector’s or waveform digitizer’s sampling rate (e.g., up a residential area may likely be a swimming pool, for in-
to a picosecond or femtosecond) can also help improve stance. The intention of preparing this article was to alert
the range resolution, which can help resolve the issue ad- the community about the existence and implications of
dressed in the “Unwanted Points Present Between Two these data artifacts. On one hand, this article hopefully
Nearby Objects” section. All these advancements open serves as a guide for beginner lidar users to understand
up avenues for the future development of lidar hardware the causes of data artifacts and their possible solutions.
with improved precision and quality in both radiometric On the other hand, it hopefully urges the research com-
and geometric perspectives. munity and industry to incorporate functions handling
these artifacts in their lidar data processing bundles. The
PREPROCESSING FUNCTIONS IN SOFTWARE current unavailability of these functions severely handi-
The availability of airborne lidar data, particularly those caps lidar users in resolving data defects. Aside from
collected from unmanned aerial vehicle platforms, algorithmic development, advancements in optics and
dramatically increases at a rate that is faster than soft- photonics also drive the development of next-generation
ware development for data exploitation. Unlike aerial lidar systems with high stability and an improved signal-
photogrammetry and satellite remote sensing, there to-noise ratio. Quantum computing will also likely boost
is a lack of off-the-shelf software tools for airborne li- the computational power and efficiency of point cloud
dar point cloud processing, especially embedded with processing in the near future.

40 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


ACKNOWLEDGMENT [5] A. Habib, K. I. Bang, A. P. Kersting, and D.-C. Lee, “Error bud-
This work was supported by the General Research Fund get of LiDAR systems and quality control of the derived data,”
(Project 15221022, Quantifying Airborne Lidar Data Ar- Photogrammetric Eng. Remote Sens., vol. 75, no. 9, pp. 1093–1108,
tifacts) by the Research Grants Council of the Hong Kong Sep. 2009, doi: 10.14358/PERS.75.9.1093.
Special Administrative Region (HKSAR). The author [6] A. Habib, “Accuracy, quality assurance, and quality control of Li-
would like to express sincere appreciation to Dr. Paul E. DAR data: Principles and processing,” in Topographic Laser Ranging
LaRocque, Dr. Ana P. Kersting, and Alex Yeryomin, of and Scanning. Boca Raton, FL, USA: CRC Press, 2017, pp. 269–294.
Teledyne Optech, for discussing lidar system problems [7] A. Swatantran, H. Tang, T. Barrett, P. DeCola, and R. Dubayah,
and providing sample datasets. Acknowledgment also “Rapid, high-resolution forest structure and terrain mapping
goes to Prof. Ayman Habib, of Purdue University; Prof. over large areas using single photon LiDAR,” Scientific Rep.,
Ahmed Shaker, of Toronto Metropolitan University; Dr. vol. 6, no. 1, pp. 1–12, Jun. 2016, doi: 10.1038/srep28277.
Karin van Ewijk and Prof. Paul Treitz, of Queen’s Univer- [8] D. F. Laefer, S. Abuwarda, A.-V. Vo, L. Truong-Hong, and H.
sity; McElhanney Consulting Services, British Columbia, Gharibi, “2015 aerial laser and photogrammetry survey of
Canada; and the Department of Civil Engineering and Dublin city collection record,” NYU Spatial Data Repository,
Development, Government of the ­HKSAR, for providing New York, NY, USA, 2017. [Online]. Available: https://geo.nyu.
sample airborne lidar datasets. In addition, the author edu/catalog/nyu_2451_38684
thanks Dr. Ernest Ho for his meticulous proofreading [9] M. Isenburg. “Efficient tools for LiDAR processing.” LAStools.
and the anonymous referees for suggesting improve- Accessed: Jun. 1, 2022. [Online]. Available: https://rapidlasso.de/
ments. This work specifically pays tribute to the late Dr. [10] N. Pfeifer, G. Mandlburger, J. Otepka, and W. Karel, “OPALS –
Martin ­Isenburg for his dedication and contribution to A framework for airborne laser scanning data analysis,” Com-
data processing tools and open data/formats/methods for put., Environ. Urban Syst., vol. 45, pp. 125–136, May 2014, doi:
lidar point clouds. 10.1016/j.compenvurbsys.2013.11.002.
[11] J.-R. Roussel et al., “LidR: An R package for analysis of Airborne
AUTHOR INFORMATION Laser Scanning (ALS) data,” Remote Sens. Environ., vol. 251,
Wai Yeung Yan (waiyeung.yan@polyu.edu.hk) received his Dec. 2020, Art. no. 112061, doi: 10.1016/j.rse.2020.112061.
Ph.D. degree in civil engineering from Toronto Metropoli- [12] C. Mallet and F. Bretar, “Full-waveform topographic LiDAR:
tan University (formerly Ryerson University) in 2012. He State-of-the-art,” ISPRS J. Photogrammetry Remote Sens., vol. 64,
is currently an assistant professor with the Department of no. 1, pp. 1–16, Jan. 2009, doi: 10.1016/j.isprsjprs.2008.
Land Surveying and Geo-Informatics, Hong Kong Poly- 09.007.
technic University, Kowloon, Hong Kong, and an adjunct [13] H. K. Heidemann, “LiDAR base specification version 2.1,” in
professor with the Department of Civil Engineering, To- U.S. Geological Survey Techniques and Methods. Reston, VA, USA:
ronto Metropolitan University, Toronto, ON M5B 2K3, U.S. Geological Survey, 2019, ch. B4.
Canada. His research interests include improving lidar data [14] “Federal airborne LiDAR data acquisition guideline version
quality, point cloud processing, laser scanning, and remote 2.0,” Natural Resources Canada and Public Safety Canada,
sensing. He is a Senior Member of IEEE. ­Ottawa ON, Canada, General Inf. Product 117e, 2018. [Online].
Available: https://www.afn.ca/wp-content/uploads/2021/03/2.
REFERENCES -Federal-Airborne-LiDAR-Data-­Acquisition-Guideline-Version
[1] “LiDAR market by technology (2D, 3D, 4D), component (Laser -2.0.pdf
Scanners, Navigation and Positioning Systems), installation type [15] L. Zhou and G. Vosselman, “Mapping curbstones in airborne and
(Airborne and Ground Based), range, service, end-use applica- mobile laser scanning data,” Int. J. Appl. Earth Observ. Geoinf.,
tion, and region (2021-2026).” Markets and Markets. Accessed: vol. 18, pp. 293–304, Aug. 2012, doi: 10.1016/j.jag.2012.01.024.
Jun. 1, 2022. [Online]. Available: https://www.marketsandmarkets. [16] P. T.-Y. Shih and C.-M. Huang, “The building shadow problem
com/Market-Reports/lidar-market-1261.html of airborne LiDAR,” Photogrammetric Rec., vol. 24, no. 128, pp.
[2] “LiDAR market size, share & trends analysis report by type, 372–385, Nov. 2009, doi: 10.1111/j.1477-9730.2009.00550.x.
application, end-user, and segment forecasts, 2020–2027.” [17] S. A. N. Gilani, M. Awrangjeb, and G. Lu, “Segmentation of
Valuates Reports. Accessed: Jun. 1, 2022. [Online]. Avail- airborne point cloud data for automatic building roof extrac-
able: https://reports.valuates.com/market-reports/ALLI-Manu tion,” GIScience Remote Sens., vol. 55, no. 1, pp. 63–89, 2018,
-4A11/lidar doi: 10.1080/15481603.2017.1361509.
[3] “LiDAR market size worth $4.71 billion by 2030.” Grand [18] M. Feng, T. Zhang, S. Li, G. Jin, and Y. Xia, “An improved mini-
View Research. Accessed: Jun. 1, 2022. [Online]. Available: mum bounding rectangle algorithm for regularized building
https://www.grandviewresearch.com/press-release/global boundary extraction from aerial LiDAR point clouds with par-
-lidar-market tial occlusions,” Int. J. Remote Sens., vol. 41, no. 1, pp. 300–319,
[4] W. Y. Yan, A. Shaker, and N. El-Ashmawy, “Urban land cover 2020, doi: 10.1080/01431161.2019.1641245.
classification using airborne LiDAR data: A review,” Remote [19] L. Liu and S. Lim, “A framework of road extraction from airborne
Sens. Environ., vol. 158, pp. 295–310, Mar. 2015, doi: 10.1016/ LiDAR data and aerial imagery,” J. Spatial Sci., vol. 61, no. 2, pp.
j.rse.2014.11.001. 263–281, Apr. 2016, doi: 10.1080/14498596.2016.1147392.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 41


[20] L. Cheng, Y. Wu, Y. Wang, L. Zhong, Y. Chen, and M. Li, “Three- [32] T. Hilker et al., “Comparison of terrestrial and airborne Li-
dimensional reconstruction of large multilayer interchange DAR in describing stand structure of a thinned lodgepole pine
bridge using airborne LiDAR data,” IEEE J. Sel. Topics Appl. Earth forest,” J. Forestry, vol. 110, no. 2, pp. 97–104, Mar. 2012, doi:
Observ. Remote Sens., vol. 8, no. 2, pp. 691–708, Feb. 2015, doi: 10.5849/jof.11-003.
10.1109/JSTARS.2014.2363463. [33] B. B. Worstell, S. Poppenga, G. A. Evans, and S. Prince, “LiDAR
[21] A. Kobler, N. Pfeifer, P. Ogrinc, L. Todorovski, K. Oštir, and point density analysis: Implications for identifying water bod-
S. Džeroski, “Repetitive interpolation: A robust algorithm for ies,” U.S. Geological Survey, Reston, VA, USA, USGS Num-
DTM generation from aerial laser scanner data in forested ter- bered Series 2014-5191, 2014.
rain,” Remote Sens. Environ., vol. 108, no. 1, pp. 9–23, May 2007, [34] J. B. Coleman, X. Yao, T. R. Jordan, and M. Madden, “Holes in
doi: 10.1016/j.rse.2006.10.013. the ocean: Filling voids in bathymetric LiDAR data,” Comput.
[22] X. Liu, “Airborne LiDAR for DEM generation: Some critical is- Geosci., vol. 37, no. 4, pp. 474–484, Apr. 2011, doi: 10.1016/j.
sues,” Prog. Physical Geography, vol. 32, no. 1, pp. 31–49, Feb. cageo.2010.11.008.
2008, doi: 10.1177/0309133308089496. [35] B. Höfle, M. Vetter, N. Pfeifer, G. Mandlburger, and J. Stötter,
[23] A. L. Montealegre, M. T. Lamelas, and J. De La Riva, “A compari- “Water surface mapping from airborne laser scanning using sig-
son of open-source LiDAR filtering algorithms in a m ­ editerranean nal intensity and elevation data,” Earth Surf. Processes Landforms,
forest environment,” IEEE J. Sel. Topics Appl. Earth Observ. Remote vol. 34, no. 12, pp. 1635–1649, Sep. 2009, doi: 10.1002/esp.1853.
Sens., vol. 8, no. 8, pp. 4072–4085, Aug. 2015, doi: 10.1109/ [36] W. Y. Yan, A. Shaker, and P. E. LaRocque, “Scan line intensity-
JSTARS.2015.2436974. elevation ratio (SLIER): An airborne LiDAR ratio index for auto-
[24] R. Hill and R. K. Broughton, “Mapping the understorey of de- matic water surface mapping,” Remote Sens., vol. 11, no. 7, Apr.
ciduous woodland from leaf-on and leaf-off airborne LiDAR 2019, Art. no. 814, doi: 10.3390/rs11070814.
data: A case study in lowland Britain,” ISPRS J. P­ hotogrammetry [37] S. Tamari, J. Mory, and V. Guerrero-Meza, “Testing a near-in-
Remote Sens., vol. 64, no. 2, pp. 223–233, Mar. 2009, doi: frared LiDAR mounted with a large incidence angle to moni-
10.1016/j.isprsjprs.2008.12.004. tor the water level of turbid reservoirs,” ISPRS J. Photogramme-
[25] D. Kükenbrink, F. D. Schneider, R. Leiterer, M. E. Schaepman, try Remote Sens., vol. 66, no. 6, pp. S85–S91, Dec. 2011, doi:
and F. Morsdorf, “Quantification of hidden canopy volume of 10.1016/j.isprsjprs.2011.01.009.
airborne laser scanning data using a voxel traversal algorithm,” [38] A. V. Jelalian, Laser Radar Systems. Norwood, MA, USA: Artech
Remote Sens. Environ., vol. 194, pp. 424–436, Jun. 2017, doi: House, 1992.
10.1016/j.rse.2016.10.023. [39] J. L. Bufton, F. E. Hoge, and R. N. Swift, “Airborne measurements
[26] I. Korpela, A. Hovi, and F. Morsdorf, “Understory trees in of laser backscatter from the ocean surface,” Appl. Opt., vol. 22,
airborne LiDAR data — Selective mapping due to transmis- no. 17, pp. 2603–2618, Sep. 1983, doi: 10.1364/AO.22.002603.
sion losses and echo-triggering mechanisms,” Remote Sens. [40] L. S. Rose et al., “Challenges and lessons from a wetland Li-
Environ., vol. 119, pp. 92–104, Apr. 2012, doi: 10.1016/j. DAR project: A case study of the Okefenokee Swamp, Geor-
rse.2011.12.011. gia, USA,” Geocarto Int., vol. 28, no. 3, pp. 210–226, 2013, doi:
[27] K. Richter and H.-G. Maas, “Radiometric enhancement of 10.1080/10106049.2012.681707.
full-waveform airborne laser scanner data for volumetric rep- [41] L. P. Rampi, J. F. Knight, and C. F. Lenhart, “Comparison of
resentation in environmental applications,” ISPRS J. Photogram- flow direction algorithms in the application of the CTI for
metry Remote Sens., vol. 183, pp. 510–524, 2022, doi: 10.1016/j. mapping wetlands in Minnesota,” Wetlands, vol. 34, no. 3, pp.
isprsjprs.2021.10.021. 513–525, Feb. 2014, doi: 10.1007/s13157-014-0517-2.
[28] L. Chasmer, C. Hopkinson, B. Smith, and P. Treitz, “Examin- [42] J. A. Czuba, S. R. David, D. A. Edmonds, and A. S. Ward, “Dy-
ing the influence of changing laser pulse repetition frequencies namics of surface-water connectivity in a low-gradient mean-
on conifer forest canopy returns,” Photogrammetric Eng. Remote dering river floodplain,” Water Resour. Res., vol. 55, no. 3, pp.
Sens., vol. 72, no. 12, pp. 1359–1367, Dec. 2006, doi: 10.14358/ 1849–1870, Mar. 2019, doi: 10.1029/2018WR023527.
PERS.72.12.1359. [43] A. Shaker, W. Y. Yan, and P. E. LaRocque, “Automatic land-­
[29] B. M. Wing, M. W. Ritchie, K. Boston, W. B. Cohen, A. Gitel- water classification using multispectral airborne LiDAR data
man, and M. J. Olsen, “Prediction of understory vegetation for near-shore and river environments,” ISPRS J. Photogramme-
cover with airborne LiDAR in an interior ponderosa pine for- try Remote Sens., vol. 152, pp. 94–108, Apr. 2019, doi: 10.1016/j.
est,” Remote Sens. Environ., vol. 124, pp. 730–741, Sep. 2012, isprsjprs.2019.04.005.
doi: 10.1016/j.rse.2012.06.024. [44] W. Y. Yan, “Scan line void filling of airborne LiDAR point
[30] M. E. Hodgson et al., “An evaluation of LiDAR-derived eleva- clouds for hydroflattening DEM,” IEEE J. Sel. Topics Appl. Earth
tion and terrain slope in leaf-off conditions,” Photogrammetric Observ. Remote Sens., vol. 14, pp. 6426–6437, Jun. 2021, doi:
Eng. Remote Sens., vol. 71, no. 7, pp. 817–823, Jul. 2005, doi: 10.1109/JSTARS.2021.3089288.
10.14358/PERS.71.7.817. [45] D. Gatziolis and H.-E. Andersen, “A guide to LiDAR data ac-
[31] T. Hilker et al., “Comparing canopy metrics derived from ter- quisition and processing for the forests of the Pacific North-
restrial and airborne laser scanning in a Douglas-fir dominated west,” U.S. Dept. Agriculture, Forest Service, Pacific Northwest
forest stand,” Trees, vol. 24, no. 5, pp. 819–832, Oct. 2010, doi: Research Station, Portland, OR, USA, General Tech. Rep. PNW-
10.1007/s00468-010-0452-7. GTR-768, 2008.

42 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


[46] D. Burton, D. B. Dunlap, L. J. Wood, and P. P. Flaig, “LiDAR in- generative adversarial inpainting,” Sensors, vol. 20, no. 18, Sep.
tensity as a remote sensor of rock properties,” J. Sedimentary Res., 2020, Art. no. 5029, doi: 10.3390/s20185029.
vol. 81, no. 5, pp. 339–347, May 2011, doi: 10.2110/jsr.2011.31. [61] Z. Huang, Y. Yu, J. Xu, F. Ni, and X. Le, “PF-Net: Point frac-
[47] S. Kaasalainen et al., “Effect of target moisture on laser scanner tal network for 3D point cloud completion,” in Proc. IEEE/CVF
intensity,” IEEE Trans. Geosci. Remote Sens., vol. 48, no. 4, pp. Conf. Comput. Vision Pattern Recognit., 2020, pp. 7662–7670, doi:
2128–2136, Apr. 2010, doi: 10.1109/TGRS.2009.2036841. 10.1109/CVPR42600.2020.00768.
[48] S. Kaasalainen et al., “Radiometric calibration of LiDAR inten- [62] X. Wang, M. H. Ang Jr., and G. H. Lee, “Cascaded refine-
sity with commercially available reference targets,” IEEE Trans. ment network for point cloud completion,” in Proc. IEEE/CVF
Geosci. Remote Sens., vol. 47, no. 2, pp. 588–598, Mar. 2009, doi: Conf. Comput. Vision Pattern Recognit., 2020, pp. 790–799, doi:
10.1109/TGRS.2008.2003351. 10.1109/CVPR42600.2020.00087.
[49] S. O. Elberink and G. Vosselman, “Quality analysis on 3D [63] X. Wen, T. Li, Z. Han, and Y.-S. Liu, “Point cloud completion
building models reconstructed from airborne laser scanning by skip-attention network with hierarchical folding,” in Proc.
data,” ISPRS J. Photogrammetry Remote Sens., vol. 66, no. 2, pp. IEEE/CVF Conf. Comput. Vision Pattern Recognit., 2020, pp.
157–165, Mar. 2011, doi: 10.1016/j.isprsjprs.2010.09.009. 1939–1948, doi: 10.1109/CVPR42600.2020.00201.
[50] K. Garroway, C. Hopkinson, and R. Jamieson, “Surface mois- [64] A. Zomorodian and G. Carlsson, “Computing persistent ho-
ture and vegetation influences on LiDAR intensity data in an mology,” Discrete Comput. Geometry, vol. 33, no. 2, pp. 249–274,
agricultural watershed,” Can. J. Remote Sens., vol. 37, no. 3, pp. Feb. 2005, doi: 10.1007/s00454-004-1146-y.
275–284, Jun. 2011, doi: 10.5589/m11-036. [65] H. O. Ørka, T. Gobakken, E. Næsset, L. Ene, and V. Lien, “Si-
[51] A. Vain, X. Yu, S. Kaasalainen, and J. Hyyppa, “Correcting air- multaneously acquired airborne laser scanning and multi-
borne laser scanning intensity data for automatic gain control spectral imagery for individual tree species identification,”
effect,” IEEE Geosci. Remote Sens. Lett., vol. 7, no. 3, pp. 511–514, Can. J. Remote Sens., vol. 38, no. 2, pp. 125–138, Feb. 2012, doi:
Jul. 2010, doi: 10.1109/LGRS.2010.2040578. 10.5589/m12-021.
[52] I. Korpela, H. O. Ørka, J. Hyyppä, V. Heikkinen, and T. Tokola, [66] W. Y. Yan and A. Shaker, “Airborne LiDAR intensity banding:
“Range and AGC normalization in airborne discrete-return Cause and solution,” ISPRS J. Photogrammetry Remote Sens., vol. 142,
LiDAR intensity data for forest canopies,” ISPRS J. Photogram- pp. 301–310, Aug. 2018, doi: 10.1016/j.isprsjprs.2018.06.013.
metry Remote Sens., vol. 65, no. 4, pp. 369–379, Jul. 2010, doi: [67] C. L. Leigh, D. B. Kidner, and M. C. Thomas, “The use of ­LiDAR
10.1016/j.isprsjprs.2010.04.003. in digital surface modelling: Issues and errors,” Trans. GIS,
[53] F. Pirotti, A. Guarnieri, and A. Vettore, “State of the art of vol. 13, no. 4, pp. 345–361, Aug. 2009, doi: 10.1111/j.1467-9671.
ground and aerial laser scanning technologies for high- 2009.01168.x.
resolution topography of the earth surface,” Eur. J. Remote [68] R. A. Nobrega, J. A. Quintanilha, and C. G. O’Hara, “A noise-
Sens., vol. 46, no. 1, pp. 66–78, Feb. 2013, doi: 10.5721/ removal approach for LiDAR intensity images using anisotropic
EuJRS20134605. diffusion filtering to preserve object shape characteristics,” in
[54] P. Liepa, “Filling holes in meshes,” in Proc. Eurograph./ACM SIG- Proc. ASPRS Annu. Conf., Tampa, FL, USA, May 2007.
GRAPH Symp. Geometry Process., 2003, pp. 200–205. [69] W. Y. Yan, K. van Ewijk, P. Treitz, and A. Shaker, “Effects of radio-
[55] W. Zhao, S. Gao, and H. Lin, “A robust hole-filling algorithm metric correction on cover type and spatial resolution for mod-
for triangular mesh,” Vis. Comput., vol. 23, no. 12, pp. 987–997, eling plot level forest attributes using multispectral airborne
Dec. 2007, doi: 10.1007/s00371-007-0167-y. LiDAR data,” ISPRS J. Photogrammetry Remote Sens., vol. 169, pp.
[56] A. Ruiz, W. Kornus, J. Talaya, and J. Colomer, “Terrain model- 152–165, Nov. 2020, doi: 10.1016/j.isprsjprs.2020.09.001.
ing in an extremely steep mountain: A combination of airborne [70] M. Okhrimenko, C. Coburn, and C. Hopkinson, “Multi-
and terrestrial LiDAR,” Int. Soc. Photogrammetry Remote Sens., spectral LiDAR: Radiometric calibration, canopy spectral
vol. 36, no. 3, pp. 281–284, Jan. 2004. reflectance, and vegetation vertical SVI profiles,” Remote
[57] Z. Zhang, G. Vosselman, M. Gerke, C. Persello, D. Tuia, Sens., vol. 11, no. 13, Jun. 2019, Art. no. 1556, doi: 10.3390/
and M. Y. Yang, “Detecting building changes between air- rs11131556.
borne laser scanning and photogrammetric data,” Remote [71] M. Okhrimenko and C. Hopkinson, “Investigating the consis-
Sens., vol. 11, no. 20, Oct. 2019, Art. no. 2417, doi: 10.3390/ tency of uncalibrated multispectral LiDAR vegetation indices at
rs11202417. different altitudes,” Remote Sens., vol. 11, no. 13, Jun. 2019, Art.
[58] M. S. O’Banion, M. J. Olsen, J. P. Hollenbeck, and W. C. Wright, no. 1531, doi: 10.3390/rs11131531.
“Data gap classification for terrestrial laser scanning-derived [72] T. R. Goodbody, P. Tompalski, N. C. Coops, C. Hopkinson, P.
digital elevation models,” ISPRS Int. J. Geo-Inf., vol. 9, no. 12, Treitz, and K. van Ewijk, “Forest inventory and diversity attribute
Dec. 2020, Art. no. 749, doi: 10.3390/ijgi9120749. modelling using structural and intensity metrics from multi-
[59] J.-F. Parrot and C. R. Núñez, “LiDAR DTM: Artifacts, and cor- spectral airborne laser scanning data,” Remote Sens., vol. 12,
rection for river altitudes,” Investigaciones Geográficas, Boletín del no. 13, Jul. 2020, Art. no. 2109, doi: 10.3390/rs12132109.
Instituto de Geografía, vol. 2016, no. 90, pp. 28–39, Aug. 2016, [73] P. Knott, T. Proctor, A. Hayes, J. Ralph, P. Kok, and J. Dunning-
doi: 10.14350/rig.47372. ham, “Local versus global strategies in multiparameter estima-
[60] J. Chen, J. S. K. Yi, M. Kahoush, E. S. Cho, and Y. K. Cho, “Point tion,” Physical Rev. A, vol. 94, no. 6, Dec. 2016, Art. no. 062312,
cloud scene completion of obstructed building facades with doi: 10.1103/PhysRevA.94.062312.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 43


[74] W. Y. Yan, A. Shaker, A. Habib, and A. P. Kersting, “Improv- [89] C. K. Toth and Z. Koppanyi, “Strip adjustment,” in Topographic
ing classification accuracy of airborne LiDAR intensity data Laser Ranging and Scanning. Boca Raton, FL, USA: CRC Press,
by geometric calibration and radiometric correction,” ISPRS J. 2018, pp. 259–289.
Photogrammetry Remote Sens., vol. 67, pp. 35–44, Jan. 2012, doi: [90] A. Gruen and D. Akca, “Least squares 3D surface and curve
10.1016/j.isprsjprs.2011.10.005. matching,” ISPRS J. Photogrammetry Remote Sens., vol. 59, no. 3,
[75] B. Höfle and N. Pfeifer, “Correction of laser scanning intensity pp. 151–174, May 2005, doi: 10.1016/j.isprsjprs.2005.02.006.
data: Data and model-driven approaches,” ISPRS J. Photogram- [91] H.-G. Maas, “On the use of pulse reflectance data for laserscan-
metry Remote Sens., vol. 62, no. 6, pp. 415–433, Dec. 2007, doi: ner strip adjustment,” Int. Arch. Photogrammetry Remote Sens.
10.1016/j.isprsjprs.2007.05.008. Spatial Inf. Sci., vol. 34, no. 3/W4, pp. 53–56, Jan. 2001.
[76] A. G. Kashani, M. J. Olsen, C. E. Parrish, and N. Wilson, “A re- [92] M. Albani and B. Klinkenberg, “A spatial filter for the removal
view of LiDAR radiometric processing: From ad hoc intensity of striping artifacts in digital elevation models,” Photogrammet-
correction to rigorous radiometric calibration,” Sensors, vol. 15, ric Eng. Remote Sens., vol. 69, no. 7, pp. 755–765, Jul. 2003, doi:
no. 11, pp. 28,099–28,128, Nov. 2015, doi: 10.3390/s151128099. 10.14358/PERS.69.7.755.
[77] W. Y. Yan and A. Shaker, “Radiometric correction and normaliza- [93] J. Gallant, “Adaptive smoothing for noisy DEMs,” in Proc. Geo-
tion of airborne LiDAR intensity data for improving land-cover morphometry, Redlands, CA, USA, Sep. 2011, pp. 37–40.
classification,” IEEE Trans. Geosci. Remote Sens., vol. 52, no. 12, [94] B. Münch, P. Trtik, F. Marone, and M. Stampanoni, “Stripe and
pp. 7658–7673, Jun. 2014, doi: 10.1109/TGRS.2014.2316195. ring artifact removal with combined wavelet--Fourier filter-
[78] C. Hopkinson, “The influence of flying altitude, beam diver- ing,” Opt. Exp., vol. 17, no. 10, pp. 8567–8591, Jun. 2009, doi:
gence, and pulse repetition frequency on laser pulse return in- 10.1364/OE.17.008567.
tensity and canopy frequency distribution,” Can. J. Remote Sens., [95] T. H. Tarekegn and T. Sayama, “Correction of SRTM DEM ar-
vol. 33, no. 4, pp. 312–324, Aug. 2007, doi: 10.5589/m07-029. tefacts by Fourier transform for flood inundation modeling,”
[79] S. Kaasalainen et al., “Radiometric calibration of ALS inten- J. Jpn. Soc. Civil Eng., B1 (Hydraul. Eng.), vol. 69, no. 4, pp. I_193–I
sity,” Int. Arch. Photogrammetry, Remote Sens. Spatial Inf. Sci., _198, Jan. 2013, doi: 10.2208/jscejhe.69.I_193.
vol. 36, no. Part3/W52, pp. 201–205, Jan. 2007. [96] M. Favalli, A. Fornaciai, and M. T. Pareschi, “LiDAR strip
[80] I. S. Korpela, “Mapping of understory lichens with airborne dis- adjustment: Application to volcanic areas,” Geomorphology,
crete-return LiDAR data,” Remote Sens. Environ., vol. 112, no. 10, vol. 111, nos. 3–4, pp. 123–135, Oct. 2009, doi: 10.1016/j.geo-
pp. 3891–3897, Oct. 2008, doi: 10.1016/j.rse.2008.06.007. morph.2009.04.010.
[81] D. Gatziolis, “Dynamic range-based intensity normalization [97] A. Zlinszky, E. Boergens, P. Glira, and N. Pfeifer, “Airborne la-
for airborne, discrete return LiDAR data of forest canopies,” ser scanning for calibration and validation of inshore satellite
Photogrammetric Eng. Remote Sens., vol. 77, no. 3, pp. 251–259, altimetry: A proof of concept,” Remote Sens. Environ., vol. 197,
Mar. 2011, doi: 10.14358/PERS.77.3.251. no. 4, pp. 35–42, Aug. 2017, doi: 10.1016/j.rse.2017.04.027.
[82] H. You, T. Wang, A. Skidmore, and Y. Xing, “Quantifying the [98] J.-H. Song, S.-H. Han, K. Yu, and Y.-I. Kim, “Assessing the pos-
effects of normalisation of airborne LiDAR intensity on conif- sibility of land-cover classification using LiDAR intensity data,”
erous forest leaf area index estimations,” Remote Sens., vol. 9, Int. Arch. Photogrammetry Remote Sens. Spatial Inf. Sci., vol. 34,
no. 2, Feb. 2017, Art. no. 163, doi: 10.3390/rs9020163. no. 3/B, pp. 259–262, May 2012.
[83] M. Kukkonen, M. Maltamo, L. Korhonen, and P. Packalen, “Mul- [99] X. Lai, X. Zheng, and Y. Wan, “A kind of filtering algorithms for
tispectral airborne LiDAR data in the prediction of boreal tree spe- LiDAR intensity image based on flatness terrain,” in Proc. IEEE
cies composition,” IEEE Trans. Geosci. Remote Sens., vol. 57, no. 6, Int. Symp. Spatio-temporal Model., Spatial Reasoning, Anal., Data
pp. 3462–3471, Jan. 2019, doi: 10.1109/TGRS.2018.2885057. Mining Data Fusion, 2005.
[84] W. Y. Yan, “Intensity correction of multispectral airborne [100] A. H. Incekara, D. Z. Seker, and B. Bayram, “Qualifying the
laser scanning data,” in Proc. IEEE Int. Geosci. Remote Sens. LiDAR-derived intensity image as an infrared band in NDWI-
Symp., Brussels, Belgium: IEEE, Jul. 2021, pp. 7712–7715, doi: based shoreline extraction,” IEEE J. Sel. Topics Appl. Earth Ob-
10.1109/IGARSS47720.2021.9553543. serv. Remote Sens., vol. 11, no. 12, pp. 5053–5062, Dec. 2018,
[85] N. El-Sheimy, “Georeferencing component of LiDAR systems,” doi: 10.1109/JSTARS.2018.2875792.
in Topographic Laser Ranging and Scanning. Boca Raton, FL, USA: [101] X. Li, B. Yang, X. Xie, D. Li, and L. Xu, “Influence of waveform
CRC Press, 2017, pp. 195–214. characteristics on LiDAR ranging accuracy and precision,”
[86] A. Habib, K. I. Bang, A. P. Kersting, and J. Chow, “Alternative Sensors, vol. 18, no. 4, Apr. 2018, Art. no. 1156, doi: 10.3390/
methodologies for LiDAR system calibration,” Remote Sens., s18041156.
vol. 2, no. 3, pp. 874–907, Mar. 2010, doi: 10.3390/rs2030874. [102] X.-F. Han, J. S. Jin, M.-J. Wang, W. Jiang, L. Gao, and L. Xiao, “A
[87] A. F. Habib, A. P. Kersting, A. Shaker, and W.-Y. Yan, “Geometric review of algorithms for filtering the 3D point cloud,” Signal
calibration and radiometric correction of LiDAR data and their Process., Image Commun., vol. 57, pp. 103–112, May 2017, doi:
impact on the quality of derived products,” Sensors, vol. 11, no. 10.1016/j.image.2017.05.009.
9, pp. 9069–9097, Sep. 2011, doi: 10.3390/s110909069. [103] Y. Duan, C. Yang, H. Chen, W. Yan, and H. Li, “Low-complexity
[88] G. Vosselman, “Analysis of planimetric accuracy of airborne point cloud denoising for LiDAR by PCA-based dimension re-
laser scanning surveys,” Int. Arch. Photogrammetry, Remote Sens. duction,” Opt. Commun., vol. 482, Mar. 2021, Art. no. 126567,
Spatial Inf. Sci., vol. 37, no. 3a, pp. 99–104, 2008. doi: 10.1016/j.optcom.2020.126567.

44 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


[104] E. Schubert, J. Sander, M. Ester, H. P. Kriegel, and X. Xu, “DB- fic,” Transp. Res. B, Methodological, vol. 42, no. 9, pp. 771–781,
SCAN revisited, revisited: Why and how you should (Still) use Nov. 2008, doi: 10.1016/j.trb.2008.06.008.
DBSCAN,” ACM Trans. Database Syst., vol. 42, no. 3, pp. 1–21, [118] “Specifications for airborne LiDAR for the province of British
Jul. 2017, doi: 10.1145/3068335. Columbia,” Ministry of Forests, Lands, Natural Resources
[105] L. Zhou, G. Sun, Y. Li, W. Li, and Z. Su, “Point cloud denois- Operations and Rural Development, Victoria, BC, Canada,
ing review: From classical to deep learning-based approaches,” Version 5.0, 2020, p. 54.
Graphical Models, vol. 121, May 2022, Art. no. 101140, doi: [119] R. Chen, Y. Jiang, and H. Wang, “Calculation method of the over-
10.1016/j.gmod.2022.101140. lap factor and its enhancement for airborne LiDAR,” Opt. Com-
[106] R. A. Haugerud and D. Harding, “Some algorithms for virtual mun., vol. 331, pp. 181–188, Nov. 2014, doi: 10.1016/j.optcom.
deforestation (VDF) of LiDAR topographic survey data,” Int. 2014.05.063.
Arch. Photogrammetry Remote Sens., vol. 34, no. 3/W4, pp. 211– [120] S. Crema et al., “Can inpainting improve digital terrain analy-
218, Jan. 2001. sis? Comparing techniques for void filling, surface reconstruc-
[107] A. Khosravipour, A. K. Skidmore, and M. Isenburg, “Generating tion and geomorphometric analyses,” Earth Surf. Processes
spike-free digital surface models using LiDAR raw point clouds: Landforms, vol. 45, no. 3, pp. 736–755, Mar. 2020, doi: 10.1002/
A new approach for forestry applications,” Int. J. Appl. Earth esp.4739.
Observ. Geoinf., vol. 52, pp. 104–114, Oct. 2016, doi: 10.1016/ [121] G. Meynet, Y. Nehmé, J. Digne, and G. Lavoué, “PCQM: A full-
j.jag.2016.06.005. reference quality metric for colored 3D point clouds,” in Proc.
[108] K. Lim, P. Treitz, M. Wulder, B. St-Onge, and M. Flood, “Li- 12th IEEE Int. Conf. Qual. Multimedia Exp. (QoMEX), 2020, pp.
DAR remote sensing of forest structure,” Progr. Physical Ge- 1–6, doi: 10.1109/QoMEX48832.2020.9123147.
ography, vol. 27, no. 1, pp. 88–106, Mar. 2003, doi: 10.1191/ [122] Q. Yang, Z. Ma, Y. Xu, Z. Li, and J. Sun, “Inferring point cloud
0309133303pp360ra. quality via graph similarity,” IEEE Trans. Pattern Anal. Mach.
[109] M. A. Wulder et al., “LiDAR sampling for large-area forest char- Intell., vol. 44, no. 6, pp. 3015–3029, Jun. 2022, doi: 10.1109/
acterization: A review,” Remote Sens. Environ., vol. 121, pp. 196– TPAMI.2020.3047083.
209, Jun. 2012, doi: 10.1016/j.rse.2012.02.001. [123] Q. Yang, H. Chen, Z. Ma, Y. Xu, R. Tang, and J. Sun, “Predicting
[110] J. R. Ben-Arie, G. J. Hay, R. P. Powers, G. Castilla, and B. St- the perceptual quality of point cloud: A 3D-to-2D projection-
Onge, “Development of a pit filling algorithm for LiDAR can- based exploration,” IEEE Trans. Multimedia, vol. 23, pp. 3877–
opy height models,” Comput. Geosci., vol. 35, no. 9, pp. 1940– 3891, 2021, doi: 10.1109/TMM.2020.3033117.
1949, Sep. 2009, doi: 10.1016/j.cageo.2009.02.003. [124] A. H. Jones, S. D. March, S. R. Bank, and J. C. Campbell, “Low-
[111] A. Khosravipour, A. K. Skidmore, M. Isenburg, T. Wang, and noise high-temperature AlInAsSb/GaSb avalanche photodi-
Y. A. Hussin, “Generating pit-free canopy height models from odes for 2-μm applications,” Nature Photon., vol. 14, no. 9, pp.
airborne LiDAR,” Photogrammetric Eng. Remote Sens., vol. 80, no. 559–563, Sep. 2020, doi: 10.1038/s41566-020-0637-6.
9, pp. 863–872, Sep. 2014, doi: 10.14358/PERS.80.9.863. [125] S. Kodati et al., “Low noise AlInAsSb avalanche photodiodes
[112] C. Chen, Y. Wang, Y. Li, T. Yue, and X. Wang, “Robust and pa- on InP substrates for 1.55 μm infrared applications,” in Proc. In-
rameter-free algorithm for constructing pit-free canopy height frared Technol. Appl., Bellingham, WA, USA: Int. Soc. Opt. Eng.,
models,” Int. J. Geo-Inf., vol. 6, no. 7, Jul. 2017, Art. no. 219, doi: 2021, vol. 11741, Art. no. 117411X, doi: 10.1117/12.2587884.
10.3390/ijgi6070219. [126] A. Shahverdi, Y. M. Sua, L. Tumeh, and Y.-P. Huang, “Quantum
[113] H. Liu and P. Dong, “A new method for generating canopy parametric mode sorting: Beating the time-frequency filter-
height models from discrete-return LiDAR point clouds,” ing,” Scientific Rep., vol. 7, no. 1, pp. 1–12, 2017, doi: 10.1038/
Remote Sens. Lett., vol. 5, no. 6, pp. 575–582, Jun. 2014, doi: s41598-017-06564-7.
10.1080/2150704X.2014.938180. [127] A. Gärtner et al., “Combined antifogging and antireflective dou-
[114] W. Zhang et al., “Cloth simulation-based construction of pit- ble nanostructured coatings for LiDAR applications,” Appl. Opt.,
free canopy height models from airborne LiDAR data,” For- vol. 62, no. 7, pp. B112–B116, Jan. 2023, doi: 10.1364/AO.476974.
est Ecosystems, vol. 7, no. 1, pp. 1–13, Jan. 2020, doi: 10.1186/ [128] F. Fan, S. Turkdogan, Z. Liu, D. Shelhammer, and C.-Z. Ning, “A
s40663-019-0212-0. monolithic white laser,” Nature Nanotechnol., vol. 10, no. 9, pp.
[115] T. Allouis, J.-S. Bailly, Y. Pastol, and C. Le Roux, “Comparison of 796–803, Jul. 2015, doi: 10.1038/nnano.2015.149.
LiDAR waveform processing methods for very shallow water ba- [129] R. Murray and A. Lyons, “Two-photon interference LiDAR im-
thymetry using Raman, near-infrared and green signals,” Earth aging,” Opt. Exp., vol. 30, no. 15, pp. 27,164–27,170, Jul. 2022,
Surf. Processes Landforms, J. Brit. Geomorphological Res. Group, vol. doi: 10.1364/OE.461248.
35, no. 6, pp. 640–650, May 2010, doi: 10.1002/esp.1959. [130] H. Wright, J. Sun, D. McKendrick, N. Weston, and D. T. Reid,
[116] W. Yao, M. Zhang, S. Hinz, and U. Stilla, “Airborne traffic mon- “Two-photon dual-comb LiDAR,” Opt. Exp., vol. 29, no. 23, pp.
itoring in large areas using LiDAR data – Theory and experi- 37,037–37,047, Nov. 2021, doi: 10.1364/OE.434351.
ments,” Int. J. Remote Sens., vol. 33, no. 12, pp. 3930–3945, Jun. [131] S. S. Gill et al., “Quantum computing: A taxonomy, systematic
2012, doi: 10.1080/01431161.2011.637528. review and future directions,” Softw., Pract. Exp., vol. 52, no. 1,
[117] C. F. Daganzo and N. Geroliminis, “An analytical approxima- pp. 66–114, Jan. 2022, doi: 10.1002/spe.3039.
tion for the macroscopic fundamental diagram of urban traf- GRS

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 45


Interferometric
Phase Linking
Algorithm, application, and perspective

DINH HO TONG MINH AND


STEFANO TEBALDINI

M itigating decorrelation effects


on interferometric synthetic
aperture radar (InSAR) time series
data is challenging. The phase linking
(PL) algorithm has been the key to han-
dling signal decorrelations in the past 15
years. Numerous studies have been carried
out to enhance its precision and computational
efficiency. Different PL algorithms have been
proposed, each with unique phase optimization
approaches, such as the quasi-Newton method,
equal-weighted and coherence-weighted factors, com-
ponent extraction and selection SAR (CAESAR), and ei-
gendecomposition-based algorithm (EMI). The differenc-
es among the PL algorithms can be attributed to the weight
criteria adopted in each algorithm, which can be coherence-
based, sparsity-based, or other forms of regularization. The PL
algorithm has multiple applications, including SAR tomography
(TomoSAR), enhancing distributed scatterers (DSs) to combine
with persistent scatterers (PS) in PS and DS (PSDS) techniques, and
compressed PSDS InSAR (ComSAR), where it facilitates the retrieval
of the optimal phase from all possible measurements. This article aims
to review PL techniques developed in the past 15 years. The review also
underscores the importance of the PL technique in various SAR applica-
tions (TomoSAR, PSDS, and ComSAR). Finally, the deep learning (DL) ap-
proach is discussed as a valuable tool to improve the accuracy and efficiency
of the PL process.

INTRODUCTION
InSAR has become an increasingly popular tool for high-precision deformation
monitoring, due to its ability to detect small changes over time and provide a
unique view of Earth’s surface [1]. One critical challenge in InSAR is extracting

Digital Object Identifier 10.1109/MGRS.2023.3300974


Date of current version: 19 September 2023 IMAGE LICENSED BY INGRAM PUBLISHING

46 2473-2397/23©2023IEEE IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


meaningful information from the interferometric phase, Accurately estimating linked phases is critical in co-
which is affected by atmospheric conditions, topography, herent analysis for mitigating decorrelation effects on
and decorrelations [2], [3]. Several techniques have been SAR data. Hence, numerous studies have been carried out
proposed to address these signal decorrelations to improve to enhance the precision and computational efficiency of
the interferometric phase’s quality [1]. PL estimation since the work of Guarnieri and Tebaldini
The first category of techniques is based on PS InSAR, [10]. Ferretti et al. [11] proposed a quasi-Newton method
which utilizes individual scatterers dominating the signal for unconstrained nonlinear optimization, which is the
from a resolution cell to track deformation through time Broyden–Fletcher–Goldfarb–Shanno algorithm [12] for
[4]. PS interferometry (PSI) techniques provide high-qual- the MLE solution. This optimization technique is effi-
ity deformation information at point target locations. cient in minimizing the nonlinear cost function of the
However, in natural scenes, PSI technology, widely used PL problem. Cao et al. [13] introduced equal-weighted
for deformation estimation in urban areas, may not be and coherence-weighted factors for phase optimization,
sufficient to obtain accurate results, due to a low den- which provide flexibility for incorporating coherence
sity of PSs [5]. To address the challenges associated information. The CAESAR algorithm [14] was proposed
with the limited information in InSAR data, an for phase optimization under multiple scattering mecha-
alternative approach is based on DSs, which of- nisms. It is based on the coherence matrix’s eigenvalue
fers the potential to leverage information more decomposition (EVD) and can extract different scatter-
effectively. DS targets are commonly found ing components. The EMI algorithm proposed by Ansari
in natural environments, such as mead- et al. [15] is also EVD based and MLE based. It is efficient
ows, fields, and bare soil, where multiple in computation and estimation due to its exploitation
scatterers with similar brightness con- of coherence information. Ho Tong Minh and Ngo [16]
tribute to the information in a reso- proposed a compression technique for PL estimation. The
lution cell. However, to account for method involves dividing massive data into ministacks
signal decorrelation, one should and then compressing them to enhance the noise-free
select interferogram subsets for short-lived interferometric components. Finally, Zwie-
a temporal analysis using short back [17] proposed regulation methods to improve the
spatial and temporal base- estimation of the coherence matrix. The improvement
lines, known as small base- in the compression [16] and regulation [17] approaches
line subsets (SBASs) [6], [7], is most significant for low long-term coherences, due to
[8]. This approach has more reliable coherence matrix estimates. Overall, the
demonstrated promis- differences among the PL algorithms can be attributed
ing results in various ap- to the weight criteria adopted in each algorithm, which
plications, such as ground can be coherence-based, sparsity-based, or other forms of
deformation monitoring and regularization.
surface elevation mapping. The SBAS Addressing signal decorrelations to improve the
approach offers a valuable alternative to interferometric phase’s quality is crucial for many ap-
traditional InSAR methods by leveraging the in- plications. The PL algorithm has been instrumental in
formation from DSs [9]. However, deformation mea- handling this issue in the past 15 years. In the past five
surements on distributed targets are often of lower years, there have been 826 review articles published,
quality and require spatial multilooked filtering. while the total number of papers related to PL since
Another approach is the PL method intro- its introduction is 1,286, according to results from
duced by Guarnieri and Tebaldini [10]. PL Google Scholar as of June 2023. However, there is a
is defined as a statistical method used in need for a comprehensive review of the different PL al-
interferometry to combine multiple inter- gorithms proposed so far. Notably, no comprehensive
ferometric phases into a single equivalent sin- review has been conducted thus far on the relation-
gle-reference (ESR) phase. Suppose N single-look ship between PL and its applications in diverse SAR
complex (SLC) images are available. The PL algorithm methodologies, including TomoSAR, PSs and DSs, and
is the maximum likelihood estimation (MLE) of the N – 1 the ComSAR algorithm. This article aims to bridge
ESR phase from all possible combinations of N (N - 1) /2. this knowledge gap by providing an overview of the
Before applying the SBAS approach, it is necessary to un- PL techniques developed over the past 15 years and
wrap the interferograms. However, the PL method ex- emphasizing the significance of the PL technique in
ploits all wrapped interferometric phases to optimize the various SAR applications. Furthermore, the potential
phase quality. SqueeSAR technology [11], which uses a of employing DL as a valuable tool to enhance the
phase triangulation algorithm, is one example of the PL precision and efficiency of the PL process is explored
method. These optimized ESR phases of DS targets can be and discussed. Table 1 provides nomenclature used
used in conventional PSI processing. throughout the article.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 47


PHASE LINKING MODELS AND topography, deformation, atmosphere, and noise [2].
PROPOSED ALGORITHMS For single pixels, the closure phase W 1, 2, 3 is always zero
by definition and is called the phase consistency condi-
CLOSURE PHASE PROBLEM tion or phase triangularity condition [18], [ [19]. However,
Suppose that N SLC SAR images are available for a specific the phase triangularity condition is not necessarily val-
area of interest. The images are coregistered on a reference id for multilooked interferometric pixels, as the closure
grid, and phase contributions due to terrain topography phase can be nonzero [11], [19], [20]. Figure 1 shows
and orbit have been compensated. Let yn be the nth coregis- an example of a full-scene closure phase from three
tered SLC images in the form of multilooked interferometric phases. It has been de-
rived from Advanced Land Observing Satellite (ALOS) 2/
y n = S n exp (j{ n)(1)
Phased-Array L-Band Synthetic Aperture Radar (PALSAR)
where Sn is the amplitude and { n is the phase of yn. With 2 data acquired over Vietnam. Ho Chi Minh City occupies
three SAR images, it is possible to generate three single-look the west center of the image.
interferograms written as Notably, it has been demonstrated that nonsymmetric
volumetric targets have a nonzero phase closure [20]. This
y 12 = y 1 y )2, y 13 = y 1 y )3, y 23 = y 2 y )3 (2)
implies that the zero-closure model assumes no phase clo-
where * denotes the complex conjugate. The closure phase is sure and is inadequate for volume scattering scenarios, such
the circular combination of the three interferometric phases: as forests and glaciers. In such cases, multilooked interfero-
grams assume a mathematical model representing the volu-
W 1, 2, 3 = W {{ 12 + { 23 + { 31}(3)
metric target as an “equivalent point target” with a “phase
where W is the wrapping (modulo-2π) operator, { nm = center” position. Consistent with this interpretation, the
{ n - { m is the single-look interferometric phase of ynm, mathematical model of multilooked interferograms as-
and { nm includes contributions related to the residual sumed in PL algorithms is given as

E {y n y )m} = c nm v n v m exp ^ j (z n - z m) h (4)

TABLE 1. NOMENCLATURE USED IN THIS ARTICLE. where E denotes spatial averaging, cnm is the coherence of
the nmth interferometric pair, v n = E {S 2n}, z n is the mul-
ABBREVIATION EXPANSION
tilook interferometric phase for the nth acquisition, and
CAESAR Component extraction and selection synthetic
aperture radar z nm = z n - z m is the multilook interferometric phase of ynm.
CNN Convolutional neural network Equation (4) requires reevaluating the phase estimation
ComSAR Compressed persistent scatterers and distributed in a stack of SAR images. The selection and weighting of the
scatterers interferometric SAR algorithm interferograms can affect the accuracy of the reconstructed
DL Deep learning
CRLB Cramér‒Rao lower bound
11°0′0″N 11°10′0″N

DS Distributed scatterer
N
EMI Eigendecomposition-based maximum likelihood
estimator of interferometric phase
ESA European Space Agency
ESR Equivalent single reference +30

EVD Eigenvalue decomposition


GAN Generative adversarial network
Degree
10°30′0″N 10°40′0″N 10°50′0″N

MLE Maximum likelihood estimator


InSAR Interferometric SAR
PCA Principal component analysis –30

PL Phase linking
PS Persistent scatterer
PSDS PSs and DSs
PSI PS interferometry
PU Phase unwrapping
RNN Recurrent NN
0 5 10 20
SAR Synthetic aperture radar km

SBAS Small baseline subsets 106°30′0″E 106°40′0″E 106°50′0″E 107°0′0″E 107°10′0″E


SHP Statistically homogeneous pixel
SLC Single-look complex FIGURE 1. The closure phase corresponds to three ALOS-2/PALSAR-2
TomoSAR SAR tomography acquisitions over Vietnam, acquired on 12 January, 23 March, and
TSPA Two-stage programming approach 10 August 2018. Ho Chi Minh City is recognizable in the west center
of the image.

48 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


phase history. PL is a technique to address this statistical where
misclosure.
R jj 1 V
Se 0 g 0 W
COHERENCE MATRIX S0 e jj 2 g 0 W
H=S W
Under the hypothesis of distributed scattering, the prob-
SS h h j h W
jj NW
ability density function of the data may be regarded as 0 0 g e
a zero-mean multivariate circular normal distribution. RT 1 c 12 g
X
c 1NWV
S
Therefore, an ensemble of the second-order moments (i.e., S c 21 1 g c 2NW
G=S .
a covariance matrix or a coherence matrix) represents suf- SS h h j h WW
ficient statistics to infer information from the data. Com- c N1 c N2 f 1 W
T X
pared to a covariance matrix, the coherence matrix can
avoid amplitude disturbance among SAR images. For sim-
plicity, we assume, without any loss of generality, that the Here, m = 6j 1, j 2, ..., j N@T is the optimal phase that
images are normalized such that v n = 1 for every n [see (4)] needs to be estimated from the filtered N (N - 1)/2 phas-
in the remainder of this article. Under this assumption, the es. It should be noted that the N (N - 1) /2 interferometric
covariance matrix is identical to a coherence matrix. The phase values are not redundant, due to the spatial filter-
expression of a sampled coherence matrix can be defined as ing process. Furthermore, the estimation of the absolute
phase series is ambiguous. The phase of an arbitrary im-
t = E [yy H] . 1 / yy H (5)
C age in the time series is set to zero, and the remaining
L y !X
values are measured comparably to this arbitrary datum.
where H is the conjugate transpose and represents a ho- Without any loss of generality, the first value can be set
mogeneous patch containing L adjacent pixels with simi- to zero, such as m = 60, j 2, ..., j N@T . Therefore, only N – 1
lar scattering properties. The absolute and phase values of phase values need to be estimated. PL is a technique to
the elements of C t are the estimated coherence value ct and estimate the optimal N – 1 interferometric phases from
interferometric phase z n, respectively. Therefore, C t can be the possible N (N - 1)/2 phases. In other words, PL can be
expressed as understood as combining multiple interferometric phases
R V into an ESR phase. PL is commonly formulated as an op-
S 1 ct 12 e jz 12 g ct 1N e jz1NW
S ct 21 e jz 21 j 2NW timization problem.
t =S 1 g ct 2N e z
t obtained from (5), the MLE of R follows from
C W Given C
S h h j h W
Sct N1 e jz N1 ct N2 e jz N2 f the maximization of the Wishart probability density func-
1 W
Tt X tion given by
= | C |% U (6)

where U is an N # N matrix, with element e jz nm indicating R t | R)A-


t = argmax # ln 7p (C
R
the interferometric phases between the nth acquisition and = argmax # - tr (LR -1 C
t ) - L ln ^det (R) h- 
t represents an N # N matrix
mth acquisition. In this case, C R

t
with element c nm .
t ) + ln ^det (G) h- .
= argmin # tr (HG -1 H H C (9)
G, H

MAXIMUM LIKELIHOOD ESTIMATION To estimate H, the true coherence G is required. Since G


PHASE LINKING is unknown in practice, it is common to use the standard
Based on the central limit theorem, we assume that the nor- plug-in estimate by setting it as
malized SAR data vector y follows a complex multivariate
normal distribution with zero mean and dispersion matrix t (10)
G= C
R. The dispersion matrix R represents the scattering prop-
erties of a DS. Recalling R as a model for the underlying co- resulting in unregularized PL and guaranteeing nonnega-
variance of a complex circular Gaussian process, it is known tiveness [17]. We then can write (9) as
that the probability density function of C t follows a com-
t = argmin # tr (H | C
R t ) - . (11)
t | -1 H H C
plex Wishart distribution with L degrees of freedom [13]: H

t ) L - N exp " - tr 6LR -1 C


(L) NL det (C t @,
t | R) =
p (C (7) Equation (11) was originally proposed by Guarnieri and
det (R) L P ^gamma (L - j + 1) h
N (N - 1) N
r 2 Tebaldini [10]. To ensure a unified analysis, considering the
j=1
variations in PL approaches, (11) can be reformulated based
where tr(.) and det(.) indicate the trace and determinant on the employed weighting strategy. In detail, at row n and
operators. The R of a generic pixel can be expressed using column m, the MLE of the phase value is [13]
“true” coherence values and “true” phase values as

mt MLE = argmin ( / clnm cos (z nm - j n + j m) 2 (12)


N N

R = HGH H (8) m
/
n = 1 m= n+1

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 49


where clnm is the element of Hadamard product C t -1 & C
t, [13], [14], [15]. On the other hand, the second approach
defined as the weight factor in the optimization, and & involves estimating the coherence matrix, which can be
represents the Hadamard entry-wise product. In practice, a achieved through regularization or compression tech-
damping factor is used in the niques [16], [17].
algorithm to remove small Ferretti et al. [11] proposed a quasi-Newton method for
negative or null eigenvalues unconstrained nonlinear optimization, i.e., the Broyden–
ONE CRITICAL CHALLENGE t before ma-
of the matrix C Fletcher–Goldfarb–Shanno algorithm [12], for the MLE
IN INSAR IS EXTRACTING
trix inversion. The MLE may solution. Cao et al. [13] introduced equal-weighted and
MEANINGFUL INFORMATION be interpreted as a temporal coherence-weighted factors in the phase optimization.
FROM THE filter that compresses the in- They proposed a modified MLE algorithm incorporating
INTERFEROMETRIC PHASE, formation of N (N - 1)/2 in- the coherence matrix. The coherence-weighted algorithm
WHICH IS AFFECTED BY terferograms to a phase series can be more accurate than the equal-weighted one in tem-
ATMOSPHERIC CONDITIONS, of size N. The solution of (12) poral decorrelation. Fornaro et al. [14] proposed a specific
TOPOGRAPHY, AND is achieved by iteratively min- CAESAR algorithm for estimating the phase in the pres-
imizing [10]. After this proce- ence of multiple scattering mechanisms. The authors used
DECORRELATIONS.
dure, the DS phase values are an EVD-based approach to extract different scattering
filtered, resulting in more re- components from the coherence matrix. The algorithm
liable PU. With this assumed was shown to be accurate and computationally efficient
model, the scattering behavior of the DS neighborhood is for estimating the phase in the presence of multiple scat-
approximated by a PS-resembling mechanism. tering mechanisms. Ansari et al. [15] proposed an EVD-
based MLE (EMI) algorithm to estimate the phase. They
RECENT ADVANCES used an EVD approach to reduce the problem’s dimen-
Estimating the linked phase is the crucial step in coherent sionality and improve the algorithm’s computational effi-
analysis to account for decorrelating targets. Consequently, ciency. The initial solution can be found as the minimum
dedicated research focuses on improving the precision and eigenvector of the matrix C t -1 & C
t . Ho Tong Minh and
computational efficiency in the work of Guarnieri and Te- Ngo [16] proposed a data compression technique to im-
baldini [10]. These techniques can be categorized into two prove the precision of the phase estimation algorithm. The
main groups. The first approach is based on computation authors divided massive data into many ministacks and
methods involving optimization, iteration, or EVD [11], then compressed them. The improvement of the phase
estimation algorithm from the compression is due to the
noise-free short-lived interferometric components. Zwie-
TABLE 2. CHARACTERISTICS OF THE MAIN PL APPROACHES. back [17] proposed a method for improving the estima-
tion of the coherence matrix by regulating the coherence
METHOD/
REFERENCE NAME WEIGHT DESCRIPTION matrix estimates. For instance, with spectral regulariza-
COMPUTATION tion, Ct regularization = bI + (1 - b) C
t , with b varied from zero
Guarnieri MLE clnm The element of Hadamard to one. The regularization techniques impose constraints
and Tebaldini ­product |Ct | -1 & Ct , with the
[10] ­iterative solution
Ferretti et al. MLE clnm Similar to Guarnieri and ­Tebaldini
0.2
[11] [10], with the solution by the
Broyden–Fletcher– ­Goldfarb– 0.18
Root-Mean-Square Error (rad)

Shanno algorithm 0.16


Cao et al. Coherence ct nm The element of coherence matrix
0.14
[13] Ct (with equal-weighted factor
ct nm = 1) 0.12
Fornaro et al. EVD ct nm hlnm hlnm is the element of ma- 0.1 Coherence
[14] trix |h 1| | h 1| T , where |h 1| is
0.08 EVD
the m­ aximum eigenvector of EMI
­coherence matrix Ct 0.06 MLE
Ansari et al. EMI clnm Similar to Guarnieri and Tebaldini 0.04 Regularization
[15] [10], with the iterative solution Compression
that initializes as the minimum 0.02 CRLB
eigenvector of the matrix |Ct | -1 & Ct 0
0 20 40 60 80 100 120 140 160 180
COHERENCE MATRIX
SLC Image Index
Ho Tong MLE c compression
nm
The element of Hadamard
Minh and ­product |Ct compression | -1 & Ct compression
Ngo [16] FIGURE 2. PL performances using a Sentinel-1 temporal coherence
Zwieback MLE c regularization
nm
The element of Hadamard model. The coherence is modeled as two exponential decays and
[17] ­product |Ct regularization |-1 & Ct regularization a long-term coherent component [21], [16]. The performances are
ordered to facilitate the visualization.

50 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


on the coherence matrix estimates to improve their ac- valuable information on forest structure [26], [27], classifi-
curacy. In compression [16] and regularization [17] ap- cation [28], and biomass [23].
proaches, the improvements are most significant for low Conventional SAR imaging of forests is challenging due
long-term coherences, due to more reliable coherence ma- to the complex interaction between the radar signal and for-
trix Ct estimates. In summary, the difference among the est canopy. The forest canopy is composed of vegetation lay-
PL algorithms can be interpreted as the adopted weight ers that attenuate and scatter the radar signal in a complex
criteria, as in (12). Table 2 gives the characteristics of dif- way, making it difficult to retrieve information on the under-
ferent PL approaches. lying terrain and vegetation
Figure 2 illustrates PL performances. We employ a well- [23]. TomoSAR can overcome
documented coherence Sentinel-1 model to simulate the some of these limitations by
behavior of temporal coherence over time [21], [16]. This exploiting the 3D structure of PL IS DEFINED AS A
model generates a coherence matrix for a three-year time the forest. When the radar’s STATISTICAL METHOD USED
series of 180 temporally ordered measurements taken at wavelength is long enough to IN INTERFEROMETRY TO
six-day intervals. Each measurement includes an ensemble penetrate the forest canopy, it COMBINE MULTIPLE
of 300 statistically homogeneous samples. The simulation becomes possible to use mul- INTERFEROMETRIC PHASES
is repeated 1,000 times. The EMI solution corresponds tiple SAR acquisitions with INTO A SINGLE EQUIVALENT
to the minimum eigenvector of the matrix C t -1 & Ct . The slightly different look angles
SINGLE-REFERENCE PHASE.
spectral regularization is used with a b of 0.5. We set the over the same area to quantify
ministack as 10 for the compression method. The CRLB is the three dimensions of forest
a theoretical measure that employs simulated coherence reflectivity. This principle is
for the calculation, as described in [10]. The EVD and co- demonstrated in Figure 3(a) and (c), where acquisitions from
herence weight results are very similar. The compressed traditional SAR and TomoSAR are shown, respectively.
estimator performs better than other approaches, closely To further elaborate on the TomoSAR process, consider a
approximating the CRLB. The compressed estimator’s re- scenario where a sensor carefully flies along N parallel tracks
duced error is attributed to the lack of noise in short-lived and acquires multibaseline data of SAR images. Each pixel
components [16]. at the slant range r and azimuth location x in the nth image
It is essential to highlight that the significance of PL is denoted by y n (r, x). The azimuth axis x is defined by the
becomes more pronounced in ill-posed InSAR scenarios. direction of the aircraft platform, whereas the slant range r
However, with specific well-posed InSAR techniques, is the distance line of sight (LOS) linking the SAR’s sensor
such as the TSPA, PU methods are already integrated into to targets on the ground, as in Figure 3(c). It is assumed that
the processing workflow [22]. Thus, exploring the synergy each image within the multibaseline dataset has been coreg-
between PL and the TSPA could lead to exciting research istered and resampled on a common grid (i.e., the reference
opportunities and further advancements in handling chal- track) and that phase components due to terrain topography
lenging InSAR scenarios. and platform motion have been compensated. Thus, the
Since the introduction of PL, there have been nu- multibaseline SAR model can be written as [24], [23]
merous SAR applications. Among them, the PSDS tech-
nique has gained immense popularity, as it surpasses PSI y n ^r, x h = # S^p, r, xh exp a j 4mrr b n p kdp (13)
in terms of performance [11]. In tomographic focusing,
the PL algorithm plays a crucial role in the phase calibra- where p is the cross-range coordinate, defined by the direc-
tion, as it requires an optimal phase model to compensate tion orthogonal to the LOS and azimuth coordinate; bn is
for potential phase residuals that may affect 3D imaging the normal baseline relative to the nth image with respect
[23]. The ComSAR technique is a recent advancement that to the reference image; m is the carrier wavelength; and
utilizes PL algorithms to select the most coherent inter- S ^p, r, x h is the average scene complex reflectivity within
ferograms based on their linked phases [16]. These appli- the slant range, azimuth, and cross-range resolution cell, as
cations are discussed in the “Applications” section. described in Figure 3(d).
It is worth noting that the SAR scene and its geometric
APPLICATIONS configuration are linked. Specifically, the distribution of
the SAR scene’s reflectivity in the cross-range direction is
SYNTHETIC APERTURE RADAR TOMOGRAPHY directly related to the multibaseline SAR data. These two
TomoSAR is a relatively new technique that has emerged as components form a Fourier pair [as shown in (13)]. As a
a powerful tool for the 3D imaging of complex scenes [24]. result, it is possible to reconstruct the cross-range distribu-
TomoSAR builds upon the capability of SAR systems to ac- tion of the scene’s complex reflectivity by taking the Fourier
quire data from multiple angles, which enables the recon- transform, as follows [23]:
struction of the 3D structure of the imaged object [25]. In
/ y n ^r, xh exp a - j 4mrr b n p k.(14)
N
the case of forests, TomoSAR can be used to retrieve the ver- St (p, r, x) =
tical distribution of scatterers within the canopy, providing n=1

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 51


Tomographic analysis relies on
the outline theoretical model that
ξ assumes a disturbance-free propagat-
ing signal. Before focusing, a phase
z 30 m calibration procedure is required to
compensate for phase residuals po-
x 15 m
tentially affecting 3D focusing [20],
0m [31], [32]. In addition, analyzing the
O
backscattered power from different
y heights in a forest makes it possible
(a) (b) to gain insight into the scattering
mechanisms within the canopy [23].
To successfully implement this con-

e
ng
t0 t cept, a vertical axis reference must

Ra
1 tN–1 ξ
be established, which enables height

s
os
measurements with respect to ter-

Cr
30 m
Sl rain elevation. The ground surface is
an z h
tR t often called the “zero-meter layer.”
imu 15 m
an Az x However, the ground phase contri-
ge
r 0m bution must be separated from the
O vegetation phase to prevent it from
influencing the 3D focusing.
y
(c) (d) These two points can be made by
removing the ground phase contri-
bution in the tomographic data [23].
FIGURE 3. A comparison of traditional SAR and TomoSAR acquisitions. (a) and (b): Traits of tra- The ground phases are determined
ditional SAR. (c) and (d): Traits of TomoSAR. (a) A SAR acquisition. (b) A SAR resolution cell. (c) A not only by terrain height zg but also
TomoSAR acquisition. (d) A TomoSAR resolution cell. The figure was adapted from [29, Fig. 2]. by the phase disturbances h deriv-
ing from the platform motion. In a
Consequently, TomoSAR processing enables us to ob- formula, { ground = k z z g + h, where k z = 4rb n /m sin iR n is the
tain the cross-range distribution of the SAR scene’s reflec- height-to-phase factor and i is the local incidence angle.
tivity at every range and azimuth location. By doing so, The multipolarimetric multibaseline covariance matrix
we can obtain a 3D image that provides comprehensive W can be approximated by retaining the first two terms
information on the reflectivity of a forest in three dimen- of the sum of the Kronecker products [33]. In a formula,
sions. This information can be used to derive valuable W . C G 7 R G + C V 7 R V , where R and C are referred to as
forest structure characteristics, such as height and bio- interferometric information and polarimetric information, re-
mass [23], [29], [30]. spectively, and G and V are associated with ground-only
and volume-only contributions, respectively. PL is a fun-
damental component in facilitating the retrieval of this
ground phase contribution from RG. Indeed, applying
z PL to forested areas allows for representing forest scatter-
N
30 m ing in terms of the “equivalent point target,” with well-
x
defined distances from the radar in different trajectories.
15 m This allows for simultaneous target and radar position
estimation, after which platform motion can be corrected
0m with subwavelength accuracy [20]. Figure 4 presents an
example of SAR and TomoSAR imaging in the Paracou
tropical forest site (French Guiana, South America).

0 0.5 1
PERSISTENT SCATTERERS AND DISTRIBUTED
y
(a) (b) SCATTERERS TECHNIQUE
The PSDS technique is an approach that leverages the phase
FIGURE 4. A comparison of traditional SAR and TomoSAR imaging. change over time of both PS and DS targets [1]. The tech-
(a) A traditional SAR image from the Paracou, French Guiana, forest nique involves two main steps: 1) PL and signal decorrela-
site. (b) TomoSAR layers, with each related to a certain height tion removal and 2) estimation of parameters of interest.
above the ground. The figure was adapted from [29, Fig. 4(b)]. PSDS refers to techniques that exploit the time series phase

52 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


change of both PS and DS targets. SqueeSAR technology that exceed a certain threshold [16]. Once a DS candidate
[11] is one example of the PSDS technique. is identified, its sample coherence matrix can be computed
The PL technique is applied first to all the N (N - 1) /2 using its SHP family. The coherence matrix fully character-
interferograms available from N images. It jointly ex- izes the target statistics and can be used to invert for N – 1
ploits these interferograms to squeeze the best estimates linked phases using PL.
of N - 1 interferometric phases. Once the estimate of the The quality of the estimated N – 1 phase values can be
N - 1linked phases has been produced, the second step assessed using the PL coherence, defined as [11]
is necessary to remove the signal decorrelations and es-
N N
timate the parameters of interest, such as the elevation
c PL = N (N - 1) Re / / exp ^i (z nm - j n + j m) h). (15)
2
error and constant velocity, similar to the PS interfer- n = 1 m= n+1
ometry processing algorithm [1]. The PSDS technique is
widely used in InSAR applications, as it provides a reli-
able way to detect and monitor changes in Earth’s sur-
face [34], [35], [36], [37]. It allows for detecting small and 1,500
slow surface movements over a large area, which is help-
ful in surface deformation applications [38]. An example

Normal Baseline (m)


1,000
of how interferogram networks can be exploited in the
InSAR time series techniques is provided in Figure 5. The 500
figure was generated using Constellation of Small Satel-
0
lites for Mediterranean Basin Observation data over the
Ha Noi, Vietnam, area. It illustrates how interferogram
–500
networks can be exploited to analyze PSDS targets over
time [39]. –1,000
It is important to note that PSI can be considered a 2011 2012 2013 2014 2015
particular case of the maximum likelihood interferom- Acquisition Date
(a)
etry approach [1]. In PSI, all PS targets are assumed to be
equally correlated in all images. This assumption obviates 1,500
the need for the joint processing of all available interfero-
Normal Baseline (m)

grams to estimate interferometric phase information. In- 1,000


stead, only removing the reference phase from all other
500
phases is necessary to obtain the linked phases required
for estimating parameters, such as surface deformation.
0
The PSDS algorithm is an extension of the PSI method
that considers both PS and DS targets. The joint process- –500
ing of all available interferograms in PSDS can help ac-
count for variations in the radar signal caused by different –1,000
2011 2012 2013 2014 2015
surface characteristics, including changes in moisture
Acquisition Date
content, surface roughness, and vegetation cover. As a (b)
result, the PSDS approach can offer greater accuracy and
1,500
detail than the PSI method in detecting and measuring
surface deformation.
Normal Baseline (m)

1,000
The DS target is known for its low average temporal co-
herence, primarily due to geometrical and temporal decor- 500
relation phenomena [3]. As a result, this target often has
a low signal-to-noise ratio, making it challenging to work 0
with. However, enhancing the DS target’s signal-to-noise
ratio and treating it as a PS target is possible by identify- –500
ing pixels within a neighborhood that exhibit similar be-
–1,000
havior. These similar pixels are called SHPs. They can be 2011 2012 2013 2014 2015
identified using a two-sample Kolmogorov–Smirnov [11] Acquisition Date
or Baumgartner–Weiss–Schindler [40] test on the ampli- (c)
tude-based time series of the current pixel and its neigh-
bors within a specified window. The pixels with a similar FIGURE 5. Interferogram networks. (a) The single master network
cumulative probability distribution are grouped as “broth- in PSI processing. (b) The subset network consists of interfero-
ers,” resulting in a family of SHPs [see Figure 6(a)]. A DS grams with short spatial and temporal baselines in SBASs.
candidate is identified if it has a sufficient number of SHPs (c) The fully connected network in the PSDS technique.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 53


If the c PL coherence is above a certain threshold, a DS small scatterers with a uniform scattering mechanism
point with the linked N – 1 phase value will replace the can be relaxed by considering DS targets dominated by
original points. Finally, the selected DS will jointly process two or more scattering mechanisms [36], [45]. Engel-
using the same PSI technique brecht et al. [45] showed that incorporating multiple
as the PS. Figure 6(b) is an scattering mechanisms in L-band ALOS PALSAR data
example of PL coherence, improved deformation measurement extraction in dy-
EXPLORING THE SYNERGY which is in full resolution, al- namic agricultural regions. Recently, it has been dem-
BETWEEN PL AND THE TSPA lowing for more comprehen- onstrated that adding polarimetric information can
COULD LEAD TO EXCITING sive PL performance. increase the number of coherent pixels by a factor of
RESEARCH OPPORTUNITIES Improving PL and select- eight compared to a single-polarization channel [46].
AND FURTHER ing SHPs are critical aspects Overall, improving SHP selection and PL, as well as
ADVANCEMENTS IN of PSDS techniques. Various considering more complex scattering mechanisms, can
modified approaches have enhance the performance of PSDS techniques in defor-
HANDLING CHALLENGING
been proposed to select mation monitoring and other applications.
INSAR SCENARIOS.
SHPs, including the Ander-
son–Darling test [34], time COMPRESSED PERSISTENT SCATTERERS AND
series likelihood ratios [41], DISTRIBUTED SCATTERERS INTERFEROMETRIC
t-test [42], fast SHP selection, Baumgartner–Weiss– SYNTHETIC APERTURE RADAR TECHNIQUE
Schindler test [40], mean amplitude difference [43], and The recent trend of spaceborne SAR missions focuses on
similar time series interferometric phase [44], among systematic Earth monitoring with high temporal resolu-
others. These approaches aim to increase the density tion [47], [48]. This has resulted in unprecedented SAR
of DSs, mitigating sample coherence bias. Addition- data volumes due to short revisit cycles, as brief as six to
ally, the conventional DS assumption of independent 12 days, with missions such as the ESA’s Sentinel-1 [49]
and NASA–Indian Space Research Organization Synthet-
ic Aperture Radar mission [50]. However, interferometric
300 processing of these large data stacks with currently avail-
50 able algorithms is infeasible. To address this demand, ef-
250 forts have focused on reducing product latency through
100
parallelized computation and cloud computing. One of
200
150 the first efficient stacking techniques that allows for ef-
200 150 ficient PL in sequences using isolated data batches of the
time series is the sequential estimator [51]. The ComSAR
250 100 algorithm introduced by Ho Tong Minh and Ngo [16] has
300 been proposed for processing PSDS analysis. Since most
50
deformation phenomena develop slowly, a processing
350
0 scheme can be devised using reduced-volume data sets.
200 400 600 800 1,000 1,200 1,400 1,600 The algorithm divides massive data into many ministacks
(a) and compresses them.
1 Conventional MLE PL suffers from a high computa-
50 0.9 tional time complexity driven by the number of SLCs
0.8 involved, primarily due to the iterative maximum like-
100 lihood optimization for the phase estimation and sec-
0.7
150 0.6 ondarily due to the regularization and inversion of the
0.5
complex coherence matrix, both of which are affected by
200
the number of interferograms. To address the high-data-
0.4
250 volume problem, data compression is a classic approach
0.3
in dealing with high data volumes. In the case of multi-
300 0.2
pass SAR, the objective is to compress a stack of coregis-
350 0.1
tered SAR data in the temporal direction so that the size
0 of the time series is reduced but the spatial size of each
200 400 600 800 1,000 1,200 1,400 1,600
(b) image is intact. To perform data compression, PCA is a
well-known technique [52], [53]. However, PCA fails to
FIGURE 6. Internal results from PSDS-based processing. (a) The incorporate the statistical properties of the complex cova-
number of SHPs was identified using the Baumgartner–Weiss– riance matrix data correctly, as it is a geometrical rather
Schindler test on a 9#35 window. (b) The PL coherence corre- than a probabilistic approach [53]. On the other hand,
sponds to the SHP map. The figure was adapted from [16, Fig. 4]. MLE PL is a purely probabilistic approach and is well

54 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


known for its precise phase estimation [10], [51]. There- can set M as 10 (see Figure 7). The processing will then es-
fore, the simple idea is to use the most coherent interfero- timate the 10 optimum phases by using the PL technique
grams from their linked phases because the linked phases in (12). These phases allow us to coherently focus the stack
are optimal from all possible interferometric phases. This subset and produce a single compressed image that can
approach reduces the data volume by using the most in- represent the 10 first images of the stack. The same pro-
formative interferograms for further processing, thereby cedure will be repeated on the following 10 images until
allowing for efficient interferometric processing. the end of the stack, producing nine compressed images.
Let us assume that the N SAR dataset (ordered tempo- We note that this process has to be performed on an SHP
rally) can be divided into small batches, or ministacks, with family basis since PL estimation can be valid only locally.
M images. The compressed version St of M SAR images for The ComSAR scheme in signal processing has benefits be-
the kth sequence can be determined by a coherent summa- yond just reducing the computational burden. It also prevents
tion, as follows [51], [16]: the need for updating and re-estimating the entire phase his-
tory at the face of every single acquisition. The processing
M
St (r, x) = / S m (r, x) g m (16) scheme reduces the data volume from the entire stack to the
m=1 compressed SLCs, making storage easier. These compressed
images can then be used as a reference point to link history
where g = K t < = 6p 1, p 2, ..., p M@, K
t /< K t = exp (jmt k) is the linked ministacks with recent acquisitions and reconstruct the full
phase from M SAR images and S(r, x) is the scene complex- phase time series [see Figure 7(d)]. In detail, PL will be per-
valued SLC at the slant range azimuth position. The vector formed on the compressed components (St ), producing a
g weights each image’s contribution in the ministack in the vector mt cal = 7jt cal (1), jt cal (2), ..., jt cal (K)A that contains the
coherent summation process. calibration phases for connecting the ministacks. The datum
For each ministack, its compression is formed using connection for the kth sequence is then carried out by [16]
(16), resulting in a strong data reduction. As an example
of the algorithm’s efficiency, from a stack of 90 images, we mt kunified = mt k + jt cal (k)(17)

20 0.8

40 0.6

0.4
60
0.2
80
0
20 40 60 80

(b)
2
4
6
8 (c)

10
2 4 6 8 10
(a)

(d)

FIGURE 7. The ComSAR algorithm. (a) A full coherence matrix in the PSDS technique. The data are divided into ministacks with 10 images
to improve the process efficiency. (b) The PL technique is employed to compress each ministack. This generates linked phases that enable
a coherent focus on the stack subset, resulting in a compressed image representing the first 10 images. This compression procedure is
repeated on the following 10 images, creating nine compressed images. These compressed images can be utilized to link prior ministacks
with new acquisitions and reconstruct the full phase time series without the need to recalculate everything. ComSAR can work with full and
compressed time series, but the (c) compressed version typically outperforms the (d) full time series version [16].

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 55


where the superscripts indicate the time series sequence source TomoSAR package [16], [54], [56]. To the best of
and jt cal (k) is the kth of the calibration vector. our knowledge, this package is the first publicly acces-
Figure 8(a) displays an original interferogram cover- sible tool that enables the simultaneous handling of
ing 670 days in Mexico City, Mexico, while Figure 8(b) is a both PS and DS targets (https://github.com/DinhHoTong
compressed version (with M = 10). Compared to the origi- Minh/TomoSAR).
nal [in Figure 8(d) and (e)], the compressed interferogram
exhibits superior quality and coherence. The average coher- PERSPECTIVE WITH DEEP LEARNING
ence improves from 0.4 to 0.8 [shown in Figure 8(c)], main- This section highlights the potential of the DL approach in
ly because the noise component of the data is reduced in the SAR data, emphasizing the value of DL in enhancing the
compression process. Essentially, the noisy short-lived com- accuracy and efficiency of PL. The “Brief Review of Deep
ponents are eliminated from the artificially compressed in- Learning for Synthetic Aperture Radar Data” section serves
terferograms created from the ministacks. Consequently, as a bridge, introducing DL’s potential for SAR data process-
these interferograms have a higher signal-to-noise ratio ing and paving the way for its application in improving PL
than the initial ones. techniques, in the “Perspective for Phase Linking” section.
Since the estimation of the linked phase is ambiguous,
we set the phase of the first image in each ministack to zero. BRIEF REVIEW OF DEEP LEARNING FOR
Finally, PS values at these multireference images will be SYNTHETIC APERTURE RADAR DATA
extracted from the original SLCs and integrated into com- In recent years, the increase in SAR Earth observation
pressed phase time series for PSDS analysis [16]. missions, such as TerraSar-X, Sentinel-1, ALOS, and RA-
An interesting development by Ho Tong Minh and DARSAT, has led to a new scenario characterized by the con-
Ngo is the implementation of the PSDS and ComSAR tinuous generation of a massive amount of data. On the one
algorithms, which have been made available as an open hand, this trend has allowed us to observe the inadequacy

–π +π

(a) (b)

× 104 Compressed Coherence


12
Raw Coherence
10

8
Frequency

2
(d) (e)
0
0 0.2 0.4 0.6 0.8 1
Coherence
(c)

FIGURE 8. Mexico interferograms. (a) A raw 670-day interferogram. (b) A compressed 670-day interferogram. (c) The coherence distribution.
(d) and (e) Zoomed-in versions of (a) and (b), respectively. The figure is adapted from [54, Fig. 3].

56 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


of classical algorithms in regard to generalization capabili- pixels. Notably, PSNet outperformed the StaMPS algorithm
ties and computational performance; on the other hand, by detecting more than double the number of PSs.
it paved the way for the new artificial intelligence para-
digm, including the DL approach [57]. DL has profoundly PERSPECTIVE FOR PHASE LINKING
impacted many scientific fields, such as machine vision, The traditional PL methods rely on handcrafted algorithms,
natural language processing, video processing, speech, im- which can be time-consuming and may not always provide
age processing, and Earth data science [58], [59]. Over the optimal results. DL has the potential to revolutionize the
past decade, significant advances have been made in devel- PL process in SAR imaging by providing an end-to-end so-
oping and applying DL techniques to problems in Earth lution that can learn the complex relationships between the
observation [59], [60], [61]. These techniques have proved interferometric phases and the ESR phase from data. The
highly effective for classification and parameter retrieval principle is quite simple for DL, which consists of multiple
tasks, as they can process large amounts of data and deal layers of interconnected nodes [58], [59], [73]. The general
efficiently with complex spatial and temporal structures DL for PL can include the first
[59]. DL has been applied to the field of SAR imaging. layer of the network extract-
CNNs, GANs, and RNNs are the leading neuron network ing low-level features from the
DL architectures that have been applied to SAR data analy- input data, such as the inter- TRADITIONAL PL METHODS
sis. CNNs have been widely used in various tasks, such as ferometric phases. The subse- RELY ON HANDCRAFTED
ship detection [62], building detection [63], deformation quent layers learn increasingly ALGORITHMS, WHICH CAN
observation [64], and land cover classification [61], [65], complex data representations BE TIME-CONSUMING AND
and they have shown to be effective in these tasks. On the until the final layer produces MAY NOT ALWAYS PROVIDE
other hand, GANs have been used for SAR image super- the desired output, such as the
OPTIMAL RESULTS.
resolution and to enhance the quality of SAR images [66]. ESR phase. Unfortunately, we
RNNs have been applied to classify time series SAR data cannot find any report on the
and have shown good results [60], [67]. DL approach for the PL tech-
Several studies have explored DL approaches for SAR fil- nique in the literature. For this reason, we discuss the pos-
tering with promising results. For instance, Mullissa et al. sibility of DL as a valuable tool for PL by addressing a few
[68] proposed a deSpeckNet DL-based approach for SAR selected points:
despeckling that achieved higher accuracy and efficiency ◗◗ Multipass SAR data are typically represented as com-
than traditional methods. Similarly, Wu et al. [69] used a plex-valued time series, which cannot be directly used
deep CNN for polarimetric SAR filtering and demonstrated as input to most DL models. One common approach is
improved filtering performance. Promising outcomes have to separate the real and imaginary components and treat
been achieved by exploring DL techniques in SAR PU tasks. them as separate input channels. Alternatively, derived
Traditional PU methods assume that the phase possesses quantities, such as magnitude and phase, can be input.
spatial continuity, but their effectiveness is hampered by In both ways, the topographic phase must be subtracted
decorrelation noise and aliasing fringes that invalidate such to improve the spatial stationarity in the homogeneous
assumptions. To enhance the reliability of unwrapping out- region before feeding in the DL model [16]. Specialized
comes, Wu et al. [70] proposed a deep CNN, known as a dis- NN architectures that handle complex-valued data di-
continuity estimation network, that predicts the probabilities rectly, such as complex CNNs, also exist [74]. In address-
of phase discontinuities in interferograms. Similarly, Zhou ing the PL problem, examining the data input approach
et al. [71] converted the PU problem into a learnable image is essential. Although few algorithms leverage the com-
semantic segmentation problem and presented a DL-based plex-valued nature of radar data [75], [ [76], separating
branch cut deployment approach (BCNet). Experimental the real and imaginary components or using magnitude
results demonstrate that the proposed BCNet-based PU and phase quantities as input are commonly used and
method is a near-real-time PU algorithm with higher accu- practical approaches.
racy than traditional PU methods. Accurate identification ◗◗ One advanced CNN model is U-Net, an encoder–decoder
of PSs is crucial in obtaining reliable phase information in CNN initially employed for semantic segmentation in
PSI and PSDS. To address this, a novel deep CNN named medical images [77]. For example, a U-Net model could
PSNet has been proposed by Tianxiang et al. [72] for PS be trained to take SAR interferograms as input and out-
identification. The significant advantage of PSNet lies in put unwrapped phase maps, using significant anno-
its deep architecture, which can learn the distinguishing tated phase and deformation maps datasets [78]. This
features of PSs from vast training images with diverse to- approach could improve the accuracy of the PU process
pography and landscapes. Using the combined feature im- and make it more robust to noise and other sources of
ages of the average amplitude, amplitude dispersion, and error. Similarly, an autoencoder architecture is designed
coherence of interferograms as inputs, PSNet was trained to effectively separate ground deformation signals from
to classify PS and non-PS pixels. The results demonstrate noise in InSAR time series without requiring any pri-
that PSNet accurately distinguishes between PS and non-PS or information about the location of a fault or its slip

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 57


behavior [64]. Indeed, U-Net is an architecture designed phase, and it can then generalize these relationships to
to learn a model in an end-to-end setting. U-Net’s en- new SAR images. The DL approach’s end-to-end nature
coder path compresses the input image information by means that a deep NN can perform the entire PL pro-
extrapolating relevant features computed at different cess, from feature extraction to ESR phase estimation.
resolution scales. As a result, this hierarchical feature ex- This results in a faster and more accurate PL process, as
traction provides various representations of abstraction the network can automatically learn the most relevant
levels. On the other hand, U-Net’s decoder path recon- features and relationships from the data [59], [73].
structs the original image by mapping the intermediate ◗◗ Exploiting DL for PL can bring two benefits. The first ad-
representation back to the input spatial resolution. In vantage is that it can handle nonstationary phase noise
particular, during this reconstruction process, the in- and outliers more effectively than traditional methods.
formation is restored at different resolutions by stack- Traditional PL methods rely on handcrafted algorithms,
ing many upsampling layers. This can preserve relevant which may not always be robust to nonstationary phase
information during the decoding stage [77]. In this way, noise and outliers. On the other hand, DL can learn to
the reconstructed image accuracy can be well preserved. handle these issues from the data during the training pro-
In this sense, there is a potential link between CNN cess [79], resulting in a more robust PL process. Another
U-Net models and PL. advantage is a reduction of the computational cost
◗◗ DL training can be performed by providing a large data- of the process. Traditional PL methods can be compu-
set of SAR images and their corresponding ESR phases. tationally expensive, especially with large SAR datasets.
The network will use these data to learn the relation- DL, however, can be implemented on parallel architec-
ships between the interferometric phases and the ESR tures, such as GPUs, which can significantly reduce the

(a) (b)

(c) (d)

(e) (f)
–π 0 +π

FIGURE 9. A synthetic example of PL using DL. (a) A simulated deformed signal for interferograms, using the first acquisition as the refer-
ence image. (b) Interferograms after adding decorrelation noise. (c) The results of the MLE method using all the interferograms.
(d) The residuals of the MLE method [i.e., the difference between (a) and (c)]. (e) The results of the DL method using the U-Net model.
(f) The residuals of the DL method [i.e., the difference between (a) and (e)].

58 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


computational cost [55]. This is particularly important for patch images from various landscapes. Additionally, uti-
near-real-time processing of big SAR data, where the com- lizing a more advanced U-Net model with greater depth
putational cost of the PL process is a significant concern. can significantly enhance the accuracy and efficiency of
In summary, DL holds considerable promise for vari- the PL process.
ous applications in PL, offering opportunities to enhance
precision estimation and computational efficiency. Never- CONCLUSIONS
theless, it is important to acknowledge the challenges and Accurate estimation of linked phases is crucial in mitigat-
limitations associated with its implementation. These in- ing decorrelation effects on SAR data in the PL technique.
clude data quality, feature extraction, model complexity, Researchers have proposed various algorithms, such as
and interpretability concerns. While DL presents an entic- quasi-Newton, CAESAR, and EMI, to improve PL estima-
ing avenue for future research trends in PL, it necessitates tion’s precision and computational efficiency. Addition-
a meticulous design and thorough evaluation of both the ally, new compression and regulation techniques have
models and the data involved. been developed to enhance the estimation of the coher-
ence matrix. PL is widely used in TomoSAR, PSDS, and
EXAMPLE WITH U-NET MODEL ComSAR applications, and the adoption of DL is expected
To demonstrate the applicability of DL in PL, we con- to improve the accuracy and efficiency of the process. The
ducted a proof-of-concept test on a synthetic dataset. future of the DL approach for PL is promising, as ongo-
The simulation settings included generating a radar data ing research in various areas can shape the development of
stack with temporal noise behaviors, consisting of 10 better algorithms and techniques. This will help improve
SLC images with a revisit time of 35 days. We simulated the detection and measurement of surface deformation
a deformation signal by assuming a simple Gaussian de- and parameter estimation in SAR applications, leading to
formation bowl with a maximum LOS deformation rate more accurate and efficient results.
of 14 mm/year at the center and a radius of 600 m. The
simulation was conducted on a flat area, resulting in a ACKNOWLEDGMENT
zero topographic signal. We assumed a crop of 1,280 # This work was supported, in part, by the ESA; Center Na-
1,280 m, a radar wavelength of 56 mm, and a pixel size of tional d’Etudes Spatiales/Terre, Ocean, Surfaces Continen-
20 # 20 m. We employed the coherence Sentinel-1 model tales, Atmosphere (project MekongInSAR); UMR TETIS;
to simulate the behavior of temporal coherence over time, and Institut National de Recherche en Agriculture, Alimen-
which generates a coherence matrix for a one-year time tation, et Environnement. The ALOS-2/PALSAR-2 data were
series of 10 temporally ordered measurements taken at kindly provided by the Japanese Aerospace Exploration
35-day intervals [21], [16]. Each SLC included 64 # 64 ho- Agency, under the third Research Announcement Program
mogeneous pixels, and the simulation was repeated 5,000 on Earth Observation, with project ER3A2N097.
times for DL data training. Figure 9(a) and (b) provides
a visual representation of both the noise-free and noisy AUTHOR INFORMATION
simulated datasets. Figure 9(c) and (d) demonstrates the Dinh Ho Tong Minh (dinh.ho-tong-minh@inrae.fr) is with
application of the MLE spectral regularization algorithm UMR TETIS, INRAE, University of Montpellier, 34090
using coherence matrix estimation over 11 # 11 windows Montpellier, France. He is a Member of IEEE.
and a b value of 0.5 (see the “Recent Advances” section). Stefano Tebaldini (stefano.tebaldini@polimi.it) is with
We used a MATLAB function called unetLayers to de- Dipartimento di Elettronica, Informazione, e Bioingegne-
fine the U-Net architecture, with a depth of three in the ria, Politecnico di Milano, 20133 Milano, Italy. He is a Se-
encoder network. The complex-valued interferometric nior Member of IEEE.
data were separated by magnitude and phase quantities as
input for the U-Net model, resulting in only 2N - 1 chan- REFERENCES
nels since the first phase is zero. The patch size was speci- [1] D. Ho Tong Minh, R. Hanssen, and F. Rocca, “Radar interfer-
fied as 64 # 64 # 19, and the number of output channels ometry: 20 years of development in time series techniques and
was nine ESR phases. We modified the U-Net architecture future perspectives,” Remote Sens., vol. 12, no. 9, Apr. 2020, Art.
by replacing the original softmax layer with a regression no. 1364, doi: 10.3390/rs12091364. [Online]. Available: https://
layer and trained the model on a dataset of 4,800, mea- www.mdpi.com/2072-4292/12/9/1364
surements with 200 for validation, setting MaxEpochs [2] R. F. Hanssen, Radar Interferometry: Data Interpretation and Error
as 10. Figure 9(e) and (f) indicates that the DL approach is Analysis. Dordrecht, The Netherlands: Kluwer, 2001.
expected to be comparable to the handcrafted MLE algo- [3] H. A. Zebker and J. Villasenor, “Decorrelation in interferomet-
rithm. However, it is important to note that the synthetic ric radar echoes,” IEEE Trans. Geosci. Remote Sens., vol. 30, no. 5,
simulation used in this example was relatively basic. For pp. 950–959, Sep. 1992, doi: 10.1109/36.175330.
future studies to effectively address the PL challenge us- [4] A. Ferretti, C. Prati, and F. Rocca, “Permanent scatterers in SAR
ing DL, it is crucial to have access to high-quality datasets interferometry,” IEEE Trans. Geosci. Remote Sens., vol. 39, no. 1,
containing both synthetic and real-world interferometric pp. 8–20, Jan. 2001, doi: 10.1109/36.898661.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 59


[5] A. Hooper, H. Zebker, P. Segall, and B. Kampes, “A new method [19] F. De Zan, A. Parizzi, P. Prats-Iraola, and P. López-Dekker,
for measuring deformation on volcanoes and other natural ter- “A SAR interferometric model for soil moisture,” IEEE Trans.
rains using InSAR persistent scatterers,” Geophysical Res. Lett., Geosci. Remote Sens., vol. 52, no. 1, pp. 418–425, Jan. 2014, doi:
vol. 31, no. 23, pp. 1–5, Dec. 2004, doi: 10.1029/2004GL021737. 10.1109/TGRS.2013.2241069.
[6] P. Berardino, G. Fornaro, R. Lanari, and E. Sansosti, “A new [20] S. Tebaldini, F. Rocca, M. Mariotti d’Alessandro, and L. Ferro-
algorithm for surface deformation monitoring based on small Famil, “Phase calibration of airborne tomographic SAR data via
baseline differential SAR interferograms,” IEEE Trans. Geosci. phase center double localization,” IEEE Trans. Geosci. Remote
Remote Sens., vol. 40, no. 11, pp. 2375–2383, Dec. 2002, doi: Sens., vol. 54, no. 3, pp. 1775–1792, Mar. 2016, doi: 10.1109/
10.1109/TGRS.2002.803792. TGRS.2015.2488358.
[7] M.-P. Doin et al., “Presentation of the small baseline NSBAS [21] H. Ansari, F. De Zan, and A. Parizzi, “Study of systematic bias in
processing chain on a case example: The Etna deformation measuring surface deformation with SAR interferometry,” IEEE
monitoring from 2003 to 2010 using Envisat data,” in Proc. Trans. Geosci. Remote Sens., vol. 59, no. 2, pp. 1285–1301, Feb.
Fringe ESA Conf. Workshop, Frascati, Italy, 2011, pp. 1–8. 2021, doi: 10.1109/TGRS.2020.3003421.
[8] Z. Yunjun, H. Fattahi, and F. Amelung, “Small baseline InSAR [22] H. Yu, Y. Lan, Z. Yuan, J. Xu, and H. Lee, “Phase unwrapping
time series analysis: Unwrapping error correction and noise re- in InSAR: A review,” IEEE Geosci. Remote Sens. Mag. (replaced
duction,” Comput. Geosci., vol. 133, Dec. 2019, Art. no. 104331, doi: Newsletter), vol. 7, no. 1, pp. 40–58, Mar. 2019, doi: 10.1109/
10.1016/j.cageo.2019.104331. [Online]. Available: http://www. MGRS.2018.2873644.
sciencedirect.com/science/article/pii/S0098300419304194 [23] D. Ho Tong Minh, T. Le Toan, F. Rocca, S. Tebaldini, M. Mariotti
[9] D. Ho Tong Minh, R. Hanssen, M.-P. Doin, and E. Pathier, Ad- d’Alessandro, and L. Villard, “Relating P-band synthetic aper-
vanced Methods for Time-Series InSAR. Hoboken, NJ, USA: Wiley, ture radar tomography to tropical forest biomass,” IEEE Trans.
2022, ch. 5, pp. 125–153. Geosci. Remote Sens., vol. 52, no. 2, pp. 967–979, Feb. 2014, doi:
[10] A. M. Guarnieri and S. Tebaldini, “On the exploitation of tar- 10.1109/TGRS.2013.2246170.
get statistics for SAR interferometry applications,” IEEE Trans. [24] A. Reigber and A. Moreira, “First demonstration of airborne
Geosci. Remote Sens., vol. 46, no. 11, pp. 3436–3443, Nov. 2008, SAR tomography using multibaseline L-band data,” IEEE Trans.
doi: 10.1109/TGRS.2008.2001756. Geosci. Remote Sens., vol. 38, no. 5, pp. 2142–2152, Sep. 2000,
[11] A. Ferretti, A. Fumagalli, F. Novali, C. Prati, F. Rocca, and A. doi: 10.1109/36.868873.
Rucci, “A new algorithm for processing interferometric data- [25] S. Tebaldini, D. Ho Tong Minh, M. M. d’Alessandro, L. Villard,
stacks: SqueeSAR,” IEEE Trans. Geosci. Remote Sens., vol. 49, no. T. Le Toan, and J. Chave, “The status of technologies to measure
9, pp. 3460–3470, Sep. 2011, doi: 10.1109/TGRS.2011.2124465. forest biomass and structural properties: State of the art in SAR
[12] R. Battiti and F. Masulli, BFGS Optimization for Faster and Auto- tomography of tropical forests,” Surv. Geophys., vol. 40, no. 4,
mated Supervised Learning. Dordrecht, The Netherlands: Spring- pp. 779–801, May 2019, doi: 10.1007/s10712-019-09539-7.
er Netherlands, 1990, pp. 757–760. [26] M. Pardini, M. Tello, V. Cazcarra-Bes, K. P. Papathanassiou, and
[13] N. Cao, H. Lee, and H. C. Jung, “Mathematical framework for I. Hajnsek, “L- and P-band 3-D SAR reflectivity profiles versus
phase-triangulation algorithms in distributed-scatterer inter- lidar waveforms: The AfriSAR case,” IEEE J. Sel. Topics Appl.
ferometry,” IEEE Geosci. Remote Sens. Lett., vol. 12, no. 9, pp. Earth Observ. Remote Sens., vol. 11, no. 10, pp. 1–16, Oct. 2018,
1838–1842, Sep. 2015, doi: 10.1109/LGRS.2015.2430752. doi: 10.1109/JSTARS.2018.2847033.
[14] G. Fornaro, S. Verde, D. Reale, and A. Pauciullo, “Caesar: An ap- [27] I. El Moussawi et al., “Monitoring tropical forest structure using
proach based on covariance matrix decomposition to improve SAR tomography at L- and P-band,” Remote Sens., vol. 11, no. 16,
multibaseline-multitemporal interferometric SAR processing,” Aug. 2019, Art. no. 1934, doi: 10.3390/rs11161934. [Online].
IEEE Trans. Geosci. Remote Sens., vol. 53, no. 4, pp. 2050–2065, Available: https://www.mdpi.com/2072-4292/11/16/1934
Apr. 2015, doi: 10.1109/TGRS.2014.2352853. [28] D. Ho Tong Minh, Y.-N. Ngo, and T. T. Lê, “Potential of
[15] H. Ansari, F. De Zan, and R. Bamler, “Efficient phase esti- P-band SAR tomography in forest type classification,” Remote
mation for interferogram stacks,” IEEE Trans. Geosci. Remote Sens., vol. 13, no. 4, Feb. 2021, Art. no. 696, doi: 10.3390/
Sens., vol. 56, no. 7, pp. 4109–4125, Jul. 2018, doi: 10.1109/ rs13040696. [Online]. Available: https://www.mdpi.com/2072
TGRS.2018.2826045. -4292/13/4/696
[16] D. Ho Tong Minh and Y.-N. Ngo, “Compressed SAR interfer- [29] D. Ho Tong Minh et al., “SAR tomography for the retrieval of
ometry in the big data era,” Remote Sens., vol. 14, no. 2, pp. forest biomass and height: Cross-validation at two tropical for-
1–13, Jan. 2022, doi: 10.3390/rs14020390. [Online]. Available: est sites in French Guiana,” Remote Sens. Environ., vol. 175, pp.
https://www.mdpi.com/2072-4292/14/2/390 138–147, Mar. 2016, doi: 10.1016/j.rse.2015.12.037.
[17] S. Zwieback, “Cheap, valid regularizers for improved interfero- [30] Y.-N. Ngo, Y. Huang, D. H. T. Minh, L. Ferro-Famil, I. Fayad,
metric phase linking,” IEEE Geosci. Remote Sens. Lett., vol. 19, and N. Baghdadi, “Tropical forest vertical structure character-
pp. 1–4, Aug. 2022, doi: 10.1109/LGRS.2022.3197423. ization: From GEDI to P-band SAR tomography,” IEEE Geosci.
[18] J. Biggs, T. Wright, Z. Lu, and B. Parsons, “Multi-interferogram Remote Sens. Lett., vol. 19, pp. 1–5, Sep. 2022, doi: 10.1109/
method for measuring interseismic deformation: Denali fault, LGRS.2022.3208744.
Alaska,” Geophysical J. Int., vol. 170, no. 3, pp. 1165–1179, Sep. [31] I. El Moussawi et al., “L-band UAVSAR tomographic imaging in
2007, doi: 10.1111/j.1365-246X.2007.03415.x. dense forests: Gabon forests,” Remote Sens., vol. 11, no. 5, Feb.

60 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


2019, Art. no. 475, doi: 10.3390/rs11050475. [Online]. Avail- rs10050794. [Online]. Available: https://www.mdpi.com/2072
able: https://www.mdpi.com/2072-4292/11/5/475 -4292/10/5/794
[32] V. Wasik, P. C. Dubois-Fernandez, C. Taillandier, and S. S. Saat- [43] K. Spaans and A. Hooper, “InSAR processing for volcano moni-
chi, “The AfriSAR campaign: Tomographic analysis with phase- toring and other near-real time applications,” J. Geophysical
screen correction for P-band acquisitions,” IEEE J. Sel. Topics Res., Solid Earth, vol. 121, no. 4, pp. 2947–2960, Apr. 2016, doi:
Appl. Earth Observ. Remote Sens., vol. 11, no. 10, pp. 1–13, Oct. 10.1002/2015JB012752. [Online]. Available: https://agupubs.
2018, doi: 10.1109/JSTARS.2018.2831441. onlinelibrary.wiley.com/doi/abs/10.1002/2015JB012752
[33] S. Tebaldini, “Algebraic synthesis of forest scenarios from [44] A. B. Narayan, A. Tiwari, R. Dwivedi, and O. Dikshit, “A novel
multibaseline polinsar data,” IEEE Trans. Geosci. Remote Sens., measure for categorization and optimal phase history retrieval
vol. 47, no. 12, pp. 4132–4142, Dec. 2009, doi: 10.1109/ of distributed scatterers for InSAR applications,” IEEE Trans.
TGRS.2009.2023785. Geosci. Remote Sens., vol. 56, no. 10, pp. 5843–5849, Oct. 2018,
[34] K. Goel and N. Adam, “A distributed scatterer interferom- doi: 10.1109/TGRS.2018.2826842.
etry approach for precision monitoring of known surface [45] J. Engelbrecht and M. R. Inggs, “Coherence optimization and its
deformation phenomena,” IEEE Trans. Geosci. Remote Sens., limitations for deformation monitoring in dynamic agricultur-
vol. 52, no. 9, pp. 5454–5468, Sep. 2014, doi: 10.1109/ al environments,” IEEE J. Sel. Topics Appl. Earth Observ. Remote
TGRS.2013.2289370. Sens., vol. 9, no. 12, pp. 5647–5654, Dec. 2016, doi: 10.1109/
[35] D. Ho Tong Minh, L. Van Trung, and T. L. Toan, “Mapping JSTARS.2016.2593946.
ground subsidence phenomena in Ho Chi Minh city through [46] A. G. Mullissa, D. Perissin, V. A. Tolpekin, and A. Stein, “Po-
the radar interferometry technique using ALOS PALSAR larimetry-based distributed scatterer processing method for PSI
data,” Remote Sens., vol. 7, no. 7, pp. 8543–8562, Jul. 2015, doi: applications,” IEEE Trans. Geosci. Remote Sens., vol. 56, no. 6,
10.3390/rs70708543. [Online]. Available: https://www.mdpi. pp. 3371–3382, Jun. 2018, doi: 10.1109/TGRS.2018.2798705.
com/2072-4292/7/7/8543 [47] G. Krieger et al., “Advanced L-band SAR system concepts for
[36] N. Cao, H. Lee, and H. C. Jung, “A phase-decomposition- high-resolution ultra-wide-swath SAR imaging,” in Proc. ESA
based PSInSAR processing method,” IEEE Trans. Geosci. Remote Adv. RF Sens. Remote Sens. Instrum. (ARSI), Sep. 2017. [Online].
Sens., vol. 54, no. 2, pp. 1074–1090, Feb. 2016, doi: 10.1109/ Available: https://elib.dlr.de/113598/
TGRS.2015.2473818. [48] J. Mittermayer, G. Krieger, A. Bojarski, M. Zonno, and A.
[37] J. Cohen-Waeber, R. Bürgmann, E. Chaussard, C. Giannico, Moreira, “A mirrorSAR case study based on the X-band high
and A. Ferretti, “Spatiotemporal patterns of precipitation-­ resolution wide swath satellite (HRWS),” in Proc. Eur. Conf. Syn-
modulated landslide deformation from independent compo- thetic Aperture Radar, 2021, pp. 1–6.
nent analysis of InSAR time series,” Geophys. Res. Lett., vol. 45, [49] R. Torres et al., “GMES Sentinel-1 mission,” Remote Sens. Environ.,
no. 4, pp. 1878–1887, Feb. 2018, doi: 10.1002/2017GL075950. vol. 120, pp. 9–24, May 2012, doi: 10.1016/j.rse.2011.05.028.
[Online]. Available: https://agupubs.onlinelibrary.wiley.com/ [Online]. Available: http://www.sciencedirect.com/science/
doi/abs/10.1002/2017GL075950 article/pii/S0034425712000600
[38] B. Fruneau, D. Ho Tong Minh, and D. Raucoules, Anthropogenic [50] P. A. Rosen, Y. Kim, R. Kumar, T. Misra, R. Bhan, and V. R. Sagi,
Activity: Monitoring Surface-Motion Consequences of Human Activi- “Global persistent SAR sampling with the NASA-ISRO SAR
ties With Spaceborne InSAR. Hoboken, NJ, USA: Wiley, 2022, ch. (NISAR) mission,” in Proc. IEEE Radar Conf. (RadarConf), May
1, pp. 283–313. 2017, pp. 0410–0414, doi: 10.1109/RADAR.2017.7944237.
[39] D. Ho Tong Minh et al., “Measuring ground subsidence in Ha [51] H. Ansari, F. De Zan, and R. Bamler, “Sequential estimator: To-
Noi through the radar interferometry technique using Ter- ward efficient InSAR time series analysis,” IEEE Trans. Geosci. Re-
raSAR-X and cosmos SkyMed data,” IEEE J. Sel. Topics Appl. Earth mote Sens., vol. 55, no. 10, pp. 5637–5652, Oct. 2017, doi: 10.1109/
Observ. Remote Sens., vol. 12, no. 10, pp. 3874–3884, Oct. 2019, TGRS.2017.2711037.
doi: 10.1109/JSTARS.2019.2937398. [52] I. T. Jolliffe, Principal Component Analysis. New York, NY, USA:
[40] M. Jiang, X. Ding, R. F. Hanssen, R. Malhotra, and L. Chang, Springer-Verlag, 2002.
“Fast statistically homogeneous pixel selection for covariance [53] L. R. Brigitte and H. Rouanet, Geometric Data Analysis, From Cor-
matrix estimation for multitemporal InSAR,” IEEE Trans. Geos- respondence Analysis to Structured Data Analysis. Dordrecht, The
ci. Remote Sens., vol. 53, no. 3, pp. 1213–1224, Mar. 2015, doi: Netherlands: Kluwer, 2004.
10.1109/TGRS.2014.2336237. [54] D. Ho Tong Minh and Y.-N. Ngo, “ComSAR: A new algorithm
[41] X. Lv, B. Yazıcı, M. Zeghal, V. Bennett, and T. Abdoun, “Joint- for processing big data SAR interferometry,” in Proc. IEEE Int.
scatterer processing for time-series InSAR,” IEEE Trans. Geosci. Geosci. Remote Sens. Symp. (IGARSS), 2021, pp. 820–823, doi:
Remote Sens., vol. 52, no. 11, pp. 7205–7221, Nov. 2014, doi: 10.1109/IGARSS47720.2021.9553675.
10.1109/TGRS.2014.2309346. [55] M. Pandey et al., “The transformational role of GPU comput-
[42] R. Shamshiri, H. Nahavandchi, M. Motagh, and A. Hooper, ing and deep learning in drug discovery,” Nature Mach. Intell.,
“Efficient ground surface displacement monitoring using vol. 4, no. 3, pp. 211–221, Mar. 2022, doi: 10.1038/s42256-022
­Sentinel-1 data: Integrating distributed scatterers (DS) identi- -00463-x.
fied using two-sample t-test with persistent scatterers (PS),” Re- [56] D. Ho Tong Minh and Y.-N. Ngo, “Tomosar platform supports
mote Sens., vol. 10, no. 5, May 2018, Art. no. 794, doi: 10.3390/ for Sentinel-1 tops persistent scatterers interferometry,” in Proc.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 61


IEEE Int. Geosci. Remote Sens. Symp. (IGARSS), Jul. 2017, pp. no. 1217, doi: 10.3390/rs10081217. [Online]. Available: https://
1680–1683, doi: 10.1109/IGARSS.2017.8127297. www.mdpi.com/2072-4292/10/8/1217
[57] Y. Xu et al., “Artificial intelligence: A powerful paradigm [68] A. G. Mullissa, D. Marcos, D. Tuia, M. Herold, and J. Reiche,
for scientific research,” Innovation, vol. 2, no. 4, Nov. 2021, “deSpeckNet: Generalizing deep learning based SAR image
Art. no. 100179, doi: 10.1016/j.xinn.2021.100179. [Online]. despeckling,” IEEE Trans. Geosci. Remote Sens., vol. 60, pp. 1–15,
Available: https://www.sciencedirect.com/science/article/pii/ 2022, doi: 10.1109/TGRS.2020.3042694.
S2666675821001041 [69] W. Wu, H. Li, L. Zhang, X. Li, and H. Guo, “High-resolution
[58] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, PolSAR scene classification with pretrained deep convnets and
vol. 521, no. 7553, pp. 436–444, May 2015, doi: 10.1038/na- manifold polarimetric parameters,” IEEE Trans. Geosci. Remote
ture14539. Sens., vol. 56, no. 10, pp. 6159–6168, Oct. 2018, doi: 10.1109/
[59] R. Baraniuk, D. Donoho, and M. Gavish, “The science of TGRS.2018.2833156.
deep learning,” Proc. Nat. Acad. Sci., vol. 117, no. 48, pp. [70] Z. Wu, T. Wang, Y. Wang, R. Wang, and D. Ge, “Deep-learning-
30,029–30,032, Nov. 2020, doi: 10.1073/pnas.2020596117. based phase discontinuity prediction for 2-D phase unwrap-
[Online]. Available: https://www.pnas.org/doi/abs/10.1073/ ping of SAR interferograms,” IEEE Trans. Geosci. Remote Sens.,
pnas.2020596117 vol. 60, pp. 1–16, 2022, doi: 10.1109/TGRS.2021.3121906.
[60] D. Ho Tong Minh et al., “Deep recurrent neural networks for [71] L. Zhou, H. Yu, Y. Lan, and M. Xing, “Deep learning-based
winter vegetation quality mapping via multitemporal SAR branch-cut method for InSAR two-dimensional phase unwrap-
Sentinel-1,” IEEE Geosci. Remote Sens. Lett., vol. 15, no. 3, pp. ping,” IEEE Trans. Geosci. Remote Sens., vol. 60, pp. 1–15, 2022,
464–468, Mar. 2018, doi: 10.1109/LGRS.2018.2794581. doi: 10.1109/TGRS.2021.3099997.
[61] D. Ienco, R. Interdonato, R. Gaetano, and D. Ho Tong Minh, [72] T. Yang, H. Yu, and Y. Wang, “Selection of persistent scatterers
“Combining Sentinel-1 and Sentinel-2 satellite image time se- with a deep convolutional neural network,” in Proc. IEEE Int.
ries for land cover mapping via a multi-source deep learning Geosci. Remote Sens. Symp. (IGARSS), 2022, pp. 2912–2915, doi:
architecture,” ISPRS J. Photogrammetry Remote Sens., vol. 158, 10.1109/IGARSS46834.2022.9884199.
pp. 11–22, Dec. 2019, doi: 10.1016/j.isprsjprs.2019.09.016. [On- [73] L. Zhang and B. Du, “Deep learning for remote sensing data:
line]. Available: https://www.sciencedirect.com/science/article/ A technical tutorial on the state of the art,” IEEE Geosci. Remote
pii/S0924271619302278 Sens. Mag. (replaced Newsletter), vol. 4, no. 2, pp. 22–40, Jun.
[62] Z. Lin, K. Ji, X. Leng, and G. Kuang, “Squeeze and excitation 2016, doi: 10.1109/MGRS.2016.2540798.
rank faster R-CNN for ship detection in SAR images,” IEEE Geos- [74] J. A. Barrachina, Nov. 2022, “Negu93/CVNN: Complex-valued
ci. Remote Sens. Lett., vol. 16, no. 5, pp. 751–755, May 2019, doi: neural networks,” Zenodo, doi: 10.5281/zenodo.7303587.
10.1109/LGRS.2018.2882551. [75] T. Scarnati and B. Lewis, “Complex-valued neural networks for
[63] H. Li, F. Zhu, X. Zheng, M. Liu, and G. Chen, “MSCDUNet: A synthetic aperture radar image classification,” in Proc. IEEE Ra-
deep learning framework for built-up area change detection dar Conf. (RadarConf), 2021, pp. 1–6, doi: 10.1109/RadarConf
integrating multispectral, SAR, and VHR data,” IEEE J. Sel. Top- 2147009.2021.9455316.
ics Appl. Earth Observ. Remote Sens., vol. 15, pp. 5163–5176, Jun. [76] Y. Zhang, H. Yuan, H. Li, C. Wei, and C. Yao, “Complex-valued
2022, doi: 10.1109/JSTARS.2022.3181155. graph neural network on space target classification for defo-
[64] B. Rouet-Leduc, R. Jolivet, M. Dalaison, P. A. Johnson, and C. cused ISAR images,” IEEE Geosci. Remote Sens. Lett., vol. 19,
Hulbert, “Autonomous extraction of millimeter-scale deforma- pp. 1–5, Jun. 2022, doi: 10.1109/LGRS.2022.3185709.
tion in InSAR time series using deep learning,” Nature Com- [77] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional
mun., vol. 12, Nov. 2021, Art. no. 6480. networks for biomedical image segmentation,” 2015. [Online].
[65] Y. J. E. Gbodjo, O. Montet, D. Ienco, R. Gaetano, and S. Du- Available: https://arxiv.org/abs/1505.04597
puy, “Multisensor land cover classification with sparsely an- [78] H. H. Zeyada, M. S. Mostafa, M. M. Ezz, A. H. Nasr, and H.
notated data based on convolutional neural networks and M. Harb, “Resolving phase unwrapping in interferomet-
self-distillation,” IEEE J. Sel. Topics Appl. Earth Observ. Remote ric synthetic aperture radar using deep recurrent residual
Sens., vol. 14, pp. 11,485–11,499, Oct. 2021, doi: 10.1109/ U-Net,” Egyptian J. Remote Sens. Space Sci., vol. 25, no. 1,
JSTARS.2021.3119191. pp. 1–10, Feb. 2022, doi: 10.1016/j.ejrs.2021.12.001. [Online].
[66] C. Zhang et al., “Blind super-resolution for SAR images with Available: https://www.sciencedirect.com/science/article/pii/
speckle noise based on deep learning probabilistic degradation S1110982321000958
model and SAR priors,” Remote Sens., vol. 15, no. 2, Jan. 2023, [79] F. Zhou, H. Zhou, Z. Yang, and L. Gu, “IF2CNN: Towards
Art. no. 330, doi: 10.3390/rs15020330. [Online]. Available: non-stationary time series feature extraction by integrating
https://www.mdpi.com/2072-4292/15/2/330 iterative filtering and convolutional neural networks,” Expert
[67] E. Ndikumana, D. Ho Tong Minh, N. Baghdadi, D. Courault, Syst. Appl., vol. 170, May 2021, Art. no. 114527, doi: 10.1016/j.
and L. Hossard, “Deep recurrent neural network for agricul- eswa.2020.114527. [Online]. Available: https://www.science
tural classification using multitemporal SAR Sentinel-1 for Ca- direct.com/science/article/pii/S0957417420311714
margue, France,” Remote Sens., vol. 10, no. 8, Aug. 2018, Art. GRS

62 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


IMAGE LICENSED BY INGRAM PUBLISHING

There Are No Data


Like More Data
Datasets for deep learning in Earth observation

MICHAEL SCHMITT , SEYED ALI AHMADI , YONGHAO XU , GÜLŞEN TAŞKIN ,


UJJWAL VERMA , FRANCESCOPAOLO SICA , AND RONNY HÄNSCH

C arefully curated and annotated datasets are the foun-


dation of machine learning (ML), with particularly
data-hungry deep neural networks forming the core of
the spotlight. Based on a review of historical develop-
ments, currently available resources are described and a
perspective for future developments is formed. We hope
what is often called artificial intelligence (AI). Due to the to contribute to an understanding that the nature of
massive success of deep learning (DL) applied to Earth our data is what distinguishes the EO community from
observation (EO) problems, the focus of the community many other communities that apply DL techniques to
has been largely on the development of evermore so- image data, and that a detailed understanding of EO
phisticated deep neural network architectures and train- data peculiarities is among the core competencies of
ing strategies. For that purpose, numerous task-specific our discipline.
datasets have been created that were largely ignored by
previously published review articles on AI for EO. With INTRODUCTION
this article, we want to change the perspective and put DL techniques have enabled drastic improvements in many
ML datasets dedicated to EO data and applications into scientific fields, especially in those dedicated to the analy-
sis of image data, e.g., computer vision or remote sensing.
Digital Object Identifier 10.1109/MGRS.2023.3293459
Although it was possible to train shallow learning approach-
Date of current version: 8 August 2023 es on comparably small datasets, deep learning requires

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 2473-2397/23©2023IEEE 63


large-scale datasets to reach the desired accuracy and gen- numerous remote sensing datasets, each consisting
eralization performance. Therefore, the availability of an- of a particular combination of sensor modalities, ap-
notated datasets has become a dominating factor for many plications, and geographic locations. Yet, a review of
cases of modern EO data analysis that develops and evalu- these developments is still missing in the literature.
ates powerful, DL-based techniques for the automated inter- The only articles that make a small step toward a gen-
pretation of remote sensing data. eral review of benchmark datasets are [1], [2], [3], and
The main goal of general computer vision is the [4]. All of them provide some sort of review, however,
analysis of optical images, such as photos, which con- they are always limited to a very narrow aspect, e.g.,
tain everyday objects, e.g., furniture, animals, or road object detection or scene classification. Furthermore,
signs. Remote sensing involves a larger variety of sensor their focus is on ML approaches and their correspond-
modalities and image analysis tasks than conventional ing datasets, while the historical evolution of datasets
computer vision, rendering the annotation of remote is neither discussed in detail, nor from a sensor- and
sensing data more difficult and costly. Besides classi- task-agnostic point of view.
cal optical images, multi- or hyperspectral sensors and As an extension of our 2021 IEEE International Sympo-
different kinds of infrared sensors; active sensor tech- sium on Geoscience and Remote Sensing contribution [5],
nologies such as laser scanning, microwave altimeters, this article intends to close this gap by
and synthetic aperture radar (SAR) are regularly used, ◗◗ reviewing current developments in the creation of datas-
too. The fields of application range from computer ets for DL applications in remote sensing and EO
vision-like tasks, such as object detection and classifica- ◗◗ structuring existing datasets and discussing their prop-
tion, to semantic segmentation (mainly for land cover erties
mapping) to specialized regression tasks grounded in the ◗◗ providing a perspective on future requirements.
physics of the used remote sensing system. In this context, we additionally present the Earth
To provide an illustrative example, a dataset for bio- Observation Database (EOD) [6], which is the result of
mass regression from interferometric SAR data will adopt the effort and cooperation of voluntary scientists with-
imagery and annotations very different from the ones in the IEEE Geoscience and Remote Sensing Society
needed for the semantic segmentation of urban land cover (GRSS) Image Analysis and Data Fusion (IADF) Techni-
types from multispectral optical data. Thus, although ex- cal Committee (TC). This database aims to function as
tensive image databases, such as ImageNet, were created a centralized tool that organizes the meta information
more than 10 years ago and form the backbone of many about existing datasets in a community-driven manner.
modern ML developments in computer vision, there is
still no similar dataset or backbone network in remote EVOLUTION OF EO-ORIENTED ML DATASETS
sensing. (Note that, as a prime example of an annotated
computer vision dataset, ImageNet contains more than HISTORICAL DEVELOPMENT
14 million images depicting objects from more than High-quality benchmark datasets have played an increas-
20,000 categories.) This lack of generality renders the gen- ingly important role in EO for quite some time and are one
eration of an ImageNet-like general EO dataset extreme- of the driving factors for the recent success of DL approaches
ly complicated and thus costly: instead of photographs to analyze remote sensing data. As such, they can be seen as
openly accessible on the Internet, many different—and a tool complementary to methodological advancements to
sometimes quite expensive—sensor data would have to push accuracy, robustness, and generalizability. This section
be acquired and, instead of “mechanical turks,” trained reviews and summarizes the historical development of EO-
EO experts would have to be hired to link these different oriented ML datasets to provide insights into the evolution
sensor data to the multitude of different domain- and of this “tool,” ranging from its historical beginnings to the
task-specific annotations (see “The Mechanical Turk”). current state of the art.
Therefore, until now, the trend in ML applied to The beginnings of ML applied to remote sensing focused
EO data has been characterized by the generation of on specific applications. Datasets were mainly built by con-
sidering a very localized study site, a few specific sensor mo-
dalities, and a relatively small number of acquired samples.
The Mechanical Turk Therefore, the first datasets were relatively small compared
The name mechanical turk comes from a fraudulent chess-playing machine to what is now considered a benchmarking dataset. Train-
developed in the 18th century. Chess players were made to believe they played ing, validation, and testing samples were often taken from
against the machine, but were in fact competing against a person hidden inside the same image. Even with the use of sophisticated shallow
it. Today, the term mostly refers to Amazon Mechanical Turk (MTurk), a crowd- learning models but especially since the advent of DL, such
sourcing website run by Amazon. On MTurk, users can hire remotely located small datasets were no longer sufficient for proper training
crowdworkers to perform desired tasks. MTurk is frequently used to create and evaluation. The need for extended datasets has led to
manual annotations for supervised machine learning tasks. the creation of larger datasets containing multiple images,
often acquired at different geographic locations.

64 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Figure 1 illustrates the evolution of benchmark datas- benchmarking datasets, we gathered an extensive list of
ets for ML in EO by showing this temporal development. datasets available in the EO community, resulting in a
To provide a broad view about the recent evolution of large collection with currently 400 (i.e., 380 image-based

7,358

Change Detection
Superresolution/Pansharpening
Object Detection 76

Classification
Semantic Segmentation
191
Change Detection and Classification
352
1,386 Object Detection and Classification
Object Detection and Semantic Segmentation 338 102
260
Classification and Semantic Segmentation 341

Others 44 45 98
359
273 263
312
20
282
361 48 264
356 158 1
25997
50
47 281
208 287 288 151 54 49

256
40 160 46
325 373
Volume (GB)

199 332
292 132 320 235
331 236
107
35 42 262 371
13 36
255 41 311 303
53 127
216 349 217
73 114 37 5719
286 152 280 328 218
145
141 272370 172
224 244 251
22 77 294 119 327
220 223 229 111 270 368
109 314 206 261 38 90 175
219 269 8 283
147 333 335 243
27 266 61
12 274 18
169 309 239138
334 161 300 173
106 143
148 124 39 358 205253 174
82 339 189 231 32 201
142 291 308 301
128 364 5 221 242 135 17
366 55 168 91
30 78 240 101
234 200 336 171
153 296 207
290 3 33 15 321
69 131 6306 277 94 121 194
130 113 157 350 265
2 71 79 52 118 211
9 4 210 230 250
355 209 238 233 317
28 29 75 110 154 293 86 192 275 357
1 104 182 225 156 241 87 323 139
72 232 276
31 184 226 318 83 295 237 170 163
348
95
203 190 193 304
363 11 155 16 248 56
316 67 150 43
14 247
63 149 186 60
108 185 34 271 202 372
284
80 187 144 188 120
116
25 285 93
367
64 298 165
23 66 362
146 84 299 166 167
322
24 105 245
26 297 134
22
254 85
51
21 222 329
65 289
0
2006

2007

2008

2009

2010

2011

2012

2013

2014

2015

2016

2017

2018

2019

2020

2021

2022

Publication Year

FIGURE 1. Distribution of remote sensing datasets over the years. The x-axis shows the publication year (the datasets are placed within the
region of their publication year with a small random offset to minimize visual overlap in the graph), while the y-axis represents the volume
of each dataset in gigabytes on a logarithmic scale. The circle radius indicates the dataset size in terms of the number of pixels. Colors
denote the type of task addressed by a dataset. Each circle is accompanied by an index, allowing for identification of the dataset in the
database (see Table 2), which provides further information.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 65


and 20 point cloud-based datasets) entries (see Tables 1 please refer to Figure S1 and “How to Measure the Size of
and 2), including related metadata. We point out that, al- a Dataset.”
though being extensive, this list is far from being complete Figure 1 provides a straightforward overview of the pro-
due to the fact that a large number of new datasets are portion among size and spatial dimension, and therefore
published every year. Furthermore, the metadata required about the overall information content given by features
to generate the plot of Figure 1 are available only for a sub- such as resolution, sensor modalities, number of bands/
set of 290 datasets (roughly 73%). In the horizontal axis, channels, and so on. Each circle is accompanied by an in-
we indicate the year of publication. The vertical axis shows dex, allowing for identification of the dataset in the data-
the volume of a dataset, while the circle radius reflects the base (see Table 2) that provides further information. Note
number of spatial pixels covered by a dataset. For a more that we use the category “Others” for datasets that do not
detailed explanation of how we measure dataset size, belong to any of the other categories and are too rare to

TABLE 1. ALTHOUGH A THOROUGH ANALYSIS OF LIDAR DATASETS IS BEYOND THE SCOPE OF THIS SURVEY, WE DO PROVIDE AN
OVERVIEW OF SEVERAL EXAMPLE DATASETS. POINT CLOUD DATASETS ARE ANOTHER LARGE GROUP OF BENCHMARK DATA THAT ARE
WIDELY USED IN THE LITERATURE AND INDUSTRY. WITHIN EO THE MOST COMMON SOURCE FOR POINT CLOUD DATA ARE LIDAR SEN-
SORS THAT USE LIGHT IN THE FORM OF LASER PULSES TO MEASURE THE DISTANCE TO THE SURFACE. THE PRIMARY SOURCES ARE
AIRBORNE LASER SCANNING (ALS), TERRESTRIAL LASER SCANNING (TLS), AND MOBILE LASER SCANNING (MLS) DEVICES. OTHER
SOURCES OF POINT CLOUDS AND 3D DATA INCLUDE PHOTOGRAMMETRIC METHODS (STRUCTURE FROM MOTION, MULTI-VIEW STE-
REO, AND DENSE MATCHING APPROACHES) AND TOMOGRAPHIC SAR. AS 3D DATA TYPICALLY COME WITH FEATURES THAT ARE VERY
DIFFERENT FROM 2D IMAGE DATA, SUCH DATASETS ARE BEYOND THE SCOPE OF THIS ARTICLE. NEVERTHELESS, TABLE 1 PROVIDES A
SHORT LIST OF EXAMPLE LIDAR/POINT CLOUD DATASETS FOR INTERESTED READERS.

PUBLICATION POINT DENSITY NUMBER NUMBER OF VOLUME


TASK PLATFORM TIMESTAMPS NAME DATE (POINTS/M2) OF CLASSES POINTS (MB)
Change detection ALS Multiple Abenberg 2013 16 — 5,400,000 258
ALS
Classification ALS Single NEWFOR 2015 Varies Four — 97
Classification ALS Single DFC19 2019 — Six 167,400,000 613
Classification ALS Single ISPRS 3D 2014 8 Nine 780,879 —
Vaihingen
Classification Multiple Single ArCH 2020 Varies 10 136,138,423 —
Classification/seman- ALS Single DublinCity 2019 240–348 13 260,000,000 3,000
tic segmentation
Filtering ALS Single OpenGF 2021 6 and 14 Three 542,100,000 2,280
Object detection/se- TLS Single LiSurveying 2021 Varies 54 2,450,000,000 —
mantic segmentation
Others ALS Single RoofN3D 2018 4.72 Three 118,100 —
Semantic ALS Single LASDU 2020 3–4 Six 3,120,000 —
segmentation
Semantic ALS Single DALES 2020 50 Eight 505,000,000 4,000
segmentation
Semantic ALS Single DALES 2021 50 Eight 492,000,000 5,000
segmentation Object
Semantic Drone Single Campus3D 2020 Varies 24 937,000,000 2,500
segmentation
Semantic Drone Multiple Hessigheim 2021 800 11 73,909,354 5,950
segmentation 3D
Semantic Drone Single WildForest3D 2022 60 Six 7,000,000 81
segmentation
Semantic MLS Single Toronto3D 2020 1,000 Eight 7,830,0000 1,100
segmentation
Semantic MLS Multiple HelixNet 2022 — Nine 8,850,000,000 235,700
segmentation
Semantic Photogram- Single SensatUrban 2020 — 13 2,847,000,000 36,000
segmentation metry
Semantic Photogram- Single STPLS3D 2022 — 20 — 36,600
segmentation metry (images:
700,000)
Semantic TLS Single Semantic 3D 2017 — Eight 4000000000 23,940
segmentation

66 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


TABLE 2. ALTHOUGH THE COMPLETE COLLECTION OF 380 DATASETS IS TOO EXTENSIVE TO BE INCLUDED IN THE PRINT VERSION OF THIS ARTICLE, IT CAN BE FOUND AT [85]. HERE WE
INCLUDE A PORTION OF THIS LIST, WHICH CONTAINS ALL THE DATASETS MENTIONED IN THE ARTICLE, ALONG WITH THEIR INDEX IN THE COMPLETE DATABASE.

PUBLICATION NUMBER OF IMAGE NUMBER OF VOLUME

SEPTEMBER 2023
INDEX TASK PLATFORM SENSOR TYPE NAME DATE TIMESTAMPS IMAGES SIZE SIZE CLASSES (MB)
1 Change detection Aerial Multiple DFC21-MSD 2021 Multiple 2,250 4,000 36,000,000,000 15 325,000
10 Change detection Satellite Multiple DFC09 2009 Multiple Two 98 — — —
11 Change detection Satellite Multispectral OneraCD 2018 Multiple 24 600 8,640,000 Two 489
15 Change detection Satellite Optical LEVIR-CD 2020 Single 637 1,024 667,942,912 One 2,700
21 Classification Aerial Hyperspectral Indian Pines 2000 Single One 145 21,025 16 Six
22 Classification Aerial Hyperspectral Salinas 2000 Single One 365 111,104 16 27
24 Classification Aerial Hyperspectral Kennedy Space Center 2005 Single One 550 311,100 13 57
34 Classification Drone Optical AIDER 2019 Single 2,645 240 152,352,000 Four 275
40 Classification Satellite Multiple BigEarthNet-MM 2019 Single 590,326 × 12 120 1.02E + 11 19 121,000
52 Classification Satellite Multispectral EuroSAT 2018 Single 27,000 × 13 64 1,437,696,000 10 1,920
63 Classification Satellite Optical University of California, 2010 Single 2,100 256 137,625,600 21 317
Merced
69 Classification Satellite Optical AID 2017 Single 10,000 600 3,600,000,000 30 2,440
76 Classification Satellite Optical FMoW 2018 Single 523,846 — 1.08E + 12 63 3,500,000

IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE


98 Cloud removal Satellite Multiple SEN12MS-CR-TS 2021 Multiple 53 4,000 848,000,000 — 649,000
105 Object detection Aerial Optical SZTAKI AirChange 2012 Multiple 13 800 7920640 Two 42
109 Object detection Aerial Optical DOTA v1.0 2018 Single 2,806 4,000 44,896,000,000 15 18,000
114 Object detection Aerial Optical DOTA v2.0 2020 Single 11,268 4,000 1.80E + 11 18 34,280
117 Object detection Aerial Optical RSOC 2020 Single 3,057 2,500 3,621,481,392 Four —
127 Object detection Drone Multispectral BIRDSAI 2020 Multiple 162,000 640 49,766,400,000 — 43,200
138 Object detection Drone Optical DroneCrowd 2022 Multiple 33,600 1,920 69,672,960,000 One 10,400
151 Object detection Satellite Optical SpaceNet-4 2018 Single 60,000 900 48,600,000,000 186,000 —
162 Object detection Satellite Optical AI-TOD 2020 Single 28,036 — — — 22,000
191 Object detection Satellite SAR xView3-SAR 2021 Single 991 × 2 27,000 1.42E + 12 Two 2,000,000
204 Object detection/ Satellite SAR FUSAR-Ship 2020 Single 5,000 512 1,310,720,000 113 —
classification
208 Semantic segmentation Aerial Hyperspectral DFC08 2008 Single Five 194 — — —
211 Semantic segmentation Aerial Hyperspectral HOSD 2022 Single 18 × 224 1,700 3,622,355,072 One 2,200
212 Semantic segmentation Aerial Multiple DFC13 2013 Single 145 16 1,000 — —
213 Semantic segmentation Aerial Multiple DFC14 2014 Single Seven — — — —
214 Semantic segmentation Aerial Multiple DFC15 2015 Single Seven 10,000 700,000,000 — —
215 Semantic segmentation Aerial Multiple DFC18 2018 Single One 20 19,763 — —

(Continued)

67
68
TABLE 2. ALTHOUGH THE COMPLETE COLLECTION OF 380 DATASETS IS TOO EXTENSIVE TO BE INCLUDED IN THE PRINT VERSION OF THIS ARTICLE, IT CAN BE FOUND AT [85]. HERE WE
INCLUDE A PORTION OF THIS LIST, WHICH CONTAINS ALL THE DATASETS MENTIONED IN THE ARTICLE, ALONG WITH THEIR INDEX IN THE COMPLETE DATABASE. (Continued)

PUBLICATION NUMBER OF IMAGE NUMBER OF VOLUME


INDEX TASK PLATFORM SENSOR TYPE NAME DATE TIMESTAMPS IMAGES SIZE SIZE CLASSES (MB)
217 Semantic segmentation Aerial Multiple DFC22-SSL 2022 Single 5,000 2,000 20,000,000,000 14 42,000
219 Semantic segmentation Aerial Multispectral ISPRS 2 D-Potsdam 2011 Single 38 6,000 1,368,000,000 Six 16,000
220 Semantic segmentation Aerial Multispectral ISPRS 2 D-Vaihingen 2011 Single 33 2,200 156,750,000 Six 17,000
246 Semantic segmentation Multiple Multiple DFC17 2017 Multiple 17 — — — —
247 Semantic segmentation Multiple Multiple SpaceNet-6 2020 Single 3,401 900 275,4810,000 One 368
253 Semantic segmentation Multiple Optical OpenEarthMap 2022 Single 5,000 1,024 5,242,880,000 Eight 9,100
254 Semantic segmentation Satellite Multiple DFC07 2007 Single One 787 619,369 19 —
257 Semantic segmentation Satellite Multiple DFC20 2020 Single 180,662 17 19,200 — —
261 Semantic segmentation Satellite Multiple DFC21-DSE 2021 Single 6,000 16 1,536,000 Four 18,000
264 Semantic segmentation Satellite Multiple MapInWild 2022 Single 1,018 × 8 1,920 30,022,041,600 11 365,000
282 Semantic segmentation Satellite Multispectral OpenSentinelMap 2022 Multiple 137,045 192 5,052,026,880 15 445,000
286 Semantic segmentation Satellite Optical SpaceNet-1 2016 Single 9,735 650 4,113,037,500 One 31,000
287 Semantic segmentation Satellite Optical SpaceNet-2 2017 Single 24,586 650 10,387,585,000 One 182,200
288 Semantic segmentation Satellite Optical SpaceNet-3 2017 Single 3,711 1,300 6,271,590,000 One 182,200
292 Semantic segmentation Satellite Optical SpaceNet-5 2019 Single 2,369 1,300 4,003,610,000 One 84,103
294 Semantic segmentation Satellite Optical SpaceNet-7 2020 Multiple 1,525 1,024 1,599,078,400 One 20,582
301 Semantic segmentation Satellite Optical SpaceNet-8 2022 Multiple 1,200 1,200 3,456,000,000 Two 6,800
312 Semantic segmenta- Satellite Multiple SEN12MS 2019 Single 180,662 256 11,839,864,832 — 510,000
tion, classification
313 Semantic segmenta- Satellite Multiple DFC23 2023 Single 12 — — — —
tion, classification
316 Superresolution, pan- Aerial Multispectral DFC06 2006 Single Six 5,000 637,500,000 — 390
sharpening
318 Superresolution, pan- Satellite Multispectral Proba-V Superresolution 2018 Single 1,160 384 171,048,960 — 692
sharpening
319 Superresolution, pan- Satellite Multispectral PAirMax (Airbus) 2021 Single — — — — 153
sharpening
324 Superresolution, pan- Satellite Optical PAirMax (Maxar) 2021 Single 14 — — — 386
sharpening
325 Superresolution, pan- Satellite Optical WorldStrat 2022 Single 3,928 1,054 4,363,678,048 — 107,000

IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE


sharpening
340 Others Multiple Multiple DFC19 2019 Single — — — — —
360 Others Satellite Optical DFC16 2016 Multiple — — — — —

SEPTEMBER 2023
form their own category. Examples in the “Others” category Figure 2 provides further insights into the size distri-
are datasets on cloud removal, visual question answering, bution of available datasets in terms of the 1) number
and parameter-estimation tasks such as hurricane wind of pixels and 2) data volume in gigabytes for typical
speed prediction, satellite pose estimation, and vegetation remote sensing tasks. Illustrating these two different
phenological change monitoring. The dashed line (“—”) il- aspects allows a deeper understanding of the nature
lustrates an exponential growth of benchmark datasets cre-
ated by and for the EO community.
This map on the evolution of remote sensing datasets How to Measure the Size of a Dataset
offers several interesting insights: In this article, we look at the size of datasets from the following two perspectives:
◗◗ The beginnings: In addition to the first IEEE GRSS Data Fu- 1) Size: data volume in terms of the number of spatial pixels. We count the num-
sion Contest in 2006 (Table 2; number 316), there are a few ber of pixels in the highest-available image resolution while ignoring mul-
other pioneering datasets that have fostered ML research tiband, multichannel, and multisensor data. In other words, pixels are only
applied to remote sensing data in its early stages, e.g., counted once in the spatial coverage provided by the dataset.
•• Hyperspectral datasets [(Indian Pines, Salinas Valley, 2) Volume: data volume in terms of storage. The amount of disk space required
and Kennedy Space Center) (Table 2; 21, 22, and 24)]: for a dataset is a proxy for image resolution and the provided modalities (e.g.,
Published before 2005, these datasets triggered the multiple bands and sensor types).
ML era in remote sensing. Covering a very small area Figure S1 highlights the different factors that affect the volume and size of a
on the ground and having a very small number of dataset: the number of bits per pixel (radiometric resolution), number of spectral
pixels, such datasets are not suitable for training DL bands (spectral resolution; i.e., red, green, blue; multispectral; or hyperspectral),
models (or have to be used with excessive caution). number of images during a specific time period (temporal resolution), and num-
On the other hand, due to their rich hyperspectral ber of pixels per unit area (spatial resolution). As mentioned previously, the size is
information, they are still being used for tasks such directly related to the unique number of ground-projected resolution cells. A
as dimensionality reduction and feature extraction. larger dataset in terms of size corresponds to images with higher resolutions or
•• The University of California, Merced dataset (Table 2; broader coverage.
63) [7]: Published in 2010, it was the first dataset
dedicated to scene classification.
•• ISPRS Potsdam/Vaihingen dataset (Table 2; 219 and 220)
[8]: Published in 2012, it was initially intended to bench- Satellites
mark semantic segmentation approaches tailored to
aerial imagery. Later, it was also used for other tasks, e.g.,
single-image height reconstruction (e.g., in [9]).
•• SZTAKI-AirChange dataset (Table 2; 105) [10]: Pub-
Airplanes t1 t2 ... tn
lished in 2011, it was one of the earliest datasets de-
signed for object detection.
All of those pioneering datasets have seen massive use in
Drones
the early ML-oriented EO literature. It is interesting to note Time Series
that pansharpening, scene classification, semantic segmen- Multispectral Hyper-
tation, and object detection were the first topics in remote Width Spectral
sensing to be addressed using ML-based methodologies. t
igh Volume
◗◗ The DL boom: As discussed by several review articles on He
DL and AI applied to EO [11], [12], [13], the year 2015 Resolution
marked the beginning of the DL boom in the EO com-
munity. This is well reflected by a significantly rising
number of datasets published from that year onward. It is Images Projected
furthermore confirmed by the fact that the dataset sizes, onto Ground Surface
both in terms of spatial pixels and data volume, started
to increase significantly from approximately that time.
◗◗ The diversity of tasks: From the early days to the present,
ML-oriented EO datasets have been designed for a mul-
titude of different tasks. The historical evolution depict-
ed in Figure 1 further shows that object detection and
semantic segmentation are the most popular tasks, with FIGURE S1. A schematic illustration of the proposed size measure
a significant increase of datasets dedicated to minor- used to characterize datasets i.e., pixels are only counted once in
ity categories (denoted as “Others”) from roughly 2019 the spatial coverage provided by a dataset. For a more detailed
to the present. This indicates that the rise of DL in EO definition, see “How to Measure the Size of a Dataset.”
broadens the overall application scope of the discipline.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 69


of the data. For example, datasets counting a similar again confirms its popularity in DL-oriented remote
number of spatial pixels may differ in data volume, sensing research. However, in terms of number of pix-
which can, for example, indicate the use of multimodal els, semantic segmentation takes the lead, indicating
imagery. Object detection offers the largest data vol- that a larger spatial coverage is usually involved for this
ume among the existing benchmarking datasets, which type of application.

0.66% 3.39% 0.13%

0.16% 1.12%

1.92% 14.53%
21.07%

8.52%
22.81%
0.65%

29.94% 13.87%
57.31%

20.19% 0.24%
0.48%

0.34% 0.14%
2.52%
(a) (b)

Superresolution/Pansharpening Object Detection

Object Detection and Classification Change Detection

Change Detection and Classification Classification

Classification and Semantic Segmentation Semantic Segmentation

Object Detection and Semantic Segmentation Others

FIGURE 2. Distribution of EO dataset sizes over typical remote sensing tasks, expressed in (a) volume and (b) size, as defined in Figure S1
and “How to Measure the Size of a Dataset.” In (a), with 30%, object detection is the predominant task, followed by semantic segmentation.
In (b), in contrast to Figure 2(a), semantic segmentation is the prevailing task, illustrating that corresponding datasets involve more complex
scenarios such as leveraging multiple sensors or spectral bands. Object detection and semantic segmentation are the dominant image
analysis tasks in ML-centered EO.

Platforms Sensor Type Time Epochs

Satellite RGB Single

60.19% 54.63% 79.01%

4.63% HS
25% 7.41%
SAR
17.9% 20.99%
6.48% 15.43%
8.33%
Aerial Multiple Multiple Multiple
MS
Drone

FIGURE 3. A distribution of available EO datasets over different platforms, sensor types, and number of acquisition times. Single-image red,
green, blue (RGB) images acquired by satellites are clearly the dominating modality. MS: multispectral; HS: hyperspectral.

70 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


PLATFORMS the platform type, which then splits into different sensor
A direct overview of the occurrence of a type of platform types in the middle ring, and finally denotes the targeted
and sensor is given in Figure 3. Satellite platforms are tasks in the outer ring, respectively. This graph shows that
the most common, followed by airborne platforms and the datasets acquired by unmanned aerial vehicles (UAVs)
drones. Optical data cover more than half of the datas- and aircraft are mainly dedicated to optical sensors, while
ets, while all other sensors are almost equally distributed. satellite-based EO has a much wider and more homoge-
Interestingly, 20% of the datasets provide time series, neous distribution across all sensor and application types.
while the rest are single-temporal acquisitions. Comple- Figure 5(a) and (b) further specifies the previous find-
menting this, Figure 4 highlights the distribution of tasks ings by showing how the datasets acquired by different
between sensors and platforms. The inner ring indicates platforms are distributed across both tasks and sensors.

Class/SemSeg
SemSeg
SemSeg

eg
Others

SemS
Class

OD
OD

ss
OD hers

ers

/Cla
OD
/Cla

Oth
Ot
Se

OD
ss
mS
OD
Clalass

eg
/C
ss
Ot

OD
he
rs
Se
m
Se

Multiple
g

Cl
ss
MS

OD as
HS

SAR

/C O s la
C
Se Oth las D
RGB

m er s
Se s
CD g

B
/C CD CD

RG
las
M

s
ult

eg
ipl

mS
e

Se
Cla g
ss RG ers Se
B OthD/Sem
OD
Cla OD
ss/S C
emS Seg
eg
le Sem
OD ltip
Mu SR
Dron

OD/S OD
emS
eg
e
M
ul

MS
tip

Others
l
le

MS ria SemSeg
Ae
SR

HS
Others
OD
Class
CD
SemSeg
SemSeg

Others

CD SAR

s OD
Clas
llite

le
Sate

ltip
Mu

Seg
em
ss/S Cla
ss
Cla
ers
Oth

Se
eg mS
mS eg
Se
RGB
CD

SR
Ot
he
rs
ss
Cla

OD
Seg
Class/Sem

FIGURE 4. A distribution of tasks between sensors and platforms. Platforms are in the inner ring, sensors are distributed in the middle ring,
and the outer ring shows different tasks per sensor. OD: object detection; CD: change detection; SemSeg: semantic segmentation; SR: super-
resolution; class: classification; MS: multispectral; HS: hyperspectral.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 71


Acquisition Platforms Mulitiple Satellite Aircraft Drone
SAR
RGB
Sensor
Multiple
MS
HS

CD CD/Class Class Class/SemSeg OD OD/Class OD/SemSeg Others SR SemSeg


Tasks
(a)
Until 2018 2018
MS Multiple RGB SAR

MS Multiple RGB SAR


Sensor
HS

HS

CD CD/Class Class Class/SemSeg OD OD/ClassOD/SemSeg Others SR SemSeg CD CD/Class Class Class/SemSeg OD OD/ClassOD/SemSeg Others SR SemSeg

2019 2020
MS Multiple RGB SAR

MS Multiple RGB SAR


Sensor
HS

HS

CD CD/Class Class Class/SemSeg OD OD/ClassOD/SemSeg Others SR SemSeg CD CD/Class Class Class/SemSeg OD OD/ClassOD/SemSeg Others SR SemSeg

2021 2022
MS Multiple RGB SAR

MS Multiple RGB SAR


Sensor
HS

HS

CD CD/Class Class Class/SemSeg OD OD/ClassOD/SemSeg Others SR SemSeg CD CD/Class Class Class/SemSeg OD OD/ClassOD/SemSeg Others SR SemSeg

Tasks Tasks
(b)

FIGURE 5. Different tasks [CD, classification (class), semantic segmentation (SemSeg), object detection (OD), superresolution (SR), and
pansharpening] and task combinations (denoted by “/”) make use of very different platforms and sensors. (a) Combinations of different
platforms, sensors, and tasks accumulated over the years. (b) Combinations of different platforms, sensors, and tasks for different years.
MS: multispectral; HS: hyperspectral.

72 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Although the former provides an overview, the latter GEOGRAPHIC DIVERSITY
adds a temporal aspect by showing the corresponding The geographic diversity and spatial distribution of EO
applications per year. The x- and y-axes represent tasks imagery is an important attribute of a benchmark data-
and sensors, respectively, while the different markers set as it is directly related to the geographic generalization
indicate the type of acquisition platforms. From these performance of data-driven models. Figure 6 shows the
plots, we can affirm that object detection and classifica- geographic distribution of
tion tasks are mainly performed on optical images. At the datasets (i.e., of roughly ~300
same time, semantic segmentation is fairly evenly dis- datasets (75%) from Table 2 OBJECT DETECTION OFFERS
tributed among optical, multispectral, and other sensor that provided (or where we THE LARGEST DATA VOLUME
combinations. SAR images are mainly acquired from sat- could find) their geographic AMONG THE EXISTING
ellite platforms, while hyperspectral datasets are almost information). Many of the
BENCHMARKING DATASETS,
always acquired from airborne systems. UAVs mainly datasets are globally distrib-
WHICH AGAIN CONFIRMS
carry optical sensors in the context of semantic segmen- uted (“Global”) and contain
tation. Some tasks, such as superresolution, naturally images from around the ITS POPULARITY IN
make use of multimodal data, e.g., optical and hyper- world, while others cover a DL-ORIENTED REMOTE
spectral imagery. The year-by-year graph in Figure 5(b) limited set of multiple cities SENSING RESEARCH.
shows that superresolution datasets, together with UAV- or countries (“Multiple Loca-
based acquisitions, have received more attention in re- tions”). Maybe surprisingly,
cent years. On the other hand, the EO community has “Synthetic” datasets show a dominant presence as well,
not seen many new hyperspectral datasets since 2018. illustrating the benefits of being able to control image
Optical sensors were the most common source of infor- acquisition parameters (such as viewing angle), environ-
mation, while after 2020 an increasing number of datas- mental factors (such as atmospheric conditions, illumi-
ets were also acquired from other sensors, such as SAR or nation, and cloud cover), and scene content (e.g., type,
hyperspectral systems. size, shape, and number of objects). Figure 6 illustrates
Figures 3, 4, and 5(a) show that the number of datasets an important, and seldom-discussed issue within the EO
acquired from “Multiple” platforms or sensors is still the mi- community: there exists a strong geographic bias within
nority, which provides evidence for the earlier statement that the available EO datasets. Although 25% of the datas-
state-of-the-art datasets are usually designed to respond to a ets contain samples from globally distributed locations,
specific task in EO applications. These figures also show which nearly 40% of available datasets cover regions in Europe
combination of EO tasks, platforms, and sensors is currently (21%) and North America (18%) only. Asia is still covered
underrepresented. In particular, Figure 5(a) shows three main by 10%, however, Africa (5%), South America (4%), and
gaps: 1) SAR change detection (CD), 2) SAR superresolution, Australia (1%) are barely included. This raises the ques-
and 3) hyperspectral superresolution. From a sensor perspec- tion of whether many of the findings and conclusions in
tive alone, the lack of airborne SAR datasets and drone-based corresponding research articles would generalize to these
hyperspectral benchmarks are other obvious gaps. geographic areas. In any case, the need for more spatially

80
Geographic Diversity of Datasets
70
South America
4%
60
North
50 America
18% Multiple
Other Global Locations
40 Europe 41% 12%
25%
21%
30 Asia
Synthetic
10%
20 Australia 4%
1% Africa
10 5%

0
Brazil

Australia
Sweden
Global
United States
Multiple Locations
China
Synthetic
Germany
France
Europe
Italy
Canada
Asia
Africa
Rwanda
Greenland
Switzerland
India
Peru
South Africa
South America
Spain
Netherlands
Japan
Greece
Alaska
Belgium
Gulf of Mexico
United Kingdom
UAE

Sudan
Austria
Benin
Slovenia
Scotland
Denmark
Poland
Ecuador
Norway
North America
New Zealand
Estonia
Finland
Mexico
Mali
Malawi
Kenya
Ghana
Hungary
West Africa

FIGURE 6. A geographic distribution of EO benchmark datasets (which provided clear location information). UAE: United Arab Emirates.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 73


diverse datasets becomes apparent, in particular, covering (Table 2; 211) and Sen12MS-CR-TS (Table 2; 98), or a mix-
underdeveloped countries as well. ture of these [e.g., RSOC (Table 2; 117)] uses existing datas-
ets (DOTA, Table 2; 109) as well as manual annotations for
SOURCE OF THE REFERENCE DATA new images.
Another important aspect is the source of the reference data. The quality of class definitions varies among different
Although most of the scientific papers that introduce a new datasets, often depending on whether a dataset is designed
benchmark dataset are very detailed regarding source and for a specific real-world application (in which class defini-
properties of the provided image data, they often contain tions are driven by the application context), or whether it
only sparse information about the exact process of how the is just an example dataset used to train and evaluate a ML-
reference data were obtained. However, knowledge about based approach (in which case, class definitions are more
the source of the reference data as well as about measures arbitrary). If human interaction is involved (e.g., via manu-
for quality assessment and quality control are essential to al annotations), annotation protocols or precise class defi-
judge the potential of the benchmark and how to interpret nitions are often not shared (if they even exist). With the
obtained results. For example, achieving high “accuracy” evolution of datasets, the quality of the meta-information
on a dataset that contains a high number of annotation er- about the reference data needs to increase, together with
rors in its reference data only means that the model is repro- the quality and quantity of the image data.
ducing the same bias as the method used to annotate the
data but is not actually producing accurate results. Further- APPLICATIONS AND TASKS
more, information about the source of the reference data is Finally, we analyze a last aspect concerning the distribu-
not only scarce but also very heterogeneous. Examples in- tion of some characteristics of the datasets, namely, the
clude manual annotation for semantic segmentation [e.g., number of classes and images. In Figure 7(a), we show the
OpenEarthMap (Table 2; 253)] or object detection [e.g., evolution of the number of classes per publication year.
BIRDSAI (Table 2; 127)], DroneCrowd (Table 2; 138), the In Figure 7(b), we plot the number of classes against the
use of keywords in image search engines for classification number of images in the dataset. In both plots, the radius
[e.g., AIDER (Table 2; 34)], leveraging existing resources indicates the size of the image (width or height), while the
[e.g., BigEarthNet (Table 2; 40)] uses the CORINE land use color indicates the type of task. We can see that there is no
land cover (LULC) map, MapInWild (Table 2; 264) uses clear trend or correlation between the year of publication and
the World Database of Protected Areas, AI-TOD (Table 2; the number of classes in a dataset. Instead, we find that more
162) uses other existing datasets, ship detection datasets recently published datasets increase the variety in the num-
usually employ AIS data, OpenSentinelMap (Table 2; 282) ber of classes, again highlighting an increased interest in us-
uses OSM, automatic reference methods (e.g., HSI Oil Spill ing benchmarking datasets in a wider range of applications.

108 CD 108
97 97
Superresolution/Pansharpening
87 87
78 Object Detection 78
69 Classification 69
61 Semantic Segmentation 61
54 Change Detection and Classification 54
48 48
Object Detection and Classification
42 42
Number of Classes

Object Detection and Semantic Segmentation


36 36
31 Classification and Semantic Segmentation 31
27 Others 27
23 23
20 20
16 16
14 14
11 11
9 9
7 7
5 5
4 4
2 2
1 1
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022

1
6
28
109
357
1,000
2,000
6,000
14,000
29,000
59,000
112,000
206,000
365,000
627,000
1,048,000
1,708,000
2,719,000
4,242,000
6,492,000

Publication Year Number of Images


(a) (b)

FIGURE 7. The number of classes provided by the reference data of a given dataset not only varies for different tasks (e.g., object detection
is dominated by datasets with only a single class) but also with (a) publication year and (b) the number of images of a dataset.

74 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Furthermore, although there is no clear correlation, we can instance, Functional Map of the World (FMoW) (Table 2;
confirm that increasing the number of classes reduces the 76), introduced in 2018 [14], is the largest dataset for remote
size of the images and increases the number of images. More sensing scene classification in terms of the number of images
populated datasets, in terms of number of images, also have (1 million) but has yet to gain
smaller image sizes and typically consider a smaller number a high level of popularity [with
of classes. Conversely, larger images are found in less popu- 200 citations and 128 GitHub MORE POPULATED
lated datasets. stars, compared to EuroSAT DATASETS, IN TERMS OF
This overview is a snapshot of the currently published (Table 2; 52) with 796 cita-
NUMBER OF IMAGES, ALSO
benchmarking datasets. Although the list of datasets con- tions and 276 GitHub stars, or
HAVE SMALLER IMAGE
tinues to grow, we believe that the observed trends will not AID (Table 2; 69) with 1,000
change much over the next few years. Instead, we expect citations, which were all pub- SIZES AND TYPICALLY
long-term developments that will lead to two divergent lished in 2018]. Several other CONSIDER A SMALLER
aspects of datasets: specificity and generality. This will be factors affect the popularity NUMBER OF CLASSES.
further discussed in the “Open Challenges and Future Direc- of a dataset, too, such as ease
tions” section. of access to the hosting servers
[Google Drive, IEEE DataPort, Amazon Web Services (AWS),
EXEMPLARY DATASETS university/personal servers, and so on], accompanying doc-
This section provides short descriptions of a selection of EO umentation and presentation (standard metadata format,
datasets for various tasks: semantic segmentation, classifica- detailed description, availability of template code and sup-
tion, object detection, CD, and superresolution/pansharpen- port, suitable search engine optimization, and so on), or
ing. Given the vast number of publicly available EO datasets, ease of downloading the data (temporary or permanent links,
it is only possible to present some of them in this article. bandwidth of hosting server, sign-in/-up requirements, and
Thus, this selection cannot be comprehensive and certainly so forth). Finally, an already-established dataset is more likely
follows a subjective view influenced by the experience of the to be used in new studies to enable a comparison to related
authors. However, the selected datasets are representative for prior works, even if newer (and potentially better) datasets
their respective tasks and were selected based on their ob- might exist. Along with brief descriptions, this section pro-
served relevance: they either are the largest in terms of size vides insights into the different dataset characteristics.
(see the “Evolution of EO-Oriented ML Datasets” section for
a definition), the first to be introduced for a given task, or the SEMANTIC SEGMENTATION
most popular dataset for the chosen application. The popu- Semantic segmentation refers to assigning a class label to each
larity was determined based on the number of citations the pixel in the image. Partitioning the image into semanti-
original paper introducing the dataset received. cally meaningful parts creates a more detailed map of the
The popularity of a dataset is influenced by multiple fac- input image and provides a foundation for further analy-
tors. One is certainly the size of a dataset, i.e., larger datas- sis such as LULC classification, building footprint extrac-
ets are often preferred; however, there are exceptions. For tion, landslide mapping, and so on [15]. Figure 8 shows the

Volume (MB)
ActiveFireL8 South Sudan Crop Type

PASTIS HKH Glacier Mapping


CITY-OSM
FireCube SpaceNet-1
RescueNet
MUESLI
SpaceNet-7
MultiSenGE
DFC18
DFC22-SSL
DFC20

Other DFC21 - DSE

AIRS SEN1-2

OpenSentinelMap WildfireDanger Aerial


Kag. DLRSD
Indian
Salinas
DFC07
Image
Kennedy
Botswana
Pavia
GTA Pines
DubaiSemSeg
University
Segmentation
Space
-VCenter
Zurich
DFC08
Satellite
SpaceNet
UBC SID
Summer
AeroscapesWater
-6
MultimodalBen Center
Bodies
Kag. Satellite
DFC13
SynthAer
WHU
Synthinel -Hi Buildings
S12_ScotInd_Cloud
-1
SmallholderPlant
ManipalUAVid
MARIDA
LandCoverAI
ISPRS 2D - Vaihingen Chikusei
STGAN
RIT-18 Cloud Removal
OnCloudN AeroRIT
BHPools&WaterTanks
NDWS
HOSD
Landslide4Sense
SpaceNet-5 ISPRS 2D - Potsdam DeepGlobe-LandCover
S2-SHIPS
Kag. Massachusetts
Semantic Drone Buildings
FloatingObjects TTPLA
SemCity Toulouse
MiniFrance ETCI2021
Sat_Burn_Area CrowdAI
UAVid Mapping
SEN-12-FLOOD iSAID
SpaceNet-8
MapInWild Chesapeake Land Cover FloodNet
C2S LoveDA
GreeceFire
-MS MMFlood
GrowliFlower
SpaceNet-2
Ghana Crop Type SpaceNet-3
SeasoNet

FIGURE 8. Relative volume distribution among datasets addressing semantic segmentation.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 75


relative volume of corresponding benchmark datasets and at random locations around the globe and cover the four
illustrates a wide spread of dataset sizes. Here, we present meteorological seasons of the Northern Hemisphere.
two examples: SEN12MS (focusing on medium-resolution The patches were further processed to ensure that the
satellite data and weak annotations) and the ISPRS Vaihin- images did not contain any clouds, shadows, and arti-
gen dataset (focusing on very high-resolution airborne data facts. In addition to the images and the land cover maps,
with high-quality annotations). the results of two baseline convolutional neural net-
1) SEN12MS [49] is among the most popular and larg- work classifiers (ResNet50 and DenseNet121) are also
est datasets (Table 2; 312) in terms of volume of data discussed to demonstrate the usefulness of the dataset
as shown in Figure 1. A total of 541,986 image patches for land cover applications [17].
of size 256 × 256 pixels with high spectral information 2) ISPRS Vaihingen [18] is the earliest semantic segmenta-
content is present in this dataset [16]. It contains dual- tion dataset used for identifying land cover classes in
polarimetric SAR image patches from Sentinel-1, multi- aerial images [18], as shown in Figure 1 (Table 2; 220).
spectral image patches from Sentinel-2, and Moderate- This dataset contains a subset of images acquired by
Resolution Imaging Spectroradiometer (MODIS) land aerial cameras during the test of digital aerial cameras
cover maps (Figure 9). The image patches are acquired carried out by the German Association of Photogramme-
try and Remote Sensing [19]. The dataset was prepared
as part of a 2D semantic labeling contest organized by
the International Society for Photogrammetry and Re-
mote Sensing (ISPRS). The dataset contains images ac-
quired over the city of Vaihingen in Germany. In total,
orthorectified images of varying sizes and a digital sur-
face model (DSM) are provided for 33 patches covering
Vaihingen (Figure 10). The ground-sampling distance
(a) for the images and DSM is 9 cm. The three channels
of the orthorectified images contain infrared, red, and
green bands as acquired by the camera. The images are
manually labeled for six common land cover classes:
impervious surfaces, buildings, low vegetation, tree, car,
and clutter/background.

(b)

(c)

(d)

FIGURE 9. Several patch triplet examples from the SEM12MS


dataset. (a) False-color Sentinel-1 SAR (R: VV; G: VH; B: VV/VH), (b)
Sentinel-2 red, green, blue, (c) Sentinel-2 short-wave infrared, (d) FIGURE 10. An illustration of 33 patches from the ISPRS Vaihingen
IGBP land cover, and LCCS land cover. dataset. (Source: [18].)

76 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


SCENE CLASSIFICATION ◗   BigEarthNet-MM [51] (Table 2; 40) is a benchmark ar-
Remote sensing scene classification is closely related to se- chive that introduced an alternative nomenclature for
mantic segmentation and has been used in various appli- images as compared to the traditional CORINE Land
cation domains such as urban planning [20], environment Cover (CLC) map for multilabel image classification and
monitoring [21], LULC estimation [22], and so forth. The image-retrieval tasks. The CLC level-3 nomenclature is
main difference is that instead of pixelwise classification arranged into 19 classes for semantically coherent and
and resolution-preserving maps as output, in scene classi- homogenous land cover classes. This archive is created
fication, only one or more global labels are predicted for a using Sentinel-1 SAR and Sentinel-2 multispectral satel-
given input image, aiming at a more comprehensive, gen- lite images acquired between June 2017 and May 2018
eralized, and context-aware understanding of the underly- over 10 European countries [27]. The first version of Big-
ing scene. Similar to image classification, which has been the EarthNet included only Sentinel-2 image patches. It has
driving force behind the early days of DL in computer vision, been augmented with Sentinel-1 image patches to form a
research on remote sensing scene classification has led to multimodal benchmark archive (also called BigEarthNet-
the creation of more diverse and large-scale high-resolution MM). The archive comprises 590,326 image patches with
image datasets. Figure 11 shows the relative volume distri- 12 spectral and two polarimetric bands. As shown in Fig-
bution of remote sensing scene classification datasets. This ure 13, each image patch is annotated with several land
section covers EuroSAT as one of the earliest and FMoW as cover classes (multilabels). One of the key features of the
one of the largest datasets (representing roughly 50% of the BigEarthNet dataset is its large number of classes and
available data in this task category) as well as BigEarthNet- images, which makes it suitable for training DL models.
MM, which introduces an additional interesting aspect by However, because it is a multilabel dataset, the complex-
providing multiple labels for each image. ity of the problem is significantly increased compared to
◗ EuroSAT [50] is one of the earliest large-scale datasets single-label datasets [28].
tailored for training deep neural networks for LULC clas- ◗◗ FMoW [52] (Table 2; 76) is the largest dataset (Figure 1)
sification tasks, as depicted in Figure 1 (Table 2; 52). The for remote sensing scene classification in terms of the
dataset includes 10 land cover classes (Figure 12), each number of images [14]. The FMoW dataset is composed
containing 2,000–3,000 images, for 27,000 annotated of approximately 1 million satellite images, along with
and georeferenced images with 13 different spectral metadata and multiple temporal views as well as sub-
bands of 64 # 64 pixels. It contains multispectral images sets of the dataset that are classified into 62 classes,
with a single label acquired from Sentinel-2 satellite imag- which are used for training, evaluation, and testing.
es of cities in 34 European countries [23], [24]. This data- Each image in the dataset comes with one or multiple
set has been widely used for classification tasks, however, bounding boxes indicating the regions of interest. Some
it may be used in a variety of real-world EO applications, of these regions may be present in multiple images ac-
such as detecting changes in land use and land cover as quired at different times, adding a temporal dimension
well as improving geographical maps as Sentinel-2 im- to the dataset. The FMoW dataset is available to the pub-
ages are freely available [23], [25], [26]. lic in two image formats: the FMoW-Full and the FMoW

Volume (MB)
FMoW

So2Sat LCZ42

DND-SB
Rwanda Crop Type

Gaofen Image (GID)

BrazilDAM
Other
ZueriCrop
SAR OVERHEAD
Kag. Satellite
-based
NWPUBCS
BCSS
Hurricane
RSSCN7
WHU-RS19
Sea
UC MNIST
IceDamage
-Merced
Bijie Dataset
SanFranBay_Ships_Planet
Landslide
SIRI-WHU
USTC_SmokeRS
RSI-
SAT-4
SAT-6
Airbus Wind PatternNet
Turbines
MLRSNet Patches
BigEarthNet-MM EuroSAT
RSI-CB256
MASATI-v2
AID
LandCoverNet Au Kag. Find a Car Park
SeaIceClassification
Kag. Planet Forest CV4A Kenya
DeepGlobe -Road
VEDAI
SloveniaLandCover ADVANCE
Kaggle Cloud Detection
MultiScene
LandCoverNet Af AiRound CV-BrCT
LandCoverNet SA S2-Agri
xView ICONES-HSI
ARGO
AnthroProtect
LandCoverNet Eu

LandCoverNet As DENETHOR
LandCoverNet NA

FIGURE 11. Relative volume distribution among datasets addressing scene classification.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 77


AnnualCrop Forest HerbaceousVegetation Highway Industrial Pasture PermanentCrop Residential River SeaLake

FIGURE 12. Sample image patches of all 10 classes covered in the EuroSAT dataset.

red, green, blue (RGB). The FMoW-Full is in TIFF format 200 GB and includes all multispectral data converted to
and contains four- and eight-band multispectral imag- RGB in JPEG format. Examples of the classes in the da-
ery with a high spatial resolution resulting in 3.5 TB of taset include flooded roads, military facilities, airstrips,
data, while the FMoW-RGB has a much smaller size of oil and gas facilities, surface mines, tunnel openings,
shipyards, ponds, and towers (see
Figure 14 for examples). The FMoW
dataset has a number of important
characteristics, such as global di-
versity, multiple images per scene
captured at different times, multi-
spec t ral imager y, and metadata
linked to each image.

Urban Fabric, Arable Land, Coniferous Forest, Mixed Forest, Peatbogs


Transitional Woodland, Shrub OBJECT DETECTION
The aim of object detection is to lo-
cate and identify the presence of
one or more objects within an im-
age, including objects with clear
boundaries, such as vehicles, ships,
and buildings, as well as those with
more complex or irregular boundar-
Permanently Irrigated Permanently Irrigated Discontinuous Urban Sea and Ocean
Land, Estuaries ... Land ... Fabric ... ies, for example, LULC parcels [29].
As seen in Figure 2, object detection
is one of the most widely studied
research tasks. Figure 15 shows the
relative volume of the corresponding
datasets. The xView3-SAR is by far
the largest one. On the other hand,
Pastures, Mixed Nonirrigated Nonirrigated Coniferous Forest,
the family of DOTA datasets has a
Forest Arable Land ... Arable Land ... Water Bodies ... pioneering role in this field, placing
them among the most popular data-
FIGURE 13. Sample image patches of several classes in the BigEarthNet-MM dataset with sets for object recognition tasks from
multiple labels being assigned to each image. remote sensing images.

78 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


◗   The xView3-SAR [53] Ship Detection dataset is the largest SAR images from the Sentinel-1 mission, which are an-
labeled dataset (Table 2; 191), as shown in Figure 1, for notated using a combination of automated and manual
training ML models to detect and classify ships in SAR analysis (see Figure 16 for image samples). The images
images. The dataset includes nearly 1,000 analysis-ready are accompanied by co-located bathymetry and wind

Airport Airport Hangar Airport Terminal Amusement Park Aquaculture Archaeological Site Barn

Border Checkpoint Burial Site Car Dealership Construction Site Crop Field Dam Debris or Rubble

Educational Factory
Electric Substation Fire Station Flooded Road Fountain Gas Station
Institution or Power Plant

Ground Transportation Impoverished


Golf Course Helipad Hospital Interchange Lake or Pond
Station Settlement

Parking lot
Lighthouse Military Facility Nuclear Power Plant Office Building Oil or Gas Facility Park
or Garage

Recreational
Place of Worship Police Station Port Prison Racetrack Railway Bridge
Facility

Residential Unit Residential Unit Road Bridge Runway Shipyard Shopping Mall Smokestack

Solar Farm Space Facility Stadium Storage Tank Surface Mine Swimming Pool Tollbooth

Tower Tunnel Opening Waste Disposal Water Treatment Wind Farm Zoo
Facility

FIGURE 14. Sample image patches for several classes from the FMoW dataset.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 79


Volume (MB)
xView3-SAR

WHUS2-CD+
WHU Building Kag. Airbus Ship Detection

RampBuilding
DOTA v2.0

TopoBoundary

INRIA
BIRDSAI
Other

DOTA v1.5

AirbusTree
Urban OilAMD
SZTAKI
SAR_VD
NWPU
Storage
EoCCrownAirChange
-VHR10
NASAdebris
AIR-SARShip
ALCD
Iceberg
RSOD
WAMI Detection
-2
Detection
DIRSIG
PlanesNet
SAR Ship
MTARSI
HRSID
SRSDD Detection
-v1.0
SaRNet
UAVOD-10
TaS LEVIR-Ship
COWC DOTA v1.0 SSDD
SIMD
DSSDD
ITCVD
Bridges
built-structure
CARPK -count
VisDrone NYCPlanimetric
AU-AIR
DSTL3B SeabirdsDetection
MSAR
Oil Storage Tanks
ArcticSeals Satellite Pool Detection
SpaceNet-4 SWIM ShipRSImageNet
Agriculture-Vision
DroneCrowd PKLot
AFO
TGRS-HRRSD EU Flood
VISO DSTL16BRoads
Massachusetts
NEON Tree HRSC2016
Dataset
LS-SSDD
RarePlanes

Flying Airplanes

FIGURE 15. Relative volume distribution among the datasets addressing object detection.

state rasters, and the dataset is intended to support the The DOTA dataset is available in three different versions:
development and evaluation of ML approaches for de- DOTA-v1.0, DOTA-v1.5, and DOTA-v2.0. The image size
tecting “dark vessels” not visible to conventional moni- in the initial version ranges from 800 × 800 pixels to
toring systems. 4,000 × 4,000 pixels, with 188,282 object instances with
◗◗ DOTA [54] is one of the most popular and largest ob- various scales and angular orientations and a total of 15
ject detection datasets (Table 2; 109 and 114) in terms object categories. DOTA-v1.5 adds various small objects
of labeled object instances. It includes 2,806 images ac- (fewer than 10 pixels) and a new container crane catego-
quired from Google Earth (GE) and the China Center for ry with 402,089 instances, whereas DOTA-v2.0 adds two
Resources Satellite Data and Application [31], [32], [33]. categories, airport and helipad, with 11,268 images and
1,793,658 instances, respectively.
Some image samples of the DOTA
VH_dB VV_dB Mask dataset are presented in Figure 17.

CD
CD in remote sensing aims to iden-
tify temporal changes by analyzing
multitemporal satellite images of the
same location. CD is a popular task
in EO as it fosters monitoring envi-
ronmental changes through artifi-
(a) cial or natural phenomena. Figure 2
WindDirection WindQuality WindSpeed shows that the number of dedicated
CD datasets is small compared to
other applications. Figure 18 shows
that the available data are dominated
by the DFC20 dataset (Track MSD),
which focuses on semantic CD, fol-
lowed by xView2, which tackles
building damage assessment in the
context of natural disasters. We chose
(b) LEVIR-CD as a recent dataset exam-
ple and the Onera Satellite Change
FIGURE 16. (a) An example image stack of dual-polarimetric SAR images and a water mask and Detection dataset as one of the first
(b) several wind properties from the xView3-SAR dataset. large-scale datasets.

80 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


◗   LEVIR-CD [55] is one of the most recent and largest CD as new buildings or roads [34]. The locations were se-
datasets, as seen in Figure 1 (Table 2; 15). It was mainly lected from around the world, including Brazil, the
developed for the evaluation of DL algorithms on build- United States, Europe, the Middle East, and Asia. The
ing-related changes, including building growth (the reference data for pixel-level change are provided for
transition from soil/grass/hard-
ened ground, building under con-
struction to new build-up regions) Harbor Large-Vehicle Swimming-pool Baseball-Diamond
and building decline [33]. The
dataset comprises 637 optical im-
age patch pairings extracted from
GE with a resolution of 1,024 #
1,024 pixels acquired over a time
span of five to 14 years. LEVIR-
CD covers various structures, in- Airport Ground-Track-Field Ship Storage-Tank
cluding villa residences, tall apart-
ments, small garages, and large
warehouses, from 20 distinct re-
gions in multiple Texas cities. The
fully annotated LEVIR-CD da-
taset comprises 31,333 distinct
Tennis-Court Small-Vehicle Basketball-Court Bridge
change-building instances, some
of which are illustrated in Fig-
ure 19, generated from the bitem-
poral images by remote sensing
image interpretation specialists.
◗◗ Onera Satellite Change Detec-
tion [56] is one of the first, larger Helipad Plane Roundabout Soccer-Ball-Field
CD datasets, as shown in Fig-
ure 1 (Table 2; 11), containing
multispectral image pairs from
Sentinel-2. This dataset includes
24 pairs of multispectral images
acquired from Sentinel-2 satel-
lites between 2015 and 2018,
focusing on urban changes such FIGURE 17. Several example image samples from the DOTA dataset.

Volume (MB)

DFC21-MSD
Synthetic and Real

LEVIR-CD

Other Second
HRSCD DFC09
SYSU-CD DSIFN
S2Looking
MtS-WH
S2MTcityPair
OneraCD
MUNO21
AIST Building CD
AICD
xView2

FIGURE 18. Relative volume distribution among the datasets addressing change detection.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 81


all 14 training and 10 test image pairs. As illustrated SUPERRESOLUTION/PANSHARPENING
in Figure 20, the annotated changes are primarily as- Pansharpening is one of the oldest data fusion approach-
sociated with urban changes, such as new buildings es in remote sensing and aims to increase the spatial
or roads. resolution of a multispectral image by combining it with

FIGURE 19. Examples of annotated samples from the LEVIR-CD dataset.

Chongqing Dubai Lasvegas Brasilia Valencia Milano

FIGURE 20. Examples of annotated samples from the Onera Satellite Change Detection dataset.

82 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


a panchromatic image. Due to the resolution difference 2) PAirMax [58] is a recently published (Table 2; 319 and
between panchromatic and multispectral sensors, pan- 324) yet popular dataset for pansharpening. It contains
sharpening is an exciting topic in remote sensing as it 14 multispectral and panchromatic image pairs from
can provide a means to obtain higher-resolution data six sensors on board satellite constellations of Maxar
without better sensor technology. We select the Proba-V, Technologies and Airbus
PAirMax, and WorldStrat datasets as examples to show- [36]. As seen in Figure 22,
case the peculiarities of datasets designed for that par- most of the images are ac- CD IS A POPULAR TASK IN
ticular application. quired over urban areas, EO AS IT FOSTERS
1) Proba-V [57] is the earliest dataset available for su- highlighting several chal-
MONITORING
perresolution [35] as Figure 1 shows (Table 2; 318). lenges for pansharpening
ENVIRONMENTAL CHANGES
This dataset contains radiometrically and geometri- (such as high-contrast fea-
cally corrected Top-Of-Atmosphere reflectances in tures and adjacent regions THROUGH ARTIFICIAL OR
Plate Carre projection from the PROBA-V mission with different spectral NATURAL PHENOMENA.
of the European Space Agency (Figure 21). The RED features). The work [36]
and near-infrared (NIR) spectral band data at 300 m also discusses the best
and 100 m resolution are collected for 74 selected practices to be followed in preparing high-quality, full-
regions across the globe. Superresolution might be , and reduced-resolution images for pansharpening
affected by the presence of pixels with clouds, cloud applications. In this dataset, the panchromatic band
shadows, and ice/snow cover. Therefore, this dataset has 4 × 4-times-more pixels than the multispectral
contains a quality map indicating pixels affected by bands. The multispectral bands are near the visible
clouds that should not be considered for superreso- infrared region. The dataset also contains nine re-
lution. The dataset contains one 100 m resolution duced-resolution test cases per Wald’s protocol.
image and several 300 m resolution images of the 3) WorldStrat [59] is a recently introduced dataset for su-
same scene. perresolution [37] and the largest in terms of volume

HR LR000 LR001 LR002 ... LR014


NIR

SM QM000 QM001 QM002 ... QM014


NIR

HR LR000 LR001 LR002 ... LR025


Red

SM QM000 QM001 QM002 ... QM025


Red

FIGURE 21. Sample images from the Proba-V dataset. Each sample consists of one high-resolution (HR) and several low-resolution (LR)
images, each with a quality map (QM) showing which pixels are concealed (e.g., through clouds and so on).

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 83


FIGURE 22. Sample images from PAirMax dataset. (Source: [58].)

(107 GB), as Figure 1 shows (Table 2; 325). This dataset WORKING WITH REMOTE SENSING DATASETS
contains high-resolution images from Airbus SPOT 6 This section provides guidance on how to leverage avail-
and 7 along with 16 temporally matched low-resolu- able datasets to their full potential. Two of the main
tion images from Sentinel-2 satellites. The high-resolu- difficulties caused by information asymmetry (i.e., the in-
tion images are over five spectral bands: the panchro- formation imbalance between the data providers and the
matic band at a 1.5-m pixel resolution and RGB, and data users) [5] have found suitable datasets and easy pro-
NIR bands at 6 m per pixel. The low-resolution ranges totyping of ML approaches using such datasets. Here we
from 10 m per pixel to 60 m per pixel (Figure 23). discuss resources to gain an overview of existing datasets
In total, the dataset covers an area of approximately and download actual data, but we also provide examples
10,000 km 2 and attempts to represent all types of land of EO-oriented ML programming libraries.
use across the world. Notably, the dataset contains
nonsettlement and underrepresented locations such DATA AVAILABILITY
as illegal mining sites, settlements of persons at risk, Data availability covers two distinct aspects: on the one
and so on. hand, access to the curated benchmark datasets, i.e., how
such datasets are made available to
the public. This section prov ides
several e xamples of the most
common data sources. On the other
hand, the actual noncurated measure-
ments as acquired by the different
sensors such as satellites, planes, and
UAVs are often available too. Many
data providers offer either their com-
plete database for public access (e.g.,
the European Copernicus Open Ac-
cess Hub [60]) or at least portions
of their image acquisitions (e.g., via
Open Data Programs such as those
from Maxar [61] and Capella Space
FIGURE 23. Sample images from the WorldStrat dataset. (Source: [37].) [62] or through scientific proposals

84 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


as possible for TerraSAR-X/TanDEM-X data [63]). In prin- atmospheric data, and land cover data, and offers
ciple, these data sources are highly valuable as they offer a number of programming APIs. For example, the
free access to petabytes of EO data, which can either be Application for E xtracting and E xploring A nal-
used to compile benchmark datasets by augmenting them ysis Ready Samples API enables users to perform
with reference data for a specific task, or be leveraged for data access and transformation processes easily
modern training paradigms such as self-supervised learn- and efficiently.
ing. An important aspect that is unfortunately sometimes ◗◗ Maxar Open Data Program [61] provides access to
ignored is the licensing of the corresponding data prod- high-resolution remote sensing imagery collected
ucts: although direct usage (at least for scientific pur- since 2017, amounting to a total area of 1,834,152 km 2.
poses) is usually free, redistribution of these data is often It also features several programming APIs, such as the
prohibited. Nevertheless, such image data find their way Maxar Catalog API and the ARD API. In addition, this
into public benchmark datasets, essentially causing copy- program seeks to assist the humanitarian community
right infringements. In addition to being in direct con- by furnishing essential and useful information to sup-
flict with scientific integrity, such behavior is less likely port response efforts when crises occur.
to be tolerated in the future given the rising commercial ◗◗ OpenAerialMap [68] is a community-driven platform
value of EO imagery. Creators of future benchmark data- that provides access to openly licensed satellite and
sets should be fully aware of the license under which the UAV imagery from across the globe, accompanied by
leveraged EO data are accessible and how it limits their programming APIs for data access and processing. The
use in public datasets. In parallel, data providers should platform currently hosts 12,614 images captured by 721
consider a licensing scheme that allows noncommercial sensors, and all the images can be traced in OpenStreet-
redistribution for scientific/educational purposes to fur- Map [69].
ther facilitate the development and evaluation of modern ◗◗ EarthNets [70] is an interactive data-sharing platform
ML techniques. The following is a listing of the most com- with more than 400 publicly available datasets in the
mon data sources: geoscience and remote sensing field, which covers essen-
◗◗ IEEE DataPort [64] is a valuable and easily accessible tial EO tasks like land use/cover classification, change/
data platform that enables users to store, search, ac- disaster monitoring, scene understanding, climate
cess, and manage data. The platform is designed to ac- change, and weather forecasting [38]. Each benchmark
cept all formats and sizes of datasets (up to 2 TB), and dataset provides detailed meta information like spatial
it provides both downloading capabilities and access resolution and volume. Moreover, it also supports stan-
to datasets in the cloud. Both individuals and institu- dard data loaders and codebases for different remote
tions can indefinitely store and make datasets easily sensing tasks, which enables users to conduct a fair and
accessible to a broad set of researchers, engineers, and consistent evaluation of DL methods on the bench-
industry. In particular, most of the datasets used for mark datasets.
previous IEEE GRSS Data Fusion Contests have been ◗◗ EOD [71] is an interactive online platform for catalog-
published on this platform (see “The Role of the IEEE ing different types of datasets leveraging remote sensing
Geoscience and Remote Sensing Society Data Fusion imagery, which is developed by the IEEE GRSS IADF TC
Contests.”) However, unless the dataset is associated [6]. The key feature of EOD is to build a central c­ atalog
with a competition or being submitted as open access, that provides an exhaustive list of available datasets
an IEEE account is required to access and download it. with their basic information, which can be accessed and
◗◗ Radiant MLHub [65] enables users to access, store, extended by the community and queried in a structured
register, and share high-quality open datasets for
training ML models in EO. It’s designed to encour-
age widespread collaboration and the development
of trustworthy applications. The available datasets in
this platform cover research topics like crop type clas-
sification, flood detection, tropical storm estimation,
and so on.
◗◗ Euro Data Cube (EDC) [66] provides a global archive of
analysis-ready data (ARD) (Sentinel, Landsat, MODIS,
and so forth), where users can search for and order data
using the EDC Browser graphical interface (see Figure 24).
It also enables users to store, analyze, and distribute
user-contributed data content with simple application
programming interfaces (APIs).
◗◗ NASA Earthdata [67] provides access to a wide range FIGURE 24. An illustration of the EDC Browser, where users can
of remote sensing datasets, including satellite imagery, easily search for and order satellite data. (Source: [87].)

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 85


The Role of the IEEE Geoscience and Remote Sensing Society Data Fusion Contests
The IEEE Geoscience and Remote Sensing Society (GRSS) Data Fusion Contest nighttime images (Visible Infrared Imaging Radiometer Suite [S13]), and
(DFC) has been organized as an annual challenge since 2006 by the IEEE GRSS OpenStreetMap [S11]. Another meaningful change happened in 2017 when
Image Analysis and Data Fusion (IADF) Technical Committee (TC). The GRSS is the DFC moved away from providing data over a single region only but
an IEEE Society whose stated goal is to bring together researchers and practi- instead allowed the training of models over five cities worldwide (Berlin,
tioners to monitor and understand Earth’s ecosystems and identify potential Hong Kong, Paris, Rome, and Sao Paulo) and testing on four other cities
risks. The IADF is one of seven GRSS TCs aiming at technical contributions with- (Amsterdam, Chicago, Madrid, and Xi’an). This enabled the creation of models
in the scope of geospatial image analysis [e.g., machine learning (ML), deep that generalize to new and unseen geographic areas instead of overfitting to a
learning, image and signal processing, and big data) and data fusion (e.g., mul- single location. This commendable trend was continued in 2020 by using
tisensor, multiscale, and multitemporal data integration). In general, the con- SEN12MS (Table 2; 87) [S20] as training data [S16], providing data for the
test promotes the development of methods for extracting geospatial informa- whole state of Maryland, USA, in 2021 (Table 2; 1 and 291) [S12], in total, 19
tion from large-scale, multisensor, multimodal, and multitemporal data. It different conurbations in France in 2022 (Table 2; 241) [S8], and data from 17
focuses on challenging problem settings and establishes novel benchmark cities across six continents in 2023 [S9].
datasets for open problems in remote sensing image analysis (see Table S1). The tasks addressed by the DFC over the years have been dominated by
Historically, the DFC developed from less formal activities related to the semantic mapping [S3], [S4], [S5], [S6], [S7], [S9], [S10], [S11], [S13], [S17], [S18],
distribution of datasets between 2002 and 2006 by means of a collaboration [S19] but also include pansharpening (Table 2; 320) [S21], change detection (in
between the GRSS and the International Association of Pattern Recognition. In the context of floods) (Table 2; 10) [S1] and 3D reconstruction (Table 2; 193)
2006, the first DFC was organized by what was then called the Data Fusion TC [S3], [S9], modern challenges such as weakly (Table 2; 1/287/291) [S12], [S16]
of the GRSS. It addressed the fusion of multispectral and panchromatic imag- and semisupervised learning (Table 2; 241) [S8] as well as open task contests
es, i.e., pansharpening, and provided one of the first public benchmark datas- [S2], [S4], [S14], [S15], which allowed participants to freely explore the poten-
ets for ML in remote sensing. tial of new and uncommon EO data.
Since 2006, various sensors have played a role in the DFC, including opti- In 2006, seven teams from four different countries participated in the
cal (SPOT [S1], DEIMOS-2 [S2], WorldView-3 [S3], aerial [S4], [S5], [S6], [S7], first DFC [S21] despite public contests being a new concept within the EO
[S8], [S9]), multispectral (Landsat [S10], [S11], [S12], [S13], Worldview-2 [S14], community. Being organized by an internationally well-connected Society,
[S15], Quickbird [S15], Sentinel-2 [S11], [S13], [S16], and aerial [S12]) as well as providing exciting challenges, and establishing new benchmarks led to a
hyperspectral images [S7], [S17], [S18], SAR (ERS [S1], [S10], TerraSAR-X [S15], quick rise in popularity. From seven teams in four countries in 2006, partici-
Sentinel-1 [S13], [S16] Gaofen [S9]) and lidar [S5], [S6], [S7], [S14], [S17], [S18] pation jumped quickly to 21 teams in 2008, 42 teams in 2014, and reached
data, but also less common sources of Earth observation (EO) data such as its peak with 148 teams (distributed over four different tracks, however) in
thermal images [S4], digital elevation models [S8], video from space (ISS [S2]), 2019. The peak of the popularity was around 2012, when the DFC attracted

TABLE S1. AN OVERVIEW OF THE IEEE GRSS DATA FUSION CONTESTS FROM 2006 TO 2023.

YEAR DATA GOAL REFERENCE


2023 Very high-resolution (VHR) optical and SAR satellite images Classification, height estimation [S9]
2022 VHR aerial optical images, DEM Semisupervised learning [S8]
2021 Multitemporal multispectral (aerial and Landsat-8) imagery Weakly supervised learning [S12]
Multispectral (Landsat-8, Sentinel-2), SAR (Sentinel-1), nighttime Semantic segmentation [S13]
(Visible Infrared Imaging Radiometer Suite) images
2020 Multispectral (Sentinel-2) and SAR (Sentinel-1) imagery Weakly supervised learning [S16]
2019 Optical (Worldview-3) images and lidar data Semantic 3D reconstruction [S3], [S19]
2018 Multispectral lidar data, VHR optical and hyperspectral imagery Semantic segmentation [S7]
2017 Multispectral (Landsat, Sentinel-2) images and OpenStreetMap Semantic segmentation [S11]
2016 Very high temporal resolution imagery (DEIMOS-2) and video from space (ISS) Open for creative ideas [S2]
2015 Extremely high-resolution lidar and optical data Semantic segmentation [S5], [S6]
2014 Coarse resolution thermal/hyperspectral data and VHR color imagery Semantic segmentation [S4]
2013 Hyperspectral imagery and a lidar-derived digital surface model Semantic segmentation [S18]
2012 VHR optical (QuickBird and WorldView-2), SAR (TerraSAR-X), and lidar data Open for creative ideas [S15]
2011 Multiangular optical images (WorldView-2) Open for creative ideas [S14]
2009-2010 Multitemporal optical (SPOT) and SAR (ERS) images Change detection [S1]
2008 VHR hyperspectral imagery Semantic segmentation [S17]
2007 Low-resolution SAR (ERS) and optical (Landsat) data Semantic segmentation [S10]
2006 Multispectral and panchromatic images Pansharpening [S21]

ERS: European Remote Sensing.

86 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


more than 1,000 registrations for the contest from nearly 80 different coun- [S9] C. Persello et al., “2023 IEEE GRSS data fusion contest: Large-scale fine-
tries. Influenced by different factors, including the overwhelming success of grained building classification for semantic urban reconstruction [Technical
datasets (and connected contests) in the Computer Vision community, an Committees],” IEEE Geosci. Remote Sens. Mag., vol. 11, no. 1, pp. 94–97,
increasing number of EO sensors with easier access to their data, as well as Mar. 2023, doi: 10.1109/MGRS.2023.3240233.
improved options for data hosting, the number of available benchmark [S10] F. Pacifici, F. Del Frate, W. J. Emery, P. Gamba, and J. Chanussot, “Urban
datasets (that were not always but often connected to a contest) increased mapping using coarse SAR and optical data: Outcome of the 2007 GRSS data
dramatically around 2015 (see also Figure 1). Interestingly, the growing fusion contest,” IEEE Geosci. Remote Sens. Lett., vol. 5, no. 3, pp. 331–335,
influence of Computer Vision on remote sensing, in particular due to deep Jul. 2008, doi: 10.1109/LGRS.2008.915939.
learning, is also reflected by renaming the Data Fusion Technical Committee [S11] N. Yokoya et al., “Open data for global multimodal land use classification:
to the Image Analysis and Data Fusion (IADF) Technical Committee in 2014. Outcome of the 2017 IEEE GRSS data fusion contest,” IEEE J. Sel. Topics Appl.
Since then, participation in the DFC has been more or less constant (with a Earth Observ. Remote Sens., vol. 11, no. 5, pp. 1363–1377, May 2018, doi:
few positive outliers such as 2019), with approximately 40 teams from 10.1109/JSTARS.2018.2799698.
roughly 20 countries. Another commendable fact is that at the beginning, [S12] Z. Li et al., “The outcome of the 2021 IEEE GRSS data fusion contest—Track
participants of the DFC were dominated by well-established scientists with MSD: Multitemporal semantic change detection,” IEEE J. Sel. Topics Appl.
solid experience in their respective fields. Although this group still plays a Earth Observ. Remote Sens., vol. 15, pp. 1643–1655, Jan. 2022, doi: 10.1109/
vital role in more recent DFCs, a large number of participants (and win- JSTARS.2022.3144318.
ners!) are students. This shows that improved data availability helped lower [S13] Y. Ma et al., “The outcome of the 2021 IEEE GRSS data fusion contest - Track
the starting threshold to analyze various EO data by standardizing data for- DSE: Detection of settlements without electricity,” IEEE J. Sel. Topics Appl.
mats, easing access to theoretical knowledge, and open sourcing software Earth Observ. Remote Sens., vol. 14, pp. 12,375–12,385, Nov. 2021, doi:
libraries and tools. 10.1109/JSTARS.2021.3130446.
[S14] F. Pacifici and Q. Du, “Foreword to the special issue on optical multiangular
References data exploitation and outcome of the 2011 GRSS data fusion contest,” IEEE J.
[S1] N. Longbotham et al., “Multi-modal change detection, application to the Sel. Topics Appl. Earth Observ. Remote Sens., vol. 5, no. 1, pp. 3–7, Feb. 2012,
detection of flooded areas: Outcome of the 2009–2010 data fusion contest,” doi: 10.1109/JSTARS.2012.2186733.
IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 5, no. 1, pp. 331– [S15] C. Berger et al., “Multi-modal and multi-temporal data fusion: Outcome of
342, Feb. 2012, doi: 10.1109/JSTARS.2011.2179638. the 2012 GRSS data fusion contest,” IEEE J. Sel. Topics Appl. Earth Observ.
[S2] L. Mou et al., “Multitemporal very high resolution from space: Outcome of Remote Sens., vol. 6, no. 3, pp. 1324–1340, Jun. 2013, doi: 10.1109/
the 2016 IEEE GRSS data fusion contest,” IEEE J. Sel. Topics Appl. Earth JSTARS.2013.2245860.
Observ. Remote Sens., vol. 10, no. 8, pp. 3435–3447, Aug. 2017, doi: 10.1109/ [S16] C. Robinson et al., “Global land-cover mapping with weak supervision:
JSTARS.2017.2696823. Outcome of the 2020 IEEE GRSS data fusion contest,” IEEE J. Sel. Topics Appl.
[S3] S. Kunwar et al., “Large-scale semantic 3D reconstruction: Outcome of the Earth Observ. Remote Sens., vol. 14, pp. 3185–3199, Mar. 2021, doi: 10.1109/
2019 IEEE GRSS data fusion contest - Part A,” IEEE J. Sel. Topics Appl. Earth JSTARS.2021.3063849.
Observ. Remote Sens., vol. 14, pp. 922–935, 2021, doi: 10.1109/ [S17] G. Licciardi et al., “Decision fusion for the classification of hyperspectral data:
JSTARS.2020.3032221. Outcome of the 2008 GRS-S data fusion contest,” IEEE Trans. Geosci.
[S4] W. Liao et al., “Processing of multiresolution thermal hyperspectral and digi- Remote Sens., vol. 47, no. 11, pp. 3857–3865, Nov. 2009, doi: 10.1109/
tal color data: Outcome of the 2014 IEEE GRSS data fusion contest,” IEEE J. TGRS.2009.2029340.
Sel. Topics Appl. Earth Observ. Remote Sens., vol. 8, no. 6, pp. 2984–2996, [S18] C. Debes et al., “Hyperspectral and LiDAR data fusion: Outcome of the 2013
Jun. 2015, doi: 10.1109/JSTARS.2015.2420582. GRSS data fusion contest,” IEEE J. Sel. Topics Appl. Earth Observ. Remote
[S5] M. Campos-Taberner et al., “Processing of extremely high-resolution Sens., vol. 7, no. 6, pp. 2405–2418, Jun. 2014, doi: 10.1109/JSTARS.2014.
LiDAR and RGB data: Outcome of the 2015 IEEE GRSS data fusion contest– 2305441.
part A: 2-D contest,” IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., [S19] Y. Lian et al., “Large-scale semantic 3-D reconstruction: Outcome of the 2019
vol. 9, no. 12, pp. 5547–5559, Dec. 2016, doi: 10.1109/JSTARS.2016. IEEE GRSS data fusion contest—Part B,” IEEE J. Sel. Topics Appl. Earth
2569162. Observ. Remote Sens., vol. 14, pp. 1158–1170, 2021, doi: 10.1109/
[S6] A.-V. Vo et al., “Processing of extremely high resolution LiDAR and RGB data: JSTARS.2020.3035274.
Outcome of the 2015 IEEE GRSS data fusion contest—Part B: 3-D contest,” [S20] M. Schmitt, L. H. Hughes, C. Qiu, and X. X. Zhu, “SEN12MS - A curated data-
IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 9, no. 12, pp. 5560– set of georeferenced multi-spectral sentinel-1/2 imagery for deep learning
5575, Dec. 2016, doi: 10.1109/JSTARS.2016.2581843. and data fusion,” ISPRS Ann. Photogrammetry Remote Sens. Spatial Inf.
[S7] Y. Xu et al., “Advanced multi-sensor optical remote sensing for urban land Sci., vol. IV-2/W7, pp. 153–160, Sep. 2019, doi: 10.5194/isprs-annals-IV-
use and land cover classification: Outcome of the 2018 IEEE GRSS data fusion 2-W7-153-2019.
contest,” IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 12, no. 6, [S21] L. A lparone, L. Wald, J. Chanussot, C. Thomas, P. Gamba, and L. M.
pp. 1709–1724, Jun. 2019, doi: 10.1109/JSTARS.2019.2911113. Bruce, “Comparison of pansharpening algorithms: Outcome of the
[S8] R. Hänsch et al., “The 2022 IEEE GRSS data fusion contest: Semisupervised 2006 GRS-S data-fusion contest,” IEEE Trans. Geosci. Remote Sens.,
learning [Technical Committees],” IEEE Geosci. Remote Sens. Mag., vol. 10, vol. 45, no. 10, pp. 3012–3021, Oct. 2007, doi: 10.1109/TGRS.2007.
no. 1, pp. 334–337, Mar. 2022, doi: 10.1109/MGRS.2022.3144291. 904923.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 87


FIGURE 25. An illustration of the map view of the EOD data catalog.

and interactive manner (see Figure 25 for an example of data. As PyTorch [39] and TensorFlow [40] (note that
the user interface). For more information, see “The IEEE Keras is now a part of TensorFlow) are the most widely
Geoscience and Remote Sensing Society Earth Observa- used DL frameworks, we focus mainly on the introduc-
tion Database.” tion for the existing libraries, using these two DL frame-
works as the back end.
EO-ORIENTED ML LIBRARIES 1) TorchGeo [72] is an open source PyTorch-based li-
Most of the existing ML libraries are developed for clas- brary, which provides datasets, samplers, transforms,
sic computer vision tasks, where the input image is usu- and pretrained models specific to geospatial data [41].
ally single-channel (grayscale) or with RGB bands. EO The main goal of this library is to simplify the pro-
datasets are often of large volumes with highly diverse cess of interacting with complex geospatial data and
data types, different numbers of spectral bands, and make it easier for researchers to train ML models for
spatial resolutions, as illustrated in Figure S1. The code EO tasks. Figure 26 provides an example of sampling
base for processing such data samples is often highly pixel-aligned patch data from heterogeneous geo-
complex and difficult to maintain. One approach to spatial data layers using the TorchGeo package. As
increase readability, reusability, and maintainability is different layers usually have different coordinate
to modularize the code and encapsulate different tasks systems and spatial resolutions, patches sampled
by decoupling processing the dataset (data loaders, vi- from these layers in the same area may not be pixel
sualization, preprocessing, and so on) and applying ML aligned. Therefore, in a practical application sce-
models (training, prediction, model selection, evalua- nario, researchers need to conduct a series of pre-
tion, and so forth). A major challenge in training ad- processing operations such as reprojecting and resa-
vanced ML models for EO tasks is the implementation mpling of the geospatial data before training ML
of an easy-to-use, yet efficient data loader explicitly de- models, which is time consuming and laborious.
signed for geoscience data that loads and preprocesses To address this challenge, TorchGeo provides data
a complex dataset and produces an iterable list of data loaders tailored for geospatial data, which support
samples in a customizable way. transparently loading data from heterogeneous geo-
This section introduces several well-known packages spatial data layers with relatively simple code, per
designed explicitly for geoscience and remote sensing the following example:

88 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


1 From torch.utils.data import DataLoader
2 from torchgeo.datasets import CDL, Land- The IEEE Geoscience and Remote Sensing Society Earth
sat8, stack _ samples Observation Database
3 from torchgeo.samplers import Random- The core of the Earth Observation Database (EOD) is its database of user-sub-
GeoSampler mitted datasets. It can be searched by a combination of various queries, includ-
4 ing location as well as different combinations of sensors and tasks. Search
5 # Loading Landsat8 and CDL layers results (as well as the whole database) can be viewed either in an illustrated list
6 landsat8 = Landsat8(root=”…”) view or in an interactive map view, indicating those datasets that cover only a
7 
cdl = CDL(root=”…”, download=True, specific location with a marker (see Figure 25). Clicking on one of the markers
checksum=True) in the map view will limit the list to datasets at this specific location. Clicking on
8 a dataset in the list will open a new window, displaying detailed information
9 # Take the intersection of Landsat8 and CDL about this dataset, which includes
10 dataset = landsat8 & cdl ◗◗ geographic location
11 ◗◗ sensor modality
12 # Sample 10,000 256 x 256 image patches ◗◗ task/application
13 sampler = RandomGeoSampler(dataset, size ◗◗ data size
= 256, length = 10000) ◗◗ a Uniform Resource Locator to access the data
14 ◗◗ number of views
15 # Build a normal PyTorch DataLoader with ◗◗ a graphical representation
the sampler ◗◗ a brief description of the dataset.
16 dataloader = DataLoader(dataset, A helpful function is the compare option, which allows a direct side-to-side
batch _ size = 128, sampler=sampler, comparison of this information from two or more datasets in a newly opened
collate _ fn=stack _ samples) window. EOD is meant as a community-driven catalog maintained by the IEEE
17 Geoscience and Remote Sensing Society Image Analysis and Data Fusion (IADF)
18 for batch in dataloader: Technical Committee, i.e., adding datasets is neither limited to IADF members
19      image = batch[“image”] nor to the creators of the dataset, but to anybody with an interest who increases
20     mask = batch[“mask”] the visibility and accessibility of a certain dataset can request to include it in
21 EOD.The submission requests are reviewed by the IADF for completeness and
22 # Train a model correctness before the dataset is added to the database and visible to the public.
23 …

A more detailed introduction about the supported ning the model. It supports loading raster data from
geospatial datasets in TorchGeo can be found in [41]. local files or cloud services. Necessary preprocessing,
2) RasterVision [73] is an open source Python framework like reprojecting and resampling, is also conducted auto-
that aims to simplify the procedure for building DL- matically. Keras Spatial supports sample a­ugmentation
based computer vision models on satellite, aerial, and
other types of geospatial data (including oblique drone
Landsat 8 Scene
imagery). It enables users to efficiently construct a DL EPSG: 32617
A
pipeline, including training data preparation, model C
training, model evaluation, and model deployment,
without any expert knowledge of ML. Specifically,
RasterVision supports chip classification, object detec-
tion, and semantic segmentation with the PyTorch back
end on both CPUs and GPUs, with built-in support D
for running in the cloud using AWS. The framework
is extensible to new data sources, tasks (e.g., instance
segmentation), back ends (e.g., Detectron2), and cloud
providers. Figure 27 provides the pipeline of the Raster- B Cropland Data Layer
Vision package. A more comprehensive tutorial for this EPSG: 5072

package can be found in [74].


3) Keras Spatial [75] provides data samplers and tools de- FIGURE 26. An illustration of sampling from heterogeneous
signed to simplify the preprocessing of geospatial data geospatial data layers. (Source: [41].) With the TorchGeo package,
for DL applications with the Keras back end. It provides users can focus directly on training ML models without manually
a data generator that reads samples directly from raster preprocessing the heterogeneous geospatial data, including align-
layers without creating small patch files before run- ing data layers by reprojection and resampling.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 89


using a user-defined callback system to improve These datasets may be time consuming and expensive to
the flexibility of data management. Here, a simple acquire on one’s own but can be easily accessed through a
demo code using the SpatialDataGenerator geospatial computing platform using the cloud service.
class from Keras Spatial to prepare a training set for a ◗◗ Scalability: geospatial computing platforms are designed
DL model is given: to address large-scale geospatial data processing and
analysis. They usually provide cloud-based computing
1 from keras _ spatial import SpatialData- resources that can be easily scaled up or down to meet re-
Generator searchers’ needs. This makes it easier to perform complex
2 geospatial analysis that would be difficult to do on local
3 # Loading labels from a local file machines considering the limited computing resources.
4 labels = SpatialDataGenerator() ◗◗ Prebuilt tools and APIs: geospatial computing platforms usu-
5 labels.source = ‘/path/to/labels.tif’ ally provide prebuilt tools and programming APIs for im-
6 # Sample 128 x 128 patches age processing, geocoding, and data visualization, making
7 labels.width, labels.height = 128, 128 it much easier for researchers to work with geospatial data.
8 # Set a geodataframe with 200 x 200 Several representative geospatial computing platforms are
9 # in projection units of the original raster as follows:
10 df = labels.regular _ grid(200,200) ◗◗ GEE [76] is a cloud-based platform designed for large-
11 scale geospatial data analysis and processing. It provides
12 # Loading images from a file on the cloud a range of tools and APIs that allow users to analyze and
13 samples = SpatialDataGenerator() visualize geospatial data, including raster and vector
14 samples.source = ‘https://server.com/ processing tools, ML algorithms, and geospatial mod-
files/data.tif’ eling tools. In addition, GEE provides access to power-
15 samples.width, samples.height = labels. ful computing resources, such as virtual machines and
width, label.height storage, to enable users to perform complex geospatial
16 analyses. The GEE Data Catalog contains more than
17 # The training set generator 40 years of historical imagery and scientific datasets for
18 train _ gen = zip(labels.flow _ from _ Earth science, which are updated and expanded daily.
dataframe(df), patches.flow _ from _ These datasets cover various topics such as climate,
dataframe(df)) weather, surface temperature, terrain, and land cover.
19 Notable datasets available on the GEE Data Catalog in-
20 # Train a model clude Planet SkySat Public Ortho Imagery (collected for
21 model(train _ gen) crisis response events) [42] and NAIP [77] (agricultural
monitoring data in the United States).
GEOSPATIAL COMPUTING PLATFORMS ◗◗ AWS [78] is a powerful platform for geospatial computing
In addition to ML libraries, public geospatial computing that offers a range of services for geospatial data storage,
platforms such as Google Earth Engine (GEE) for EO tasks processing, and analysis. AWS hosts several representa-
in practical application scenarios offer a series of benefits, tive geospatial datasets, including Digital Earth Africa
including the following: [79] (Landsat and Sentinel products over Africa), National
◗◗ Access to large-scale datasets: geospatial computing platforms Oceanic and Atmospheric Administration Emergency
usually provide access to large and diverse geospatial datas- Response Imagery [80] (lidar and hyperspectral data
ets, such as satellite imagery, weather data, and terrain data. over the United States), and datasets from the SpaceNet

Deployment
Inputs RasterVision Pipeline

Batch
Images
Predictions
ANALYZE CHIP TRAIN PREDICT EVAL BUNDLE

Live
Labels
Predictions
Dataset Training Training Validation Evaluation Model
Metrics Chips Model Predictions Metrics Bundle
Custom
AOI
Integrations

FIGURE 27. An illustration of the pipeline from the RasterVision package. AOI: area of interest. (Source: https://github.com/azavea/raster-vision.)

90 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


challenges (Table 2; 134, 264, 298, 299, 300, 304, 306, samples that are extremely limited in their spatial and
and 312) [43]. Sharing data on AWS allows anyone to temporal distribution, to the extreme of consisting of a
analyze them and build services using a broad range of single, small image only. Not only are such datasets prone
computing and data analytical products like Amazon to leading to biased evaluation protocols, where informa-
EC2, which enables data users to spend more time on tion from the training set leaks into the test set via spatial
data analysis rather than data acquisition. correlation, they are usually also not sufficient to train
◗◗ Microsoft Planetary Computer [81] provides access to a models that are able to gen-
wide range of EO data and powerful geospatial comput- eralize to other geographic
ing resources. The platform is specifically designed to areas or time points (e.g., dif- THE GEE DATA CATALOG
support the development of innovative applications and ferent seasons). More mod- CONTAINS MORE THAN 40
solutions for addressing critical environmental challeng- ern datasets aim to increase YEARS OF HISTORICAL
es, such as climate change, biodiversity loss, and natural diversity regarding scene
IMAGERY AND SCIENTIFIC
resource management. It offers cloud-based computing content (e.g., more different
DATASETS FOR EARTH
services that enable users to perform complex geospa- object instances), environ-
tial analyses efficiently. The Planetary Computer Data mental factors (e.g., seasons, SCIENCE, WHICH ARE
Catalog provides access to petabytes of environmental light conditions, and so on), UPDATED AND EXPANDED
monitoring data in a consistent and analysis-ready for- or aspects regarding image DAILY.
mat. Some of the representative datasets available on the acquisition/processing (e.g.,
Planetary Computer Data Catalog include the HREA da- different look angles, resolu-
taset [82] (settlement-level measures of electricity access tions, and so forth). This increase of diversity has thus
derived from satellite images) and the Microsoft Building far always connected to an increase in dataset size by in-
Footprints dataset [83]. cluding more images and/or larger scenes, or even other
◗◗ Colab [84] provides a flexible geospatial computing modalities. Although an increase in the provided image
platform that offers users access to free GPU and ten- data is often easily possible, it is usually not feasible to
sor processing units (TPUs) computing resources for have the same increase in the reference data. This leads
analyzing geospatial data. It provides a free Jupyter to large-scale datasets where the reference data are much
notebook environment that allows users to write and less carefully curated as for early datasets, often based on
run Python code for accessing and processing geospa- other existing sources (e.g., OpenStreetMap) and con-
tial data from various platforms (e.g., the GEE Data taining more label noise. Although many ML methods
Catalog). Colab notebooks can be easily shared and can handle a certain extent of label noise during training,
collaborated on with others, which is particularly use- its effects on the evaluation are barely understood and
ful for geospatial analysis projects that involve mul- often ignored.
tiple team members.
◗◗ Kaggle [86] is an online platform widely used for data CURRENT TRENDS: SPECIFICITY AND GENERALITY
science and ML competitions. It provides a cloud-based In this context, two main characteristics of a dataset for a
computational environment that allows for reproduc- given task will be the focus of the discussion: specificity
ible and collaborative analysis in the field of geospatial and generality.
computing. Kaggle also provides access to a variety of EO applications present numerous and diverse scenarios
geospatial datasets, including satellite imagery, terrain due to varying object types, semantic classes, sensor types
data, and weather data, along with tools to analyze and and modalities, spatial, spectral, and temporal resolutions,
visualize these datasets. Users can take advantage of and coverage (global, regional, or local).
the platform’s free GPU and TPU computing resources High specificity refers to datasets that are strongly tai-
to train ML models and undertake complex geospatial lored to a specific sensor-platform-task combination,
analysis tasks. maybe even being limited to certain acquisition modes,
geographic regions, or meteorological seasons. These can
OPEN CHALLENGES AND FUTURE DIRECTIONS hardly be used for anything beyond their original pur-
The previous sections give an overview of existing bench- pose. Although different application domains such as ag-
marking datasets, presenting their main features and de- riculture, urban mapping, military target detection, and
scribing their main characteristics, eventually providing bio-/geophysical parameter extraction do require differ-
a broad yet detailed picture of the current state of the art. ent types of data, i.e., images, point measurements, tables,
This section discusses existing gaps, open challenges, and and metadata, the proliferation of different datasets spe-
potential future trends. cialized for every small task variation reduces reusability
of the datasets in a broader context and affects scientific
WHERE ARE WE NOW? transparency and reproducibility.
The most prominent issue that new dataset releases aim to High specificity also contributes to cluttered nomen-
fix is a lack of diversity. Many of the earlier datasets contain clature, causing different datasets to appear different while

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 91


actually sharing very similar content. For example, a high gaps. Furthermore, although multimodal datasets enable
level of detail in class semantics and terminology makes it data-fusion and cross-modal approaches that leverage the
difficult to compare the reference data of different datasets. different input sources, multitask datasets allow exploit-
A typical example is land cover classification, where simi- ing the mutual overlap of related tasks regarding feature
lar classes may be aggregated into different subcategories extraction and representation learning. Finally, the idea
depending on the application. As a result, models trained of loosening the previously tight relationship between
on different application-specific datasets may actually ap- input data and the target variable in datasets (up to the
proximate very similar functional relationships between point where a dataset might not offer reference data for
image data and target variables. any target variable) is to provide data that can be lever-
Virtually all of the less recent and still most of the mod- aged to learn powerful representations that are useful for
ern datasets aim for specificity. However, several of the a large variety of downstream tasks (as in pretraining or
more recent benchmarks follow another direction: gen- self-supervised learning).
erality, i.e., providing more sensor modalities than actu- However, there is not yet a single “go-to” dataset that can
ally required plus large-scale, often noisy reference data for be used for pretraining most of the newly developed mod-
multiple tasks instead of small scale and carefully curated els or for benchmarking specific tasks against state-of-the-
annotations that only address a single task. art approaches. Collecting such a high-quality benchmark
The contribution of such general datasets is manifold: dataset that enables pretraining of models for as many
first and foremost, the required number of (annotated) downstream tasks as possible is of significant value for fur-
training samples for fully supervised ML simply does not ther pushing performance boundaries.
scale very well given the effort of data annotation and cu- Figure 28 presents a schematic diagram of the proper-
ration in remote sensing. Thus, such general, large-scale ties of an ideal solution for a go-to EO benchmark data-
datasets introduce new factors that increase the relation set, covering diverse geolocations, multiple modalities,
to realistic application scenarios such as robustness to different acquisition scenarios, and various applications.
label noise (e.g., by leveraging existing semantic maps It is ideally acquired by different types of sensors and
as reference data, which are often outdated, misaligned, platforms with different viewing geometries to cover
or of coarse resolution) and weakly supervised learning objects from different look angles. The images are ob-
(where the reference data have a lower level of detail than tained from different electromagnetic spectrum bands,
the actual target variable, e.g., training semantic segmen- i.e., visible, infrared, thermal, and microwave, resulting
tation networks with labels on image level). Large-scale in multi-/hyperspectral, SAR, lidar, optical, thermal, and
datasets are the only option to realistically test the gener- passive microwave measurements. The reference infor-
alization capabilities of learning-based models, e.g., over mation or annotations are provided on a level that allows
different geographic regions, seasons, or other domain defining various tasks based on a single annotation. For
example, an image with dense semantic annotations al-
lows users to generate their desired object instance anno-
The Ideal Pretraining Dataset tation files. Extending the dataset to multiple images of
A dataset ideally suited for pretraining and/or self-supervised learning should a scene with corresponding semantic labels enables not
adhere to as many of the following characteristics as possible: only semantic segmentation but also semantic CD tasks.
◗◗ multiple platforms (vehicle, drone, airplane, and satellite) In summary, we foresee a certain duality in the future de-
◗◗ multiple sensors (Planet, SPOT, WorldView, Landsat, Sentinel 1/2, and so velopment of EO-related datasets: on the one hand, follow-
forth) ing the paradigm of data-centric ML [44], i.e., moving against
◗◗ several acquisition modalities (SAR, RGB, hyperspectral, multispectral, ther- the current trend of creating performance gains merely by
mal, lidar, passive microwave, and so on) leveraging more training data but instead focusing on datas-
◗◗ diverse acquisition geometries (viewing angles, e.g., off-nadir conditions and ets tailored toward specific problems with well-curated, high-
spatial and temporal baselines in multiview data, e.g., interferometric SAR) quality reference data (e.g., manually annotated or based on
◗◗ realistic distortion factors (cloud cover, dust, smog, fog, atmospheric influence, official sources). On the other hand, general datasets that cov-
spatial misalignments and temporal changes in multiview data, and so forth) er as many input and output modalities as possible to allow
◗◗ well distributed geographical locations (spatial distribution within the dataset, learning generic representations that are of value for a large
climate zones, socioeconomic and cultural factors, and different topographies) number of possible downstream tasks.
◗◗ diverse land cover/use (urban, rural, forest, agricultural, water, and so on)
◗◗ varying spatial resolution (0.1–1 m, 3–10 m, 10–30 m, 100–500 m, and scale FINDABILITY, ACCESSIBILITY, INTEROPERABILITY,
distribution) REUSE AND ARD
◗◗ temporally well distributed (seasonality, lighting condition, sun angle, and In addition to the content, scope, and purpose of datas-
nighttime imagery) ets, their organization will gain importance. With only a
◗◗ a diverse set of reference data that are well aligned with the EO measurements dozen public datasets available prior to 2015, it was fea-
(semantic information, change, geo/biophysical parameters, and so forth). sible that each is provided with its own data format and
meta-information, hosted on individual web pages, and

92 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Pansharpening/Superresolution
Different Satellites Multiresolution Data Applications
Different Viewing Angles Multi-Modal Data Applications

Lidar
SAR
Multi-/Hyperspectral Unsupervised CD
Different Geolocations

Time-Series Analysis Applications

t1 tn

Object Detection

Annotations
Semantic Segmentation
Spring Summer Autumn Winter Instance Segmentation
Different/Multiple Times of the Year

Supervised CD

FIGURE 28. An illustration that shows the authors’ view of the paramount properties that an ideal benchmark dataset needs to satisfy,
including the type of tasks, sensors, temporal constraints, and geolocalization.

downloaded by everyone who wanted to work with them. boom has not only led to a rise in dataset numbers but also
With the hundreds of datasets available today and many a large increase in size (as in spatial coverage and resolution
more published every year, this cannot be maintained. but also with respect to multimodal and multitemporal im-
Concepts such as findability, accessibility, interoperabili- agery) and a diversity of application tasks. Furthermore, this
ty, and reuse (FAIR) (see, for example, [45]) were proposed development has led to the implementation of dedicated
years ago and are still of high relevance. Data catalogs software packages and meta databases that help interested
such as EOD (see the “Working With Remote Sensing Da- users develop solutions for their applications. Eventually,
tasets” section) are a first step toward structure datasets we drew the conclusion that one of the critical challenges
that are scattered among different data hosts. ARD (see, in dataset design for EO tasks is the strong heterogeneity of
e.g., [46]), for example, in the form of data cubes [47], possible sensors, data, and applications, which has led to
and efforts to homogenize meta-information, e.g., in the a jungle of narrow-focused datasets. Although one of the
form of datasheets [48], will continue to evolve into new trends in DL is certainly the consideration of smaller, well-
standardized data formats. The trend of datasets growing curated, and task-specific datasets, another direction is the
in terms of size and volume (see the “Evolution of EO- creation of a generic, nearly sensor- and task-agnostic data-
Oriented ML Datasets” section) as well as the need for base similar to the well-known ImageNet dataset used in
global data products will soon put a stop to the current Computer Vision. Such a generic dataset will be especially
practice of downloading datasets and processing them lo- valuable in the pretraining of large high-capacity models
cally. Working with data on distributed cloud services will with worldwide applicability.
create new requirements regarding data formats but also
lead to new standards and best practices for processing. AUTHOR INFORMATION
Finally, the goal of any ML-centered dataset is to train an MICHAEL SCHMITT (michael.schmitt@unibw.de) received
ML model. These models should be treated similar to the his Dipl.-Ing. degree in geodesy and geoinformation, his
data they originated from, i.e., they should follow stan- Dr.-Ing. Degree in remote sensing, and his habilitation
dardized data formats and FAIR principles. in data fusion from the Technical University of Munich
(TUM), Germany, in 2009, 2014, and 2018, respectively.
SUMMARY AND CONCLUSION Since 2021, he has been a full professor for Earth obser-
This article discussed the relevance of ML-oriented datas- vation at the Department of Aerospace Engineering of the
ets in the technical disciplines of EO and remote sensing. University of the Bundeswehr Munich, 85577 Neubiberg,
An analysis of historical developments shows that the DL Germany. Before that, he was a professor of applied geodesy

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 93


and remote sensing at the Department of Geoinformatics, at Istanbul Technical University Istanbul, 34469 Turkey.
Munich University of Applied Sciences. From 2015 to 2020, She was a visiting scholar at the School of Electrical and
he was a senior researcher and deputy head at the Profes- Computer Engineering and the School of Civil Engineer-
sorship for Signal Processing in Earth Observation at TUM; ing at Purdue University from 2008 to 2009 and 2016 to
in 2019, he was additionally appointed adjunct teaching 2017. She is a reviewer for Photogrammetric Engineering and
professor at the Department of Aerospace and Geodesy of Remote Sensing, IEEE Transactions on Geoscience and Remote
TUM. In 2016, he was a guest scientist at the University of Sensing, IEEE Journal of Selected Topics in Applied Earth Ob-
Massachusetts, Amherst. He is a cochair of the Active Mi- servations and Remote Sensing, IEEE Geoscience and Remote
crowave Remote Sensing Working Group of the Interna- Sensing Letters, and IEEE Transactions on Image Processing.
tional Society for Photogrammetry and Remote Sensing, Her current research interests include machine learning
and also of the Benchmarking Working Group of the IEEE approaches in hyperspectral image analysis, dimension-
Geoscience and Remote Sensing Society Image Analy- ality reduction, explainable artificial intelligence, and
sis and Data Fusion Technical Committee. He frequently sensitivity analysis.
serves as a reviewer for a number of renowned internation- UJJWAL VERMA (ujjwal.verma@manipal.edu) received
al journals and conferences and has received several Best his Ph.D. degree from Télécom ParisTech, University of
Reviewer Awards. His research focuses on technical aspects Paris-Saclay, Paris, France, in image analysis and his M.S.
of Earth observation, in particular image analysis and ma- degree (Research) from IMT Atlantique (France) in signal
chine learning applied to the extraction of information and image processing. He is currently an associate professor
from multisensor remote sensing observations. Among his and head of the Department of Electronics and Communi-
core interests is remote sensing data fusion with a focus on cation Engineering at Manipal Institute of Technology, Ben-
synthetic aperture radar and optical data. He is a Senior galuru 560064, India. He is a recipient of the ISCA Young
Member of IEEE. Scientist Award 2017–2018 by the Indian Science Congress
SEYED ALI AHMADI (cpt.ahmadisnipiol@yahoo.com) Association, a professional body under the Department of
received his B.Sc. degree in surveying engineering and his Science and Technology, Government of India. He is also a
M.Sc. degree in remote sensing from the Faculty of Geod- recipient of the Young Professional Volunteer Award 2020
esy and Geomatics, K.N. Toosi University of Technology, by the IEEE Mangalore Subsection in recognition of his
Tehran 19697, Iran, in 2015 and 2017, respectively, where outstanding contribution to IEEE activities. He is a co-lead
he is currently pursuing his Ph.D. thesis on building dam- of the Working Group on Machine/Deep Learning for Im-
age assessment. He worked on image classification and age Analysis of the Image Analysis and the IEEE Geosci-
segmentation techniques, machine learning algorithms, ence and Remote Sensing Society Data Fusion Technical
and lidar data processing. His thesis was focused on clas- Committee. He is a guest editor for Special Stream in IEEE
sifying hyperspectral and lidar datasets by combining Geoscience and Remote Sensing Letters and a reviewer for sev-
spectral and spatial features to increase classification eral journals including IEEE Transactions on Image Process-
accuracy. He is a cochair of the Benchmarking Working ing, IEEE Transactions on Geoscience and Remote Sensing, IEEE
Group of the IEEE Geoscience and Remote Sensing Soci- Geoscience and Remote Sensing Letters. He is also a sectional
ety Image Analysis and Data Fusion Technical Committee, recorder for the Information and Communication Technol-
frequently serves as a reviewer for a number of interna- ogy Section of the Indian Science Congress Association for
tional journals, and received a Best Reviewer Award in 2020–2024. His research interests include computer vision
2018. His research interests include machine learning, and machine learning, focusing on variational methods in
deep learning, geospatial data analysis, image processing, image segmentation, deep learning methods for scene un-
and computer vision techniques for remote sensing and derstanding, and semantic segmentation of aerial images.
Earth observation applications. He is a Senior Member of IEEE.
YONGHAO XU (yonghaoxu@ieee.org) received his B.S. FRANCESCOPAOLO SICA (francescopaolo.sica@unibw.
and Ph.D. degrees in photogrammetry and remote sensing de) received his laurea (M.S.) degree (summa cum laude)
from Wuhan University, China, in 2016 and 2021, respec- in telecommunications engineering and his Dr. Ing.
tively. He is currently a postdoctoral researcher at the Insti- (Ph.D.) degree in information engineering from the
tute of Advanced Research in Artificial Intelligence, 1030 University of Naples Federico II, Italy, in 2012 and 2016,
Vienna, Austria. His research interests include remote sens- respectively. Since 2022, he has been deputy head of the
ing, computer vision, and machine learning. He is a Mem- Earth Observation Laboratory at the Department of Aero-
ber of IEEE. space Engineering of the University of the Bundeswehr
GÜLS,EN TAS,KIN (gulsen.taskin@itu.edu.tr) received Munich, 85577 Neubiberg, Germany. Between 2016 and
her B.S. degree in geomatics engineering and her M.S. 2022, he was a researcher at the German Aerospace Cen-
and Ph.D. degrees in computational science and engi- ter. He received a Living Planet Post-Doctoral Fellowship
neering from Istanbul Technical University, Turkey, in from the European Space Agency for the High-Resolution
2001, 2003, and 2011, respectively. She is currently an as- Forest Coverage with InSAR & Deforestation Surveillance)
sociate professor at the Institute of Disaster Management project. He is a cochair of the Benchmarking Working

94 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Group of the IEEE Geoscience and Remote Sensing Soci- Sel. Topics Appl. Earth Observ. Remote Sens., vol. 14, pp. 4205–
ety Image Analysis and Data Fusion Technical Committee 4230, Apr. 2021, doi: 10.1109/JSTARS.2021.3070368.
and a regular reviewer for international journals and con- [5] M. Schmitt, S. A. Ahmadi, and R. Hänsch, “There is no data
ferences. His research interests cover a wide range of activ- like more data - Current status of machine learning datasets in
ities related to synthetic aperture radar (SAR) technology, remote sensing,” in Proc. Int. Geosci. Remote Sens. Symp., 2021,
from mission design to SAR signal and image processing, pp. 1206–1209, doi: 10.1109/IGARSS47720.2021.9555129.
to end-user applications. He is a Member of IEEE. [6] M. Schmitt, P. Ghamisi, N. Yokoya, and R. Hänsch, “EOD:
RONNY HÄNSCH (ronny.haensch@dlr.de) received his The IEEE GRSS Earth observation database,” in Proc. Int.
diploma in computer science and his Ph.D. degree from Geosci. Remote Sens. Symp., 2022, pp. 5365–5368, doi: 10.1109/
TU Berlin, Germany, in 2007 and 2014, respectively. He IGARSS46834.2022.9884725.
is a scientist at the Microwave and Radar Institute of the [7] Y. Yang and S. Newsam, “Bag-of-visual-words and spatial ex-
German Aerospace Center, 82234 Weßling, Germany, tensions for land-use classification,” in Proc. SIGSPATIAL Int.
where he leads the Machine Learning Team in the Signal Conf. Adv. Geographic Inf. Syst., 2010, pp. 270–279, doi: 10.1145/
Processing Group of the SAR Technology Department. 1869790.1869829.
He continues to lecture at TU Berlin in the Computer Vi- [8] F. Rottensteiner et al., “The ISPRS benchmark on urban object
sion and Remote Sensing Group. He serves as chair of classification and 3D building reconstruction,” ISPRS Ann.
the IEEE Geoscience and Remote Sensing Society Image Photogrammetry Remote Sens. Spatial Inf. Sci., vol. I-3, no. 1, pp.
Analysis and Data Fusion Technical Committee, cochair 293–298, Sep. 2012, doi: 10.5194/isprsannals-I-3-293-2012.
of the ISPRS Working Group on Image Orientation and [9] P. Ghamisi and N. Yokoya, “IMG2DSM: Height simulation from
Fusion, IEEE Geoscience and Remote Sensing Society single imagery using conditional generative adversarial net,”
(GRSS) membership chair, organizer of the IEEE Confer- IEEE Geosci. Remote Sens. Lett., vol. 15, no. 5, pp. 794–798, May
ence on Computer Vision and Pattern Recognition Work- 2018, doi: 10.1109/LGRS.2018.2806945.
shops with EarthVision (2017–2023), Photogrammetric [10] C. Benedek, X. Descombes, and J. Zerubia, “Building devel-
Computer Vision (2019 and 2023), the Machine Learn- opment monitoring in multitemporal remotely sensed image
ing for Remote Sensing Data Analysis Workshop at the pairs with stochastic birth-death dynamics,” IEEE Trans. Pat-
International Conference on Learning Representations tern Anal. Mach. Intell., vol. 34, no. 1, pp. 33–50, Jan. 2012, doi:
(2023), and the IEEE International Geoscience and Re- 10.1109/TPAMI.2011.94.
mote Sensing Symposium Tutorial on Machine Learning [11] L. Zhang, L. Zhang, and B. Du, “Deep learning for remote sens-
in Remote Sensing (2017–2023). He also serves as editor ing data: A technical tutorial on the state of the art,” IEEE Geosci.
of IEEE Geoscience and Remote Sensing Society eNewsletter Remote Sens. Mag. (replaced Newsletter), vol. 4, no. 2, pp. 22–40,
and associate editor of Geoscience and Remote Sensing Let- Jun. 2016, doi: 10.1109/MGRS.2016.2540798.
ters and ISPRS Journal of Photogrammetry and Remote Sens- [12] X. X. Zhu et al., “Deep learning in remote sensing: A compre-
ing. He has extensive experience in organizing remote hensive review and list of resources,” IEEE Geosci. Remote Sens.
sensing community competitions (e.g., the IEEE GRSS Mag. (replaced Newsletter), vol. 5, no. 4, pp. 8–36, Dec. 2017, doi:
Data Fusion Contest 2018–2023), serves as the GRSS 10.1109/MGRS.2017.2762307.
representative within SpaceNet, and was technical lead [13] Q. Yuan et al., “Deep learning in environmental remote sens-
of the SpaceNet 8 Challenge. His research interest is ing: Achievements and challenges,” Remote Sens. Environ., vol.
computer vision and machine learning with a focus on 241, Feb. 2020, Art. no. 111716, doi: 10.1016/j.rse.2020.111716.
remote sensing (in particular, synthetic aperture radar [14] G. Christie, N. Fendley, J. Wilson, and R. Mukherjee, “Func-
processing and analysis). He is a Senior Member of IEEE. tional map of the world,” in Proc. IEEE Conf. Comput. Vis.
Pattern Recognit., 2018, pp. 6172–6180, doi: 10.1109/CVPR.
REFERENCES 2018.00646.
[1] G. Cheng, J. Han, and X. Lu, “Remote sensing image scene classifi- [15] I. Kotaridis and M. Lazaridou, “Remote sensing image seg-
cation: Benchmark and state of the art,” Proc. IEEE, vol. 105, no. 10, mentation advances: A meta-analysis,” ISPRS J. Photogrammetry
pp. 1865–1883, Oct. 2017, doi: 10.1109/JPROC.2017.2675998. Remote Sens., vol. 173, pp. 309–322, Mar. 2021, doi: 10.1016/j.
[2] D. Hong, J. Hu, J. Yao, J. Chanussot, and X. X. Zhu, “Multimodal isprsjprs.2021.01.020. [Online]. Available: https://www.science
remote sensing benchmark datasets for land cover classification direct.com/science/article/pii/S0924271621000265
with a shared and specific feature learning model,” ISPRS J. Pho- [16] M. Schmitt, L. H. Hughes, C. Qiu, and X. X. Zhu, “SEN12MS – A
togrammetry Remote Sens., vol. 178, nos. 9–10, pp. 68–80, Aug. curated dataset of georeferenced multi-spectral sentinel-1/2 im-
2021, doi: 10.1016/j.isprsjprs.2021.05.011. agery for deep learning and data fusion,” ISPRS Ann. Photogram-
[3] K. Li, G. Wan, G. Chen, L. Meng, and J. Han, “Object detection metry Remote Sens. Spatial Inf. Sci., vol. IV-2/W7, pp. 153–160,
in optical remote sensing images: A survey and a new bench- Sep. 2019, doi: 10.5194/isprs-annals-IV-2-W7-153-2019.
mark,” ISPRS J. Photogrammetry Remote Sens., vol. 159, pp. 296– [17] M. Schmitt and Y.-L. Wu, “Remote sensing image classification
307, Jan. 2020, doi: 10.1016/j.isprsjprs.2019.11.023. with the sen12ms dataset,” ISPRS Ann. Photogrammetry, Remote
[4] Y. Long et al., “On creating benchmark dataset for aerial image Sens. Spatial Inf. Sci., vol. V-2-2021, pp. 101–106, Apr. 2021, doi:
interpretation: Reviews, guidances, and million-AID,” IEEE J. 10.5194/isprs-annals-V-2-2021-101-2021.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 95


[18] F. Rottensteiner, G. Sohn, M. Gerke, and J. D. Wegner, “ISPRS [30] G.-S. Xia et al., “DOTA: A large-scale dataset for object detec-
test project on urban classification and 3D building reconstruc- tion in aerial images,” in Proc. IEEE Conf. Comput. Vis. Pattern
tion - 2D semantic labeling - Vaihingen data,” Int. Soc. Photo- Recognit., Jun. 2018.
grammetry Remote Sens., Hannover, Germany, 2013. Accessed: [31] J. Ding, N. Xue, Y. Long, G.-S. Xia, and Q. Lu, “Learning RoI
Oct. 14, 2022. [Online]. Available: https://www.isprs.org/educa transformer for oriented object detection in aerial images,” in
tion/benchmarks/UrbanSemLab/2d-sem-label-vaihingen.aspx Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2019, pp.
[19] M. Cramer, “The DGPF-Test on digital airborne camera evalu- 2844–2853, doi: 10.1109/CVPR.2019.00296.
ation - Overview and test design,” Photogrammetrie Ferner- [32] J. Ding et al., “Object detection in aerial images: A large-scale
kundung Geoinf., vol. 2010, no. 2, pp. 73–82, May 2010, doi: benchmark and challenges,” IEEE Trans. Pattern Anal. Mach. In-
10.1127/1432-8364/2010/0041. tell., vol. 44, no. 11, pp. 7778–7796, Nov. 2022, doi: 10.1109/
[20] Y. Wang and M. Li, “Urban impervious surface detection from TPAMI.2021.3117983.
remote sensing images: A review of the methods and challeng- [33] H. Chen and Z. Shi, “A spatial-temporal attention-based meth-
es,” IEEE Geosci. Remote Sens. Mag., vol. 7, no. 3, pp. 64–93, Sep. od and a new dataset for remote sensing image change detec-
2019, doi: 10.1109/MGRS.2019.2927260. tion,” Remote Sens., vol. 12, no. 10, May 2020, Art. no. 1662, doi:
[21] D. Wen et al., “Change detection from very-high-spatial-resolu- 10.3390/rs12101662. [Online]. Available: https://www.mdpi.
tion optical remote sensing images: Methods, applications, and com/2072-4292/12/10/1662
future directions,” IEEE Geosci. Remote Sens. Mag., vol. 9, no. 4, [34] R. C. Daudt, B. Le Saux, A. Boulch, and Y. Gousseau, “Urban
pp. 68–101, Dec. 2021, doi: 10.1109/MGRS.2021.3063465. change detection for multispectral earth observation using convo-
[22] A. Karpatne, Z. Jiang, R. R. Vatsavai, S. Shekhar, and V. Kumar, lutional neural networks,” in Proc. Int. Geosci. Remote Sens. Symp.,
“Monitoring land-cover changes: A machine-learning perspec- Jul. 2018, pp. 2115–2118, doi: 10.1109/IGARSS.2018.8518015.
tive,” IEEE Geosci. Remote Sens. Mag., vol. 4, no. 2, pp. 8–21, Jun. [35] M. Märtens, D. Izzo, A. Krzic, and D. Cox, “Super-resolution of
2016, doi: 10.1109/MGRS.2016.2528038. PROBA-v images using convolutional neural networks,” 2019.
[23] P. Helber, B. Bischke, A. Dengel, and D. Borth, “EuroSAT: A nov- [Online]. Available: https://arxiv.org/abs/1907.01821
el dataset and deep learning benchmark for land use and land [36] G. Vivone, M. Dalla Mura, A. Garzelli, and F. Pacifici, “A bench-
cover classification,” IEEE J. Sel. Topics Appl. Earth Observ. Re- marking protocol for pansharpening: Dataset, preprocessing,
mote Sens., vol. 12, no. 7, pp. 2217–2226, Jul. 2019, doi: 10.1109/ and quality assessment,” IEEE J. Sel. Topics Appl. Earth Ob-
JSTARS.2019.2918242. serv. Remote Sens., vol. 14, pp. 6102–6118, 2021, doi: 10.1109/
[24] P. Helber, B. Bischke, A. Dengel, and D. Borth, “Introducing JSTARS.2021.3086877.
EuroSAT: A novel dataset and deep learning benchmark for [37] J. Cornebise, I. Orsolic, and F. Kalaitzis, “Open high-resolution
land use and land cover classification,” in Proc. Int. Geosci. satellite imagery: The worldstrat dataset – With application to
Remote Sens. Symp., 2018, pp. 204–207, doi: 10.1109/IGARSS. super-resolution,” in Proc. 36th Conf. Neural Inf. Process. Syst. Da-
2018.8519248. tasets Benchmarks Track, 2022. [Online]. Available: https://open
[25] S. A. Yamashkin, A. A. Yamashkin, V. V. Zanozin, M. M. Rado- review.net/forum?id=DEigo9L8xZA
vanovic, and A. N. Barmin, “Improving the efficiency of deep [38] Z. Xiong, F. Zhang, Y. Wang, Y. Shi, and X. X. Zhu, “Earth-
learning methods in remote sensing data analysis: Geosystem Nets: Empowering A I in Ear th obser vation,” 2022,
approach,” IEEE Access, vol. 8, pp. 179,516–179,529, Sep. 2020, arXiv:2210.04936.
doi: 10.1109/ACCESS.2020.3028030. [39] A. Paszke et al., “PyTorch: An imperative style, high-perfor-
[26] C. Broni-Bediako, Y. Murata, L. H. Mormille, and M. Atsumi, mance deep learning library,” in Proc. Neural Inf. Process. Syst.,
“Searching for CNN architectures for remote sensing scene clas- 2019, vol. 32, pp. 8026–8037.
sification,” IEEE Trans. Geosci. Remote Sens., vol. 60, pp. 1–13, [40] M. Abadi et al., “TensorFlow: A system for large-scale machine
2022, doi: 10.1109/TGRS.2021.3097938. learning,” in Proc. USENIX Symp. Oper. Syst. Des. Implementa-
[27] G. Sumbul et al., “BigEarthNet-MM: A large-scale, multimod- tions, 2016, pp. 265–283.
al, multilabel benchmark archive for remote sensing image [41] A. J. Stewart, C. Robinson, I. A. Corley, A. Ortiz, J. M. L. Ferres,
classification and retrieval [Software and Data Sets],” IEEE Geosci. and A. Banerjee, “TorchGeo: Deep learning with geospatial
Remote Sens. Mag., vol. 9, no. 3, pp. 174–180, Sep. 2021, doi: data,” 2021, arXiv:2111.08872.
10.1109/MGRS.2021.3089174. [42] H. Tamiminia, B. Salehi, M. Mahdianpari, L. Quackenbush,
[28] U. Chaudhuri, S. Dey, M. Datcu, B. Banerjee, and A. Bhattacha- S. Adeli, and B. Brisco, “Google earth engine for geo-big data
rya, “Interband retrieval and classification using the multila- applications: A meta-analysis and systematic review,” ISPRS J.
beled sentinel-2 BigEarthNet archive,” IEEE J. Sel. Topics Appl. Photogrammetry Remote Sens., vol. 164, pp. 152–170, Jun. 2020,
Earth Observ. Remote Sens., vol. 14, pp. 9884–9898, Sep. 2021, doi: 10.1016/j.isprsjprs.2020.04.001.
doi: 10.1109/JSTARS.2021.3112209. [43] A. Van Etten, D. Lindenbaum, and T. M. Bacastow, “SpaceNet:
[29] G. Cheng and J. Han, “A survey on object detection in optical re- A remote sensing dataset and challenge series,” 2018, arX-
mote sensing images,” ISPRS J. Photogrammetry Remote Sens., vol. iv:1807.01232.
117, pp. 11–28, Jul. 2016, doi: 10.1016/j.isprsjprs.2016.03.014. [44] E. Strickland, “Andrew Ng, AI minimalist: The machine-learn-
[Online]. Available: https://www.sciencedirect.com/science/ ing pioneer says small is the new big,” IEEE Spectr., vol. 59, no. 4,
article/pii/S0924271616300144 pp. 22–50, Apr. 2022, doi: 10.1109/MSPEC.2022.9754503.

96 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


[45] M. D. Wilkinson et al., “The fair guiding principles for scientific [66] “The earth in a cube.” Euro Data Cube. Accessed: Jul. 23, 2023.
data management and stewardship,” Scientific Data, vol. 3, no. 1, [Online]. Available: https://eurodatacube.com
Mar. 2016, Art. no. 160018, doi: 10.1038/sdata.2016.18. [67] “Your gateway to NASA earth observation data,” Earth Data,
[46] J. L. Dwyer, D. P. Roy, B. Sauer, C. B. Jenkerson, H. K. Zhang, Nat. Aeronaut. Space Admin., Washington, DC, USA. Accessed:
and L. Lymburner, “Analysis ready data: Enabling analysis of Jul. 23, 2023. [Online]. Available: https://www.earthdata.
the Landsat archive,” Remote Sens., vol. 10, no. 9, Aug. 2018, Art. nasa.gov
no. 1363, doi: 10.3390/rs10091363. [Online]. Available: https:// [68] “OpenAerialMap.” Accessed: Jul. 23, 2023. [Online]. Available:
www.mdpi.com/2072-4292/10/9/1363 https://openaerialmap.org
[47] G. Giuliani et al., “Building an earth observations data cube: [69] “OpenStreetMap.” Accessed: Jul. 23, 2023. [Online]. Available:
Lessons learned from the Swiss data cube (SDC) on generating https://openstreetmap.org
analysis ready data (ARD),” Big Earth Data, vol. 1, nos. 1–2, pp. [70] “EarthNets for earth observation.” EarthNets. Accessed: Jul. 23,
100–117, Sep. 2017, doi: 10.1080/20964471.2017.1398903. 2023. [Online]. Available: https://earthnets.nicepage.io
[48] T. Gebru et al., “Datasheets for datasets,” 2018. [Online]. Avail- [71] “Earth observation database.” Accessed: Jul. 23, 2023. [Online].
able: https://arxiv.org/abs/1803.09010 Available: https://eod-grss-ieee.com
[49] “SEN12MS toolbox.” GitHub. Accessed: Jul. 23, 2023. [Online]. [72] “Torchgeo.” GitHub. Accessed: Jul. 23, 2023. [Online]. Avail-
Available: https://github.com/schmitt-muc/SEN12MS able: https://github.com/microsoft/torchgeo
[50] “EuroSAT: Land use and land cover classification with senti- [73] “Azavea/raster-vision.” GitHub. Accessed: Jul. 23, 2023. [On-
nel-2.” GitHub Accessed: Jul. 23, 2023. [Online]. Available: line]. Available: https://github.com/azavea/raster-vision
https://github.com/phelber/EuroSAT [74] “Raster vision.” Accessed: Jul. 23, 2023. [Online]. Available:
[51] “Accurate and scalable processing of big data in earth observa- https://docs.rastervision.io/
tion.” BigEarth. Accessed: Jul. 23, 2023. [Online]. Available: [75] “Keras spatial.” Github. Accessed: Jul. 23, 2023. [Online]. Available:
https://bigearth.eu/ https://github.com/IllinoisStateGeologicalSurvey/keras-spatial
[52] “Functional map of the world (fMoW) dataset.” GitHub. Ac- [76] “A planetary-scale platform for Earth science data & analysis.”
cessed: Jul. 23, 2023. [Online]. Available: https://github.com/ Google Earth Engine. Accessed: Jul. 23, 2023. [Online]. Avail-
fMoW/dataset able: https://earthengine.google.com
[53] “xView3: Dark vessels.” Accessed: Jul. 23, 2023. [Online]. Avail- [77] “NAIP: National agriculture imagery program.” Earth Engine
able: https://iuu.xview.us/ Data Catalog. Accessed: Jul. 23, 2023. [Online]. Available:
[54] “A large-scale benchmark and challenges for object detection in https://developers.google.com/earth-engine/datasets/catalog/
aerial images.” DOTA. Accessed: Jul. 23, 2023. [Online]. Avail- USDA_NAIP_DOQQ
able: https://captain-whu.github.io/DOTA/ [78] “Start building on AWS today.” AWS. Accessed: Jul. 23, 2023.
[55] “LEVIR-CD.” Accessed: Jul. 23, 2023. [Online]. Available: [Online]. Available: https://aws.amazon.com
https://justchenhao.github.io/LEVIR/ [79] “Welcome to Digital Earth Africa (DE Africa).” Digital Earth Af-
[56] “Onera satellite change detection dataset.” GitHub. Accessed: Jul. rica. Accessed: Jul. 23, 2023. [Online]. Available: https://www.
23, 2023. [Online]. Available: https://rcdaudt.github.io/oscd/ digitalearthafrica.org
[57] “Data,” European Space Agency, Paris, France, Jun. 2019. [On- [80] “NOAA’s emergency response imagery,” National Ocean Ser-
line]. Available: https://kelvins.esa.int/proba-v-super-resolution/ vice, Silver Spring, MD, USA, 2023. Accessed: Jul. 23, 2023.
data/ [Online]. Available: https://oceanservice.noaa.gov/hazards/
[58] “PAirMax-Airbus.” Accessed: Jul. 23, 2023. [Online]. Available: emergency-response-imagery.html
https://perscido.univ-grenoble-alpes.fr/datasets/DS353 [81] “A planetary computer for a sustainable future.” Planetary
[59] “The WorldStrat dataset.” WorldStrat. Accessed: Jul. 23, 2023. Computer. [Online]. Available: https://planetarycomputer.
[Online]. Available: https://worldstrat.github.io/ microsoft.com
[60] “Copernicus open access hub.” Copernicus. Accessed: Jul. 23, [82] “High resolution electricity access.” Planetary Computer. Ac-
2023. [Online]. Available: https://scihub.copernicus.eu/ cessed: Jul. 23, 2023. [Online]. Available: https://planetary
[61] “Open data program.” Maxar. Accessed: Jul. 23, 2023. [Online]. computer.microsoft.com/dataset/hrea
Available: https://www.maxar.com/open-data [83] “USBuildingFootprints.” GitHub. [Online]. Available: https://
[62] “Capella space synthetic aperture radar (SAR) open dataset.” github.com/Microsoft/USBuildingFootprints
AWS. Accessed: Jul. 23, 2023. [Online]. Available: https://regis- [84] “Welcome to Colab!” Colab. Accessed: Jul. 23, 2023. [Online].
try.opendata.aws/capella_opendata/ Available: https://colab.research.google.com
[63] “TerraSAR-X science service system.” TerraSAR. Accessed: Jul. [85] Accessed: Jul. 23, 2023. [Online]. Available: https://www.grss
23, 2023. [Online]. Available: https://sss.terrasar-x.dlr.de/ -ieee.org/wp-content/uploads/2023/05/EODatasets.pdf
[64] IEEE DataPort. Accessed: Jul. 23, 2023. [Online]. Available: [86] “Start with more than a blinking cursor.” Kaggle. [Online].
https://ieee-dataport.org/datasets Available: https://www.kaggle.com
[65] “Open library for earth observations machine learning,” Radi- [87] “EDC browser.” Accessed: Jul. 23, 2023. [Online]. Available:
ant Earth Foundation, Washington, DC, USA. Accessed: Jul. 23, https://browser.eurodatacube.com/
2023. [Online]. Available: https://mlhub.earth GRS

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 97


SOFTWARE AND DATA SETS
YI WANG , NASSIM AIT ALI BRAHAM, ZHITONG XIONG ,
CHENYING LIU , CONRAD M. ALBRECHT, AND XIAO XIANG ZHU

SSL4EO-S12
A large-scale multimodal, multitemporal dataset for self-supervised learning in Earth observation

S elf-supervised pretraining bears the potential to


generate expressive representations from large-scale
Earth observation (EO) data without human annota-
information. Therefore, big EO-specific datasets for un-
supervised pretraining are necessary to be developed.
In this work, we introduce a large-scale, globally
tion. However, most existing pretraining in the field is distributed, multitemporal and multisensor dataset,
based on ImageNet or medium-sized, labeled remote SSL4EO-S12. The dataset samples 251,079 locations
sensing (RS) datasets. In this article, we share an unla- around the globe, each providing Sentinel-2 level-1C
beled dataset Self-Supervised Learning for Earth Observa- (L1C), Sentinel-2 level-2A (L2A), and Sentinel-1 ground
tion-Sentinel-1/2 (SSL4EO-S12) to assemble a large-scale, range detected (GRD) images with four snapshots from
global, multimodal, and multiseasonal corpus of satel- different seasons (in total: 3 million 2, 640-m # 2, 640-m
lite imagery. We demonstrate SSL4EO-S12 to succeed patches). Additionally, we guarantee optimal geospa-
in self-supervised pretraining for a set of representative tial coverage by avoiding the overlap of the randomly
methods: momentum contrast (MoCo), self-distillation sampled locations. This renders SSL4EO-S12 the largest
with no labels (DINO), masked autoencoders (MAE), and most generic multispectral/synthetic aperture radar
and data2vec, and multiple downstream applications, (SAR) dataset in the RS literature [5].
including scene classification, semantic segmentation, We demonstrate the potential of the SSL4EO-S12 data-
and change detection. Our benchmark results prove the set through a series of extensive experiments. Specifically,
effectiveness of SSL4EO-S12 compared to existing da- we evaluate four representative SSL algorithms—namely:
tasets. The dataset, related source code, and pretrained MoCo [6], DINO [7], MAE [8], and data2vec [9]—on
models are available at https://github.com/zhu-xlab/ three different downstream tasks: scene classification, se-
SSL4EO-S12. mantic segmentation, and change detection. Our results
indicate that pretraining on SSL4EO-S12 improves the
INTRODUCTION downstream performance compared to existing datasets.
Self-supervised learning (SSL) has attracted wide atten- Moreover, our ablation studies prove the benefits of RS-
tion in the RS community with the ability to learn ge- specific data augmentations, including multisensor, mul-
neric representations from unlabeled data. Numerous titemporal, and atmospheric correction.
studies in the literature have proven the potential of SSL
in EO beyond natural images [1]. Despite the focus SSL RELATED WORK
for EO receives, only limited effort is dedicated to pro-
viding large-scale datasets and benchmarks for pretrain- SELF-SUPERVISED LEARNING
ing. On the one hand, relying on computer vision datas- Over the past years, SSL has reached important mile-
ets like ImageNet [2] is not a preferred option due to the stones in computer vision, especially through contras-
domain gap. On the other hand, while RS datasets like tive methods with joint-embedding architectures. These
SEN12MS [3] or seasonal contrast (SeCo) [4] exist, they methods get trained to promote similarity between
are limited by geospatial overlap, sparse geographical augmented views of the same input, thereby enforcing
distribution, or lack diversity in seasonal or multimodal invariance to data augmentation. Several families of
such methods emerge: 1) contrasting negative samples
Digital Object Identifier 10.1109/MGRS.2023.3281651
for which the representations are encouraged to be dis-
Date of current version: 19 September 2023 similar [6]; 2) knowledge distillation between an asymmetric

98 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


teacher–student network [7]; 3) redundancy reduction Consequently, there is a need for large-scale pretraining
among the embedding dimensions [26]; and 4) clustering datasets in RS. Two datasets closely related to our efforts are
latent features to common prototypes from different views SEN12MS [3] and SeCo [4]. However, SEN12MS is limited
[10]. Meanwhile, recent developments in masked image by temporal coverage, SeCo has only optical data, and both
modeling reveal promising results in generative methods, datasets contain strongly overlapping patches that limit the
which reconstruct the masked input at pixel- [8] or feature- geospatial coverage. With the above in mind, our proposed
[9] level. SSL4EO-S12 dataset provides an improved spatiotemporal
We benchmark four representative methods—MoCo coverage by sampling more locations and removing over-
[6], DINO [7], MAE [8], and data2vec [9]—on the pro- lapping patches, enclosing multiple seasons, and including
posed dataset. This way, we cover a reasonably diverse set Sentinel-1 as well as two Sentinel-2 products (Table 1).
of representative methods from different categories: MoCo
contrasts negative samples, DINO represents a distillation SSL4EO-S12 DATASET
method, MAE is based on masked reconstruction, and
data2vec combines the masking mechanism with a joint- DATA CURATION AND ASSEMBLY
embedding architecture. The SSL4EO-S12 dataset (Figure 1) exploits openly avail-
able SAR/optical satellite data collected by the Euro-
PRETRAINING DATASETS pean Space Agency’s Sentinel mission. Following a well-
Pretrained models on ImageNet are widely used for various organized baseline provided by SeCo [4], we utilized
computer vision tasks. However, this is less appropriate in the Google Earth Engine [14] to download and process
the context of RS: 1) RS images are not object-centric, 2) the data. We filtered image patches to retrieve from the
there exist various types of sensors in RS, and 3) temporal 10,000 most populated cities (https://simplemaps.com/
effects yield variations on the ground surface. Therefore, data/world-cities) in the world to guarantee reasonable
EO-specific datasets are needed to provide the above in- global coverage. To obtain diverse land cover, we sampled
domain knowledge. The literature has proven the benefits 251,079 locations close by the cities following a Gaussian
of pretraining on existing labeled RS datasets [11], [12], yet distribution peaking at the city center and standard devia-
there are limitations, such as class bias and temporal and tion of 50 km, assuming most of the variability cast to the
geographical coverage. downtown and suburbs of cities [4]. At each location, we

TABLE 1. SUMMARY OF POPULAR MEDIUM-RESOLUTION PRETRAINING DATASETS IN RS.

SPATIAL TEMPORAL NO. OF NO. OF


DATASET COVER COVER MODALITY OVERLAP PATCH SIZE LOCATIONS PATCHES
BigEarthNet [13] Europe One timestamp SAR/Optical No 120 # 120 590,326 1.2 million
SEN12MS [3] Global One timestamp SAR/optical/land cover Yes 256 # 256 180,662 541,986
SeCo [4] Global Five timestamps Optical Yes 264 # 264 ~200,000 1 million
SSL4EO-S12 Global Four timestamps SAR/optical*2 + No 264 # 264 251,079 3 million

Earth Observation Multiple Seasons Global Coverage Multiple Modalities

FIGURE 1. Sample images of SSL4EO-S12 dataset assembled.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 99


downloaded four images drawn from four annual seasons for existing patches within this grid cell and exclusively
to capture seasonal variation. We searched for Sentinel-2 calculated distances to those local patches. Assuming po-
tiles with a cloud coverage lower than 10%. We also filtered tential overlap of sampled patches from distinct grid cells
out most overlapping patches with an efficient grid search is statistically negligible, we significantly reduced com-
strategy. In total, we obtained about 1 million S1-GRD/ puting time compared to a global overlap search. Indeed,
S2-L1C/S2-L2A image triplets. for SSL4EO-S12 we recorded an overlap for approximate-
ly 3% tiles of densely populated Tokyo, Japan, 1.5% in
DATA IDENTIFICATION Chicago, IL, USA, and below 1% for locations such as Bei-
The collection of SSL4EO-S12 differs from SeCo mainly by jing, China, Munich, Germany, Kampala, Uganda, and
introducing overlap filtering and multiple sensors (in ital- Brasilia, Brazil.
ics below). The workflow is shown as follows:
1) Uniformly sample one city from top-10,000 popu- DATA CHARACTERISTICS AND VOLUME
lated cities. The presented SSL4EO-S12 dataset contains 251,079
2) Sample one location from a Gaussian distribution with globally distributed Sentinel-1 dual-pol SAR, Sentinel-2
a standard deviation of 50 km around the city center. top-of-atmosphere multispectral, and Sentinel-2 surface
3) Check if a 2,640-m # 2,640-m image patch centered around reflectance multispectral triplets over four seasonal time-
that location has significant overlap with previous patches. If stamps. As of summer 2022, SSL4EO-S12 constitutes the
not, continue to step 4, otherwise return to step 1. biggest geospatial–temporal, multimodal dataset in terms
4) For a 30-d interval around four reference dates (20 of medium-resolution PolSAR and multispectral imagery,
March, 21 June, 22 September, 21 December) in 2021 serving more than 3 million images. The total data volume
(additionally look for 2020 as a buffer), check if there equates to an uncompressed size of 251, 079 # 4 # [2 $ 4B +
exist Sentinel-2 tiles with less than 10% of cloud cover- (13 + 12) $ 2B] # 264 2 . 3.7TB.
age (for both L1C and L2A) and corresponding Sentinel-1 Figure 2 depicts the geospatial distribution of the
GRD tiles. SSL4EO-S12 dataset, highlighting the dense coverage
5) If there exist valid Sentinel-1/2 tiles close to all the four across the globe. Figure 3 depicts the effect of overlap filter-
dates, process and download them into curated image ing around the Tokyo area.
patches, otherwise return to step 1.
EXPERIMENTAL SETUP
OVERLAP FILTERING We evaluated the SSL4EO-S12 dataset by self-supervised
A simple way to check significant overlap between two pretraining and transfer learning on RS downstream tasks.
patches is to calculate the distance between the two cen- Specific implementation details and additional results are
ters. If the distance is smaller than three-quarters of the provided in the supplemental material (available at https://
width of a patch, there is a nonnegligible overlap (>25%). doi.org/10.1109/MGRS.2023.3281651).
Naively, we needed to execute this computation for ev-
ery new patch relative to all existing patches. However, SELF-SUPERVISED PRETRAINING
this becomes inefficient when the number of patches We performed pretraining using four representative
grows large, 250,000+ for us. Therefore, we employed a SSL methods: MoCo-v2/v3 [15], [16], DINO [7], MAE
grid search strategy to perform efficient overlap filter- [8], and data2vec [9]. We pretrained ResNet [17] back-
ing. Instead of calculating the distance to all previous bones with MoCo(-v2) and DINO, and Vision Trans-
patches, we distributed the patch center coordinates into former (ViT) [18] backbones for all four SSL methods
360 # 180 geographical longitude–latitude, 1c # 1c grids. listed above. Unless explicitly noted, Sentinel-2 L1C
For each new patch, we converted the center coordinates was used for pretraining. To utilize multitemporal in-
into integer grid coordinates. Subsequently, we searched formation, we used RandomSeasonContrast as a data
augmentation strategy. For MoCo and DINO, the input
views were randomly picked from two seasons. For
MAE and data2vec, one random season was assigned
for each patch.
Pretraining one ResNet/ViT model for 100 epochs takes
7–25 h on four NVIDIA A100 GPUs, as shown in Table 2.

TRANSFER LEARNING
The pretrained models were transferred to various down-
stream tasks. For scene classification, we evaluated EuroSAT
[19] (single-label land cover classification), BigEarthNet [13]
(multilabel land cover classification), and So2Sat-LCZ42
FIGURE 2. Geographical distribution of SSL4EO-S12 dataset. [20] (local climate zone classification, culture-10 ­version).

100 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


For semantic segmentation, we included the 2020 Data COMPARISON OF PRETRAINING DATASETS
Fusion Contest (DFC2020) [21] (land cover segmenta- To compare SSL 4EO-S12 with other RS pretraining
tion) and Onera Satellite Change Detection (OSCD) [22] datasets, we report corresponding linear probing re-
(change detection). sults pretrained with MoCo-v2 (ResNet50 backbone)
We performed commonly used linear probing (freez-
ing the pretrained encoder) and fine-tuning for the down-
stream tasks. The results are reported in percentage scores.

BENCHMARK RESULTS

CLASSIFICATION

COMPARISON OF SELF-SUPERVISED
LEARNING METHODS
We first benchmarked different SSL methods through
linear probing on EuroSAT, BigEarthNet, and So2Sat-
LCZ42. As detailed in Table 3, all methods outperformed
(a)
random initialization by a substantial margin. As ex-
pected, linear probing on BigEarthNet with all labels
performs worse than fully supervised training. Prom-
isingly, the gap stays below 5%. On small datasets like
BigEarthNet with 10% labels or EuroSAT, linear probing
provides results comparable to supervised training with-
in approximately ±1%. The trends are slightly different
for So2Sat-LCZ42, where the training and testing sets are
built upon different cities with a challenging geographi-
cal split. Because of this significant domain shift, adding
labeled training data does not necessarily improve the
testing performance. In fact, fitting the training data dis-
tribution does not guarantee out-of-distribution gener-
(b)
alization. Nevertheless, the best pretrained models with
linear probing beat the supervised baseline by at least 1% FIGURE 3. Image patches without (a) and with (b) overlap filter-
up to about 4%. ing in Tokyo metropolitan area. We plotted red circles of radius
Furthermore, we benchmarked fine-tuning results in 1.32 km (132 pixels) for better visualization.
Table 4. All self-supervised methods outperform super-
vised learning with a margin from 1% to 6%. Top SSL
models score 99.1% on EuroSAT (MoCo/DINO) and over TABLE 2. 100 EPOCH TRAINING TIME OF THE STUDIED
SSL METHODS.
90% on BigEarthNet (MoCo/DINO). Comparing linear
probing and fine-tuning results, one interesting phe- MOCO DINO MAE DATA2VEC
nomenon shows up: in linear probing, contrastive meth- ResNet50 18 h 25 h — —
ods (MoCo and DINO) consistently score better than ViT-S/16 24 h 25 h 7h 14 h
their image-masking (MAE and data2vec) counterparts.

TABLE 3. LINEAR PROBING RESULTS FOR EUROSAT, BIGEARTHNET (BE), AND SO2SAT-LCZ42 (10% AND 100% LABELS).

DOWNSTREAM DATASET EUROSAT BE-10% BE-100% SO2SAT-10% SO2SAT-100%


MODEL\BACKBONE RN50 VIT-S/16 RN50 VIT-S/16 RN50 VIT-S/16 RN50 VIT-S/16 RN50 VIT-S/16
Random initialization 82.0 81.3 63.6 62.3 70.1 70.2 48.8 49.3 49.0 50.2
Supervised 98.0 96.7 83.4 81.3 88.7 87.4 57.5 59.7 57.5 59.3
MoCo 98.0 97.7 82.1 82.3 84.2 83.1 61.3 59.6 61.8 62.2
DINO 97.2 97.7 82.0 81.7 83.9 83.4 55.5 60.9 57.0 62.5
MAE — 94.1 — 77.5 — 78.2 — 59.5 — 60.0
Data2vec — 96.9 — 77.3 — 79.4 — 58.2 — 59.7
We report overall accuracy for EuroSAT and So2Sat-LCZ42, and mean average precision (micro) for BigEarthNet. Two backbone networks get trained: ResNet-50 (RN50) and a small(-S)
embedding dimension ViT subdividing input patches into 16 × 16 tiles(/16). Bold values indicate best per column performance.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 101


TABLE 4. FINE-TUNING RESULTS FOR EUROSAT, BIGEARTHNET, AND SO2SAT-LCZ42.

DOWNSTREAM DATASET EUROSAT BE-10% BE-100% SO2SAT-10% SO2SAT-100%


MODEL\BACKBONE RN50 VIT-S/16 RN50 VIT-S/16 RN50 VIT-S/16 RN50 VIT-S/16 RN50 VIT-S/16
MoCo 99.1 98.6 86.2 86.1 91.8 89.9 60.4 61.2 60.9 61.6
DINO 99.1 99.0 87.1 86.9 90.7 90.5 63.2 61.5 63.6 62.2
MAE — 98.7 — 84.8 — 88.9 — 60.8 — 63.9
Data2vec — 99.1 — 85.6 — 90.3 — 63.2 — 64.8
All beat supervised training; compare with Table 3. Bold values indicate best per column performance.

in Table 5. Similar to SSL4EO-S12, RandomSeasonCon-


trast is used to pick one timestamp image for each 90
geospatial patch in the SeCo dataset. In the first set RN50

Mean Average Precision (%)


of comparison, we used red/green/blue (RGB) bands 85
only. SSL4EO-S12 significantly outperforms ImageNet
by about 10%, SeCo by about 6%, and SEN12MS by 80
1.7% to 3.5%.
In a second set of experiments we evaluated all mul- 75
tispectral bands. Results indicate consistent perfor-
mance gain as in RGB setting comparing SSL4EO-S12 70
with SEN12MS and SeCo. In addition, pretraining on
SSL4EO-S12 outperforms BigEarthNet on itself and 65
EuroSAT (both are European Union only). This proves 1 10 50 100
Labeled Data (%)
SSL4EO-S12’s benefits to improve model transferability
(a)
by learning valuable knowledge from a larger scale and
wider geographical coverage. 90
ViT-S/16
Mean Average Precision (%)

COMPARISON OF DIFFERENT AMOUNTS OF LABELS 85


Figure 4 visualizes performance results of transfer learn-
ing on BigEarthNet with a varying fraction of labeled 80
samples. Compared to the supervised baseline, self-su-
pervised pretraining on SSL4EO-S12 provides significant 75
benefits when the amount of labeled samples is limited.
In fact, fine-tuning on 10% of the labels outperforms 70
50%-labels supervised training; and with ViT-S/16, fine-
tuning on 50% of the labels outperforms 100%-labels 65
1 10 50 100
supervised training. Labeled Data (%)
(b)

TABLE 5. COMPARISON OF DIFFERENT PRETRAINING Linear Fine-Tune Supervise


DATASETS UNDER LINEAR PROBING EVALUATION.

PRETRAIN DATASET EUROSAT BE-10% BE-100% BE Label Percentage 1% 10% 50% 100%
ImageNet (RGB) [4] 86.4 70.5 71.8 Linear 75.9 82.1 82.7 84.2
ViT-S/16 RN50

SeCo (RGB) [4] 89.5 74.5 76.3 Fine-Tune 80.3 86.2 87.7 91.8
SEN12MS (RGB) 94.9 76.6 79.6 Supervise 75.7 83.4 85.2 88.7
SSL4EO-S12 (RGB) 96.6 80.1 82.3 Linear 78.2 82.3 83 83.1
SeCo* (all bands) 89.2 73.7 76.6 Fine-Tune 78.9 86.1 88.2 89.9
SEN12MS (all bands) 95.5 79.6 82.1 Supervise 69.3 81.3 84.9 87.4
BigEarthNet (all bands) 94.4 80.6 83.9
SSL4EO-S12 (all bands) 98.0 82.1 84.2
FIGURE 4. BigEarthNet (BE) performance depending on the
Italics means cited from the literature. Bold values indicate best per column performance.
*SeCo is meant to have 200,000 geographical patches as described in the article, but the number of labels available to train downstream task. We report
available data at https://github.com/ServiceNow/seasonal-contrast has only about linear probing and fine-tuning results with ResNet50 and ViT-S/16
160,000 patches, which may affect our reproduced performance.
encoders pretrained using MoCo-v2. Bold values indicate best per
column performance.

102 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


SEGMENTATION ABLATION OF SEASONAL INFORMATION
We evaluated the effectiveness of multitemporal informa-
LAND COVER SEGMENTATION tion by replacing seasonal augmentation (compare with
We used the DFC2020 [21] dataset to evaluate land cover “Experimental Setup” section) by random season: the same
semantic segmentation. We pretrained ResNet50 with randomly selected season for the two positive views; and
MoCo-v2 on SSL4EO-S12 L1C products, and fine-tuned fixed season: the same season for each patch during train-
a DeepLabv3+ [23] for segmentation. Table 6 lists results ing. We pretrained on a 50,000 subset of SSL4EO-S12,
with notable improvements when compared to SeCo and evaluated on BigEarthNet-10% and EuroSAT. Table 9
pretraining. However, SSL4EO-S12 performs worse than clearly proves the benefits of seasonal augmentation.
SEN12MS in average accuracy and mean intersection
over union. This can be expected, since DFC2020 was ATMOSPHERIC CORRECTION AS DATA AUGMENTATION
built with direct reference to SEN12MS and they have The motivation to include Sentinel-2 L1C and L2A
similar data distribution. Nevertheless, the results are products in SSL4EO-S12 is to match corresponding
still comparable, proving again the transferability of the
proposed dataset.
TABLE 6. DFC2020 LAND COVER SEGMENTATION RESULTS.
CHANGE DETECTION
PRETRAIN DATASET OA AA MIOU
We evaluated the pretrained models for change detec- Random initialization 81.97 56.46 42.11
tion on the OSCD [22] dataset. We pretrained ResNet50 SeCo 87.31 57.05 49.68
with MoCo-v2 on SSL4EO-S12 L1C products, froze the SEN12MS 88.64 67.69 54.83
backbone, and fine-tuned a U-Net [24] for segmenta- SSLEO-S12 89.58 64.01 54.68
tion. The differences in feature maps between two time- Bold values indicate best per column performance. AA: average accuracy; mIoU: mean
stamps were input to the network. As Table 7 indicates, intersection over union; OA: overall accuracy.
pretraining on SSL4EO-S12 yields superior performance
in recall and F1-score when referenced to SeCo and
SEN12MS. While SSL4EO-S12 performs worse in preci-
sion, this is due to the significant class unbalance that TABLE 7. OSCD CHANGE DETECTION RESULTS.
predicting all pixels as unchanged would result in a
PRETRAIN DATASET PRECISION RECALL F1
good precision score.
Random initialization 72.31 13.75 23.10
SeCo 74.85 17.47 28.33
ADDITIONAL STUDIES
SEN12MS 74.67 19.26 30.62
We complete our benchmark by reporting a set of addition-
SSL4EO-S12 70.23 23.38 35.08
al results to document key characteristics of the SSL4EO-
Bold values indicate best per column performance.
S12 dataset, namely: multitemporal, multimodal, multi-
product-level, and data scale. For all studies, we pretrained
ResNet50 with MoCo-v2 as a common setting.

TABLE 8. LINEAR PROBING RESULTS OF MULTIMODAL SSL.


ABLATION STUDIES
BE-1% BE-100%
BENEFITS OF MULTIMODALITY S1 S2 S1+S2 S1 S2 S1+S2
While the “Benchmark Results” section employs only MoCo-S1/2 71.1 75.9 – 75.9 84.2 –
optical data for fair comparison to existing literature, we MoCo-MM 73.3 76.7 76.8 79.5 85.1 85.2
highlight the benefits of multimodal pretraining in this Supervised 66.7 75.7 76.4 77.2 88.7 88.9
section. We integrate SAR data by early fusion, and use MoCo-S1/2 represents pretraining with one single modality, and MoCo-MM represents pre-
training with both modalities. Bold values indicate best per column performance.
RandomSensorDrop [12] as an additional data augmen-
tation strategy. During training, the model gets fed ran-
dom combinations of SAR/optical patches, thus learning
both inner- and intermodality representations. Then, the
pretrained model gets transferred to different scenarios, TABLE 9. LINEAR PROBING RESULTS OF MULTITEMPORAL
ABLATION STUDY.
where either both modalities or a single one is available.
We compare multimodal pretraining (MM) to unimodal PRETRAIN SEASON BE-10% EUROSAT
pretraining (S1/2) on BigEarthNet. Table 8 presents results Fixed 75.1 93.1
with notable improvement of 1–3% for 100% and 1% la- Random 76.7 94.0

bel splits. While single-modality pretraining already works Augment 77.6 96.2

well for both Sentinel-2 and Sentinel-1 data, pretraining Bold values indicate best per column performance.

exploiting both modalities further improves performance.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 103


­ ownstream tasks. However, these product levels with
d a notable improvement of up to 4% compared to pretrain-
or without atmospheric correction can also be consid- ing on a single modality.
ered natural data augmentation for SSL. Accordingly, we
conducted an ablation study on a 50,000 SSL4EO-S12 IMPACT OF PRETRAINING SCALE
subset utilizing Sentinel-2 L1C, L2A, or both (L1C+L2A). An aspect relevant to large-scale data mining in EO is scal-
Table 10 summarizes our findings: 1) models pretrained ing of results with training data volume: why don’t we add
on the same product level as the downstream task have a more images to SSL4EO-S12? One reason concerns com-
slight edge (~ 1%) over models trained on the other prod- putational costs. We believe the current dataset (1 million
uct level, and 2) pretraining on both modalities generates patches for each Sentinel product) is comparable to the scale
of ImageNet, and can serve as a good baseline in RS for
further development. Moreover, as observed by [25], satu-
TABLE 10. LINEAR PROBING RESULTS OF DIFFERENT rating downstream performance kicks in beyond 500,000
PRODUCT LEVELS OF SENTINEL-2. pretraining images on ImageNet, with 250,000 images
yielding acceptable results with as little as 1–2% accuracy
PRETRAIN PRODUCT BE-10% (L2A) EUROSAT (L1C)
L1C 74.0 93.1
loss. We observe such a trend in our dataset, too. As dem-
L2A 75.1 92.0
onstrated by Table 11, we pretrained on various amounts of
L1C+L2A 78.0 93.8 data to report linear probing results for BigEarthNet-10%.
Bold values indicate best per column performance. While 50% (500,000) or less pretraining data yields sig-
nificant performance drops, there’s fewer diminishing
gaps from 75% (750,000) on. Note, this saturation effect
depends also on the model size.

TABLE 11. LINEAR PROBING RESULTS ON BIGEARTHNET-10%


FOR VARIOUS SENTINEL-2 L1C PRETRAINING DATA SIZES. REPRESENTATION VISUALIZATION
We qualitatively evaluated the data representations learned
PRETRAIN DATA 100,000 250,000 500,000 750,000 1 MILLION from self-supervised pretraining by visualizing the latent
Accuracy (%) 64 73 78 81 82
distributions with t-distributed stochastic neighbor em-
bedding (t-SNE) (Figure 5). We pretrained a ResNet50
with MoCo-v2 on SSL4EO-S12, and transferred the frozen
encoder to EuroSAT to calculate one 128-d representation
vector for each image. We then visualized all the vectors
with t-SNE, and compared the distribution with a random-
ly initialized encoder.

CONCLUSION
In this work, we present SSL4EO-S12, a large-scale mul-
timodal, multitemporal unlabeled dataset for SSL in EO.
An extensive benchmark on various SSL methods and
RS applications proves the promising benefits of the pro-
posed dataset.
(a) SSL4EO-S12 has some limitations: 1) there’s little cover-
age of polar regions, 2) geographical bias exists due to cloud
filtering, 3) it is not strictly free of geospatial overlap, and
4) medium-resolution radar and multispectral images are a
limited subset of EO data. Despite these limitations, we be-
lieve SSL4EO-S12 renders a valuable basis to advance self-
supervised pretraining and large-scale data mining in RS.

ACKNOWLEDGMENT
This work is jointly supported by the Helmholtz Association
through the Framework of Helmholtz Artificial Intelligence
(b) (Grant ZT-I-PF-5-01)–Local Unit “Munich Unit @Aeronau-
tics, Space and Transport (MASTr)” and Helmholtz Excel-
FIGURE 5. t-SNE visualization of EuroSAT image representations. lent Professorship “Data Science in Earth Observation–Big
One color represents one class. (a) Random-encoded features; Data Fusion for Urban Research”(Grant W2-W3-100); by the
(b) SSL-encoded features. SSL-encoded features are well clustered German Federal Ministry of Education and Research in the
even without label information. framework of the International Future Artificial I­ ntelligence

104 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Laboratory “AI4EO–Artificial Intelligence for Earth Ob- [6] K. He et al., “Momentum contrast for unsupervised visual
servation: Reasoning, Uncertainties, Ethics and Beyond” representation learning,” in Proc. IEEE/CVF Conf. Comput. Vi-
(Grant 01DD20001); and by the German Federal Ministry sion Pattern Recognit., 2020, pp. 9726–9735, doi: 10.1109/
for Economic Affairs and Climate Action in the framework CVPR42600.2020.00975.
of the “National Center of excellence ML4Earth” (Grant [7] M. Caron et al., “Emerging properties in self-supervised vision
50EE2201C). The computing resources were supported transformers,” in Proc. IEEE/CVF Int. Conf. Comput. Vision, 2021,
by the Helmholtz Association’s Initiative and Networking pp. 9650–9660.
Fund on the HAICORE@FZJ partition. This article has sup- [8] K. He et al., “Masked autoencoders are scalable vision learners,”
plementary downloadable material available at https://doi. 2021, arXiv:2111.06377.
org/10.1109/MGRS.2023.3281651. [9] A. Baevski et al., “Data2vec: A general framework for self-supervised
learning in speech, vision and language,” 2022, arXiv:2202.03555.
AUTHOR INFORMATION [10] M. Caron et al., “Unsupervised learning of visual features by
Yi Wang (yi.wang@dlr.de) is with the Chair of Data Science contrasting cluster assignments,” in Proc. Adv. Neural Inf. Process.
in Earth Observation, Technical University of Munich, Syst., vol. 33, 2020, pp. 9912–9924.
80333 Munich, Germany, and the Remote Sensing Technol- [11] M. Neumann et al., “Training general representations for re-
ogy Institute, German Aerospace Center, 82234 Weßling, mote sensing using in-domain knowledge,” in Proc. IEEE Int.
Germany. He is a Graduate Student Member of IEEE. Geosci. Remote Sens. Symp., 2020, pp. 6730–6733, doi: 10.1109/
Nassim Ait Ali Braham (nassim.aitalibraham@dlr.de) is IGARSS39084.2020.9324501.
with the Chair of Data Science in Earth Observation, Tech- [12] Y. Wang, C. M. Albrecht, and X. X. Zhu, “Self-supervised vision
nical University of Munich, 80333 Munich, Germany, and transformers for joint SAR-optical representation learning,”
the Remote Sensing Technology Institute, German Aero- 2022, arXiv:2204.05381.
space Center, 82234 Weßling, Germany. [13] G. Sumbul et al., “BigEarthNet: A large-scale benchmark ar-
Zhitong Xiong (zhitong.xiong@tum.de) is with the chive for remote sensing image understanding,” in Proc. IEEE Int.
Chair of Data Science in Earth Observation, Technical Geosci. Remote Sens. Symp., 2019, pp. 5901–5904, doi: 10.1109/
University of Munich, 80333 Munich, Germany. He is a IGARSS.2019.8900532.
Member of IEEE. [14] N. Gorelick et al., “Google Earth Engine: Planetary-scale geospa-
Chenying Liu (chenying.liu@dlr.de) is with the Chair of tial analysis for everyone,” Remote Sens. Environ., vol. 202, Dec.
Data Science in Earth Observation, Technical University of 2017, pp. 18–27, doi: 10.1016/j.rse.2017.06.031.
Munich, 80333 Munich, Germany, and the Remote Sens- [15] X. Chen et al., “Improved baselines with momentum contrastive
ing Technology Institute, German Aerospace Center, 82234 learning,” 2020, arXiv:2003.04297.
Weßling, Germany. She is a Graduate Student Member of IEEE. [16] X. Chen, S. Xie, and K. He, “An empirical study of training self-
Conrad M. Albrecht (conrad.albrecht@dlr.de) is with the supervised vision transformers,” in Proc. IEEE/CVF Int. Conf.
Remote Sensing Technology Institute, German Aerospace Comput. Vision, 2021, pp. 9640–9649.
Center, 82234 Weßling, Germany. [17] K. He et al., “Deep residual learning for image recognition,” in
Xiao Xiang Zhu (xiaoxiang.zhu@tum.de) is with the Proc. IEEE Conf. Comput. Vision Pattern Recognit., 2016, pp. 770–
Chair of Data Science in Earth Observation, Technical Uni- 778, doi: 10.1109/CVPR.2016.90.
versity of Munich, 80333 Munich, Germany. She is a Fellow [18] A. Dosovitskiy et al., “An image is worth 16x16 words: Trans-
of IEEE. formers for image recognition at scale,” 2020, arXiv:2010.11929.
[19] P. Helber et al., “EuroSAT: A novel dataset and deep learning
REFERENCES benchmark for land use and land cover classification,” IEEE J.
[1] Y. Wang et al., “Self-supervised learning in remote sensing: A re- Sel. Topics Appl. Earth Observ. Remote Sens., vol. 12, no. 7, pp.
view,” IEEE Geosci. Remote Sens. Mag., vol. 10, no. 4, pp. 213–247, 2217–2226, Jul. 2019, doi: 10.1109/JSTARS.2019.2918242.
Dec. 2022, doi: 10.1109/MGRS.2022.3198244. [20] X. X. Zhu et al., “So2Sat LCZ42: A benchmark data set for
[2] J. Deng et al., “ImageNet: A large-scale hierarchical image data- the classification of global local climate zones,” IEEE Geos-
base,” in Proc. IEEE Conf. Comput. Vision Pattern Recognit., 2009, ci. Remote Sens. Mag. (replaced Newsletter), vol. 8, no. 3, pp.
pp. 248–255, doi: 10.1109/CVPR.2009.5206848. 76–89, Sep. 2020, doi: 10.1109/MGRS.2020.2964708.
[3] M. Schmitt et al., “SEN12MS – A curated dataset of georeferenced [21] M. Schmitt et al., “IEEE GRSS data fusion contest,” 2020. [On-
multi-spectral Sentinel-1/2 imagery for deep learning and data line]. Available: https://ieee-dataport.org/competitions/2020
fusion,” ISPRS Ann. Photogrammetry, Remote Sens. Spatial Inf. Sci., -ieee-grss-data-fusion-contest
vol. IV-2/W7, pp. 153–160, Sep. 2019, doi: 10.5194/isprs-an- [22] R. C. Daudt et al., “Urban change detection for multispectral
nals-IV-2-W7-153-2019. earth observation using convolutional neural networks,” in Proc.
[4] O. Mañas et al., “Seasonal contrast: Unsupervised pre-training IEEE Int. Geosci. Remote Sens. Symp., 2018, pp. 2115–2118, doi:
from uncurated remote sensing data,” in Proc. IEEE/CVF Int. Conf. 10.1109/IGARSS.2018.8518015.
Comput. Vision, 2021, pp. 9414–9423. [23] L.-C. Chen et al., “Encoder-decoder with Atrous separable con-
[5] Z. Xiong et al., “EarthNets: Empowering AI in earth observa- volution for semantic image segmentation,” in Proc. Eur. Conf.
tion,” 2022, arXiv:2210.04936. Comput. Vision (ECCV), 2018, pp. 801–818.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 105


[24] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional net- [26] J. Zbontar, L. Jing, I. Misra, Y. LeCun, and S. Deny, “Barlow twins:
works for biomedical image segmentation,” in Proc. Int. Conf. Med. Self-supervised learning via redundancy reduction,” in Proc. Int.
Image Comput. Comput.-Assisted Intervention, 2015, pp. 234–241. Conf. Mach. Learn., 2021, pp. 12310–12320. [Online]. Available:
[25] E. Cole et al., “When does contrastive visual representation https://proceedings.mlr.press/v139/zbontar21a.html
learning work?” in Proc. IEEE/CVF Conf. Comput. Vision Pattern
Recognit., 2022, pp. 14,755–14,764.

DANIELE REGE CAMBRIN , LUCA COLOMBA,


AND PAOLO GARZA

CaBuAr
California burned areas dataset for delineation

F orest wildfires represent one of the catastrophic events


that, over the last decades, have caused huge environ-
mental and humanitarian damage. In addition to a signifi-
quired by Sentinel [1], Landsat [2], and MODIS [3] missions
as an example, it is possible to retrieve information at a con-
tinental scale in a short amount of time. This, in conjunc-
cant amount of carbon dioxide emission, they are a source tion with the development of modern methodologies in the
of risk to society in both short-term (e.g., temporary city field of machine learning and deep learning, represents an
evacuation due to fire) and long-term (e.g., higher risks of extremely interesting area of research for scientists and au-
landslides) cases. Consequently, the availability of tools thorities from different fields, such as governments and first
to support local authorities in automatically identifying responders involved in disaster response and disaster recov-
burned areas plays an important role in the continuous ery missions. Phenomena such as climate change and ex-
monitoring requirement to alleviate the aftereffects of such treme climate events have a tremendous societal, economic,
catastrophic events. The great availability of satellite acquisi- and environmental impact, also leading to humanitarian
tions coupled with computer vision techniques represents and environmental losses (e.g., a higher risk of landslides
an important step in developing such tools. due to a forest fire). Indeed, leveraging EO and modern deep
This article introduces a novel open dataset that tackles learning methodologies can provide useful tools in the area
the burned area delineation problem, a binary segmentation of disaster management and disaster recovery.
problem applied to satellite imagery. The presented resource Within the research community, numerous previous
consists of pre- and postfire Sentinel-2 L2A acquisitions of works proved the effectiveness of computer vision archi-
California forest fires that took place from 2015 to 2022. tectures in the field of disaster response, such as flood
Raster annotations were generated from the data released delineation [4], change detection [5], [6], and burned area
by California’s Department of Forestry and Fire Protection. delineation [7], [8], [9]. This article fits in the last men-
Moreover, in conjunction with the dataset, we release three tioned context. Specifically, we release a dataset to tackle
different baselines based on spectral index analyses, Seg- the burned area delineation problem, i.e., a binary image
Former, and U-Net models. segmentation problem that aims to identify areas damaged
by a forest wildfire. Tackling such a problem with modern
INTRODUCTION methodologies requires great data availability. However,
The Earth observation (EO) field has greatly increased the the time and cost needed to produce high-quality annota-
number of applications in the last decades thanks to the tions severely limit the ability to investigate ad hoc solu-
greater data availability, storage capacity, and computa- tions in the EO field. For these reasons, we propose a new
tional power of modern systems. In fact, leveraging data ac- dataset related to forest fires in California, collecting data
from the Sentinel-2 mission [10]. The dataset is publicly
Digital Object Identifier 10.1109/MGRS.2023.3292467
available to the research community at https://huggingface.
Date of current version: 19 September 2023 co/datasets/DarthReca/california_burned_areas. Compared to

106 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


the few other datasets about wildfires [11], [12], our dataset vegetation independent is difficult [24]. These methods
covers a larger area and spans more years. assume that burned and unburned areas are linearly sepa-
Ground truth masks for the task of binary image seg- rable, which is usually untrue.
mentation were generated starting from the public vector More recently, researchers started adopting supervised
data provided by California’s Department of Forestry and learning techniques to solve several tasks in computer vi-
Fire Protection [13] and rasterized. Satellite acquisitions, sion and EO. More precisely, convolutional neural network
i.e., the raw input data, were instead collected from the (CNN)-based models proved their effectiveness in image
Sentinel-2 L2A mission through Copernicus Open Ac- classification and segmentation tasks, achieving state-of-
cess Hub. More precisely, we collected and released both the-art performances compared to index-based methodolo-
prefire and postfire information associated with the same gies [25], [26]. Deep models proved their effectiveness in
area of interest. similar tasks covering wildfire detection [27] and spreading
The contributions of this article can be summarized as [28], too. The main drawback is the need for a significant
follows: amount of labeled data, pos-
◗◗ A novel image segmentation dataset was tailored to burned sibly covering heterogeneous
area delineation consisting of Sentinel-2 pre- and postfire regions with different mor-
IT IS WELL KNOWN THAT
acquisitions. We provide more samples than existing datas- phological characteristics, to
learn better representations. MORE CURATED DATA
ets to facilitate the training of (large) deep learning models.
◗◗ Three different baselines were evaluated on the pro- Over the years, many of the PROVIDE BETTER MACHINE
posed dataset: one consisting of the evaluation of sev- proposed frameworks have LEARNING MODELS, AND
eral burned area indexes and Otsu’s automatic thresh- limited their analyses to a few THE DATASET PROVIDES
olding method [14], one based on the SegFormer model samples collected from a lim- MANY READY-TO-USE
[15], and one based on the U-Net model [16]. ited number of countries or SAMPLES WITHOUT
The article is structured as follows. The “Related Works” locations [29]. In a few cases, LEVERAGING OTHER
section introduces the related works, the “Dataset” section larger datasets were adopted
SOURCES.
introduces the collected dataset and the preprocessing steps to tackle the semantic seg-
performed, and the “Tasks” and “Experiments” sections mentation problem without
formally introduce the tasks and the experimental settings disclosing the dataset [27].
and results. Finally, the final section concludes the article. In the EO domain, different public datasets are avail-
able to the research community tackling different prob-
RELATED WORKS lems, such as flood delineation [17], [30], deforestation
Before the development of deep learning-based method- [31], wild area monitoring [32], sustainable development
ologies, domain experts based their analyses on satellite goal monitoring [33], and crop classification and seg-
imagery leveraging spectral index computation and evalu- mentation [34], but, to the best of our knowledge, only
ation. Considering the synthetic aperture radar context, two public datasets are available for the burned area de-
thresholding-based techniques have been adopted to dis- lineation problem covering some countries in Europe
tinguish between flooded and unflooded areas [17]. Dif- [11] and Indonesia [12]. Our dataset collects more data
ferent analyses have been performed on various tasks con- than these, considering more wildfires and a larger area.
cerning several spectral indexes, such as in cloud detection It comprises pre- and postfire Sentinel-2 L2A data about
(cloud mask) [18], water presence (water pixels and the nor- California forest fires.
malized difference water index) [19], [20], and vegetation Table 1 shows a comparison among the three datas-
analysis (the normalized difference vegetation index) [21]. ets. The proposed dataset consists of the highest number
Considering the burned area delineation problem, do- of considered wildfires (340), globally covering the larg-
main experts have developed several indexes: the Normal- est amount of burned areas (28 million pixels covering
ized Burn Ratio (NBR), NBR2, Burn Area Index (BAI), and 11,000 km2) and a higher total covered surface (450,000 km2).
BAI for Sentinel-2 (BAIS2) [19]. They are computed using Figure 1(a) shows the covered areas. Even though the
different spectral bands to generate an index highlighting proposed dataset has the greatest amount of burned sur-
the affected areas of interest. Such techniques are often face, it achieves the lowest percentage of burned area
coupled with thresholding methodologies: either fixed or compared to the others. However, the CaBuAr dataset
manually calibrated threshold values are chosen [22], or au- provides the highest number of training samples in super-
tomatic thresholding algorithms are used [23]. Additional vised (supervised learning in binary segmentation and the
studies evaluate index-based techniques with additional highest amount of burned areas) and unsupervised cases
in situ information, namely, the Composite Burned Area (self-supervised learning and the highest area covered). It
Index, which, indeed, provides insightful information but is well known that more curated data provide better ma-
does not represent a scalable solution because in situ data chine learning models, and the dataset provides many
are incredibly costly to collect. Furthermore, studies con- ready-to-use samples without leveraging other sources.
firmed that finding a unique threshold that is region and Images are larger in terms of pixels (5,490) and disclosed

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 107


as raw data in the original and unaltered state, as direct- DATASET
ly collected from satellite instrumentation. On the other The newly created dataset comprises L2A products of Senti-
hand, the European dataset provides data collected from a nel-2, a European Space Agency mission. The area of inter-
third-party service for which preprocessing operations are est is California, with the geographical distributions of the
performed. The availability of raw data enables researchers events shown in Figure 1(b). We collected images of the same
to apply the preferred preprocessing steps without any loss area before and after the wildfire. It is essential to note that
of information. Furthermore, the monitored range of dates the L2A product contains red, green, blue (RGB) channels as
of the new dataset spans from 2015 to 2022, whereas the well as other spectral bands in the infrared region and ultra-
other two datasets span a smaller time period. blue for a total of 12 channels. Depending on the band, they
have a resolution of 10 m, 20 m, or 60 m per pixel.

TABLE 1. THE COMPARISON AMONG DATASETS. PREPROCESSING


FEATURE CaBuAr (OURS) [11] [12] The California Department of Forestry and Fire Protection
Region California Europe Indonesia publicly provides the ground truth vector information,
Mission Sentinel-2 Sentinel-1/2 Landsat-8 which we converted into raster images. Each pixel contains
Resolution (m) 20 10 (S2) 30 a binary value: one in the burned area and zero in the case
Image size 5,490 × 5,490 Up to 5,000 512 × 512 of undamaged areas. Although the registered wildfires span
× 5,000 from 1898 to 2022, we collected data only for wildfires from
Raw data ü û ü 2015 to 2022 because there were no Sentinel-2 images be-
Channels 12 12 8 fore 2015. We gathered the Sentinel-2 images directly from
Forest fires 340 73 227 Copernicus Open Access Hub.
Start and end dates January 2015 to July 2017 to January 2019 To minimize the effects of vegetation regrowth and
December 2022 July 2019 to December
2021 postwildfire modifications, images are collected within
Total surface (km2) ~450,000 ~19,000 ~46,000 one month after the wildfires are fully contained and ex-
Burned surface ~28/~11,000 ~20/~2,000 ~8/~7,000 tinguished. A total number of 340 acquisitions associated
[MP/km2] with 340 wildfires were downloaded, each being of size
Postfire ü ü ü 5,490 × 5,490 pixels with a resolution of 20 m per pixel. The
Prefire ü ü û few Sentinel-2 bands with different resolutions were either
TD ~1 year # 2 months / upsampled or downsampled with bicubic interpolation to
The highest value in each line is highlighted in bold, except for the resolution case, in reach the target resolution.
which the lowest numerical value is highlighted. MP: the number of burned pixels in mil-
lions; TD: the time difference between prefire and postfire acquisitions. Prefire images have the same size and resolution as the
postfire acquisitions. To enforce coherence and ­similar

California 42
42
Satellite Tiles

40
40
Latitude (°)
Latitude (°)

38 38

36 36

34 34

32
–124 –122 –120 –118 –116 –114 –124 –122 –120 –118 –116 –114
Longitude (°) Longitude (°)
(a) (b)
Map tiles by Stamen Design, CC BY 3.0 -- Map data (C) OpenStreetMap contributors

FIGURE 1. (a) The satellite tiles coverage: the California administrative boundaries (red) versus the satellite tiles of the proposed dataset
(blue). (b) The location of the wildfires (red) inside the California boundaries (blue).

108 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


seasonal and phenological conditions, we downloaded in supervised and unsupervised scenarios. Having at our
prefire data considering a temporal window of four weeks, disposal two sets of images, called PS and PR, containing
centered one year before the date postfire data were col- postfire samples and prefire ones, respectively, the tasks we
lected. For example, given a postfire acquisition collected considered in this article can be formulated as follows:
on 1 April 2018, we downloaded the products available 1) Binary segmentation through machine learning methods based
between 18 March 2017 and 15 April 2017, with the center on postfire acquisitions only: This involves a supervised
on 1 April 2017. This ensures similar climatic and seasonal learning algorithm (AS) to perform pixel-level predic-
conditions, thus limiting environmental changes as much tion based on samples from PS. AS labels pixels of im-
as possible. In some cases, retrieving these products was ages as burned or undamaged, creating a binary mask
impossible due to data unavailability; i.e., not all wildfires for each new image.
have a prefire acquisition satisfying such a constraint. 2) Binary segmentation through machine learning models based
Given the 340 wildfires considered in this study, 208 have on prefire and postfire acquisitions: This involves a super-
prefire availability satisfying the mentioned constraint. vised learning algorithm (AS) to perform pixel-level pre-
The dataset was randomly split into five nonoverlapped diction considering samples from PS and PR.
folds to perform cross validation. 3) Binary segmentation through spectral indexes: This involves
a spectral index (SI) designed for burned area identifi-
MANUAL INSPECTION cation. Taking samples from PS, SI outputs a value for
After collecting data, we manually evaluated each postfire
image using RGB channels. This was done to 1) discard in-
valid samples and 2) enrich the dataset with metadata and
comments based on our subjective evaluation. We remark
that such comments are not helpful for the final prediction
Postfire

task but can be used to better characterize the data. Our


evaluation is associated with each satellite acquisition.
Each image has a metadata field with a list of numeric
codes generated from the manual inspection. Figure 2 re-
ports the code-to-comment association. As can be seen,
different climatic conditions can be found in the dataset.
Prefire

Figure 3 reports some examples of postfire images. For each


postfire acquisition, Figure 3 reports its RGB version (first
line), its binary mask (second line), and the comment(s) as-
signed to it (on top of the RGB image). For instance, the
second acquisition has two comments: comments 2 and
11. We noted that some masks seem to overestimate the
Mask

burned area. However, our perception refers to the RGB


version of the images, i.e., to a subset of the available infor-
mation. Moreover, our subjective perception can be biased
also because the regions at the borders of burned areas are
usually less damaged than the central ones. These notes FIGURE 2. An example of prefire and postfire RGBs and relative masks.
must be extended to other mask-
related comments, but they are rarer.
Comments are almost equally dis- 9 2,11 5,11 6,13
tributed among the folds. All of the
340 acquisitions do not include any
Postfire

comments that can negatively affect


results and the dataset’s quality (i.e.,
comments 4, 8, and 12).
Finally, each prefire image was
manually inspected to verify its va-
lidity, but no new comment types
were added. All invalid prefire acqui-
Mask

sitions were discarded.

TASKS
The proposed dataset can be used
as a benchmark for different tasks FIGURE 3. A sample of postfire RGBs and masks with the associated comments.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 109


each pixel, creating a matrix PSl. Then, a binary mask, 3) Usage of postfire and prefire images: setting 3, tackling tasks
burned–unburned, can be made thresholding PSl. 2 and 4. Thus, two input images are considered for each
4) Binary segmentation through differential spectral indexes: area. Spectral indexes in this setting were evaluated by
This involves a spectral index (SI) designed for burned computing the difference between prefire and postfire
area identification based on the comparison of pre- and indexes.
postwildfire images. Taking samples from PS and PR, SI The code for the experiments can be found at https://github.
outputs a value for each pixel, creating a matrix PSl and com/DarthReca/CaBuAr.
PRl. Then, a binary mask, burned–unburned, can be
made thresholding the difference PSl - PRl. EXPERIMENTAL SETTINGS
The encoder of SegFormer is initialized with the original
EXPERIMENTS weights for Image-Net duplicated four times to handle
Our experiments test various classical threshold-based and the 12 available channels for Sentinel-2 L2A acquisitions.
deep learning-based methods considering three different U-Net is, instead, randomly initialized. The batch size was
data settings: set to eight. We used the AdamW optimizer with an ini-
1) Usage of all the available postfire images: setting 1, tackling tial learning rate of 0.001, decreased by a factor of 10 every
tasks 1 and 3. 15 epochs, and a weight decay of 0.01 for every consid-
2) Usage of the subset of postfire images for which the correspond- ered model. We used the well-known Dice loss [35] as the
ing prefire image is available, without using the prefire image loss function. All models were trained on one Tesla V100
to train the models: setting 2, tackling tasks 1 and 3. 32-GB GPU. The testing was made using the weights associ-
ated with the best validation loss.
Due to the size of the original input images (5,490
TABLE 2. THE ASSOCIATION BETWEEN CODES AND COMMENTS. × 5,490), we split them into patches of size 512 × 512.
COMMENT MEANING
­Furthermore, due to class imbalance, we kept only those
0 The affected area is in the incomplete region.
patches containing at least one pixel associated with the
1 The image is incomplete. positive class and no clouds over the area of interest (com-
2 There is a small burned area. ment 6 in Table 2). A total of 534 patches for setting 1 and
3 The mask has a small offset. 356 for ­settings 2 and 3 were obtained.
4 The mask is totally wrong. The statistics and performances reported in the re-
5 There is an extensive burned area. mainder of the article refer to the data obtained after the
6 There are clouds over the burned area. split-and-filter process mentioned earlier. All training and
7 There are too many clouds over the image. evaluation procedures were performed with a cross-valida-
8 The wildfire is ongoing. tion approach. The same criterion was applied for spectral
9 There is snow on the burned area. index methodologies to obtain comparable results despite
10 The mask seems smaller than the burned area. the absence of a trainable model. The reported values are
11 The mask seems bigger than the burned area. expressed as mean and standard deviation computed over
12 The mask is in the missing data area. the five folds.
13 Part of the mask is outside the area. In Figure 4, we highlight the percentage of burned pixels
per image in each fold. Even if data were split randomly,

1 1
Burned Pixels per Image (%)

Burned Pixels per Image (%)

0.8 0.8

0.6 0.6

0.4 0.4

0.2 0.2

0 0
0 1 2 3 4 0 1 2 3 4
Fold Fold
(a) (b)

FIGURE 4. The burned pixels percentage per image per fold: (a) setting 1 and (b) settings 2 and 3.

110 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


fold 0 is characterized by a larger variability in terms of the setting 3. SegFormer-B3 does not justify the greater com-
number of burned pixels per image. plexity considering its results. In this case, prefire images
do not provide any improvements, either. This can be justi-
SPECTRAL INDEXES fied because of the curse of dimensionality, which affects
We evaluated several spectral indexes (NBR, NBR2, BAI, almost all machine learning models. In fact, the concatena-
and BAIS2) for the burned area delineation task. In Table 3, tion approach we applied increases the number of features
we summarize the spectral bands exploited by the various without increasing the number of input samples. An open
indexes, and it is possible to note many bands are common
to many of them. They take as the input some bands of a
Sentinel product and output a value for each pixel. This val- TABLE 3. SPECTRAL BANDS USED BY THE INDEXES.
ue is generally thresholded to create a binary mask provid-
INDEX ULTRABLUE VISIBLE VNIR SWIR
ing the burned/unburned information for each pixel. NBR û û B8 B12
In particular, to assess the performances on the da- NBR2 û û û B11 and B12
taset, we computed the Separability Index (SI) [19] (see BAI û B4 B8 û
Table 4), which quantifies how well the index under BAIS2 û B4 B8A, B7, and B6 B12
analysis discerns between burned and unburned re-
VNIR, SWIR, and visible are exploited. B: band; SWIR: shortwave infrared; VNIR: visible
gions; i.e., a higher value of SI implies that classes are and near infrared.
more separable from each other. We apply Otsu’s thresh-
olding method to quantify the indexes’ segmentation
performances. The results, which are shown in Table 4,
confirm the poor performances in terms of F1 score and
TABLE 4. THE SI AND METRICS COMPUTED FOR EACH
Intersection over Union. Additionally, the availability of ­SETTING AND EACH EVALUATED INDEX.
prefire images (setting 3) does not significantly improve
SETTING INDEX SI F1 SCORE IOU
the evaluation metrics.
1 NBR 0.294 0.15 ± 0.231 0.103 ± 0.18
Figure 5 shows an example of predictions for the cited
NBR2 0.224 0.226 ± 0.269 0.159 ± 0.209
indexes applying Otsu’s method. In this example case,
BAI 0.044 0.04 ± 0.121 0.026 ± 0.086
BAI and NBR achieve the best scores, but many “distur-
BAIS2 0.027 0.194 ± 0.292 0.148 ± 0.252
bances” affect the final result in the unburned regions.
2 NBR 0.32 0.106 ± 0.196 0.071 ± 0.15
Figure 5(b), which is a zoom on an actual unburned
NBR2 0.349 0.243 ± 0.278 0.172 ± 0.218
area, shows that many false positive points are inside the
BAI 0.052 0.037 ± 0.115 0.024 ± 0.079
considered unburned area. We refer to this situation as BAIS2 0.002 0.086 ± 0.174 0.057 ± 0.138
­“ disturbances.” 3 dNBR 0.247 0.114 ± 0.212 0.079 ± 0.168
dNBR2 0.189 0.218 ± 0.281 0.157 ± 0.225
DEEP LEARNING MODELS dBAI 0.04 0.066 ± 0.161 0.045 ± 0.127
We tested two deep learning architectures for semantic seg- dBAIS2 0.027 0.047 ± 0.126 0.03 ± 0.099
mentation: a CNN (U-Net [16]) and a vision transformer dNBR: delta NBR; dNBR2: delta NBR2; dBAI: delta BAI; dBAIS2: delta BAIS2; IOU:
(SegFormer [15]). We decided to take into account two dif- Intersection over Union.

ferent versions of SegFormer (B0, the smallest version, and


B3, a mid-range version) that differ
only in size and, so, in the number
Image NBR NBR2 BAIS2 BAI Ground Truth
of parameters. U-Net, SegFormer-B0,
and SegFormer-B3 consist of 31 mil-
lion, 3.8 million, and 47 million pa-
rameters, respectively. To deal with
setting 3, the two input images (pre-
(a)
and postfire) are concatenated along
the channel axis, creating patches of Image NBR NBR2 BAIS2 BAI Ground Truth
size 24 # 152 # 152 (C # H # W) . We
report the results for the different set-
tings and models in Table 5.
Without any specific pretrain-
(b)
ing, U-Net provides the best perfor-
mance in every setting. SegFormer- FIGURE 5. An example of segmentation using Otsu’s method for different indexes:
B0, which is also lighter than U-Net, the (a) predictions and (b) zoomed predictions. In (b), the zoom was applied to the top left
provides comparable performance, corner of the images in (a), boxed in red. This is done to show more clearly an area with false
having some difficulties only with positive predictions.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 111


research direction is the design of more sophisticated mod- We provided baselines based on different approaches
els to exploit both images effectively. to assess the quality of basic methods and encourage fur-
Figure 6 reports the predictions of these models on ther research activities. This publicly available dataset can
the same input sample shown in Figure 5. The most evi- benefit researchers and public authorities for further tasks,
dent difference is that deep models tend to be more pre- such as recovery planning, constant monitoring of affect-
cise and less affected by false positives in the unburned ed areas, and the development of deep learning models for
areas (i.e., there are fewer “disturbances”). Looking at- burned area delineation. We plan to extend the dataset to
tentively at Figure 5(b), many false positive points can be new regions and satellite acquisitions continuously. The
seen. The greatest problem of collection of satellite acquisitions is made publicly avail-
thresholding techniques is able to encourage future use and research activities.
that they try to find a linear
WE PLAN TO EXTEND THE separation between classes, AUTHOR INFORMATION
which is frequently unreal- Daniele Rege Cambrin (daniele.regecambrin@polito.it) is
DATASET TO NEW REGIONS
istic. This is why deep learn- with Politecnico di Torino, 10129 Turin, Italy.
AND SATELLITE
ing models, which support Luca Colomba (luca.colomba@polito.it) is with Politecnico
ACQUISITIONS
nonlinearities, perform bet- di Torino, 10129 Turin, Italy.
CONTINUOUSLY. ter [36], [37]. The substantial Paolo Garza (paolo.garza@polito.it) is with Politecnico di
variability of the threshold Torino, 10129 Turin, Italy.
tec hnique, show n by t he
higher variance (see Table 4), REFERENCES
could be caused by the fact some areas are more sepa- [1] “Sentinel missions,” European Space Agency, Paris, France, 2023.
rable than others. [Online]. Available: https://sentinel.esa.int/web/sentinel/missions
[2] “Landsat mission,” Nat. Aeronaut. Space Admin., Washington,
CONCLUSION DC, USA, 2023. [Online]. Available: https://landsat.gsfc.nasa.gov/
This article introduced a new dataset for burned area de- [3] “Moderate-resolution imaging spectroradiometer (MODIS),” Nat.
lineation containing samples with different morphologi- Aeronaut. Space Admin., Washington, DC, USA , 2023. [Online].
cal features of the terrain and different sizes of burned re- Available: https://www.earthdata.nasa.gov/sensors/modis
gions. The dataset includes both prefire and postfire data. [4] Z. Dong, G. Wang, S. O. Y. Amankwah, X. Wei, Y. Hu, and A.
Feng, “Monitoring the summer flooding in the Poyang Lake area
of China in 2020 based on sentinel-1 data and multiple convo-
TABLE 5. THE METRICS CALCULATED FOR EACH DEEP lutional neural networks,” Int. J. Appl. Earth Observ. Geoinf., vol.
­LEARNING MODEL EVALUATED. 102, Oct. 2021, Art. no. 102400, doi: 10.1016/j.jag.2021.102400.
[5] A. Asokan and J. Anitha, “Change detection techniques for remote
SETTING METRIC SEGFORMERB3 SEGFORMERB0 U-NET
1 F1 score 0.62 ± 0.009 0.686 ± 0.004 0.707 ± 0.004
sensing applications: A survey,” Earth Sci. Informat., vol. 12, no. 2,
IoU 0.497 ± 0.008 0.563 ± 0.004 0.583 ± 0.004
pp. 143–160, Feb. 2019, doi: 10.1007/s12145-019-00380-5.
2 F1 score 0.583 ± 0.014 0.654 ± 0.003 0.705 ± 0.002 [6] J. Sublime and E. Kalinicheva, “Automatic post-disaster damage
IoU 0.447 ± 0.012 0.535 ± 0.003 0.577 ± 0.002 mapping using deep-learning techniques for change detection:
3 F1 score 0.533 ± 0.003 0.499 ± 0.009 0.625 ± 0.002 Case study of the Tohoku tsunami,” Remote Sens., vol. 11, no. 9,
IoU 0.401 ± 0.003 0.37 ± 0.007 0.494 ± 0.002 May 2019, Art. no. 1123, doi: 10.3390/rs11091123.
[7] R. Lasaponara and B. Tucci, “Identification of burned areas and se-
verity using SAR sentinel-1,” IEEE Geosci. Remote Sens. Lett., vol. 16,
no. 6, pp. 917–921, Jun. 2019, doi: 10.1109/LGRS.2018.2888641.
[8] D. R. Cambrin, L. Colomba, and P. Garza, “Vision transformers
Image SegFormerB3 SegFormerB0
for burned area delineation,” in Proc. Conf. Mach. Learn. Princ.
Pract. Knowl. Discovery Databases, 2022.
[9] M. A. Tanase et al., “Burned area detection and mapping: Inter-
comparison of sentinel-1 and sentinel-2 based algorithms over
tropical Africa,” Remote Sens., vol. 12, no. 2, Jan. 2020, Art. no.
334, doi: 10.3390/rs12020334.
U-Net Ground Truth
[10] “Sentinel-2 mission guide,” European Space Agency, Paris,
France, 2023. Accessed: Apr. 14, 2023. [Online]. Available: https://
sentinel.esa.int/web/sentinel/missions/sentinel-2
[11] L. Colomba et al., “A dataset for burned area delineation and se-
verity estimation from satellite imagery,” in Proc. 31st ACM Int.
Conf. Inf. Knowl. Manage., 2022, pp. 3893–3897, doi: 10.1145/
FIGURE 6. Examples of prediction with deep learning models. 3511808.3557528.

112 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


[12] Y. Prabowo et al., “Deep learning dataset for estimating burned [26] A. Farasin, L. Colomba, G. Palomba, G. Nini, and C. Rossi, “Su-
areas: Case study, Indonesia,” Data, vol. 7, no. 6, Jun. 2022, Art. pervised burned areas delineation by means of sentinel-2 imag-
no. 78, doi: 10.3390/data7060078. ery and convolutional neural networks,” in Proc. Int. Conf. Inf.
[13] “The department of forestry and fire protection,” California Syst. Crisis Response Manage. (ISCRAM), Blacksburg, VA, USA:
Department of Forestry and Fire Protection, Sacramento, Virginia Tech, 2020, pp. 24–27.
CA, USA, 2023. Accessed: Apr. 14, 2023. [Online]. Available: [27] D. Rashkovestsky, F. Mauracher, M. Langer, and M. Schmitt,
https://www.fire.ca.gov/ “Wildfire detection from multisensor satellite imagery using
[14] N. Otsu, “A threshold selection method from gray-level histo- deep semantic segmentation,” IEEE J. Sel. Topics Appl. Earth Ob-
grams,” IEEE Trans. Syst., Man, Cybern., vol. 9, no. 1, pp. 62–66, serv. Remote Sens., vol. 14, pp. 7001–7016, Jun. 2021, doi: 10.1109/
Jan. 1979, doi: 10.1109/TSMC.1979.4310076. JSTARS.2021.3093625.
[15] E. Xie, W. Wang, Z. Yu, A. Anandkumar, J. M. Alvarez, and [28] F. Huot, R. L. Hu, N. Goyal, T. Sankar, M. Ihme, and Y.-F. Chen, “Next
P. Luo, “SegFormer: Simple and efficient design for semantic day wildfire spread: A machine learning dataset to predict wildfire
segmentation with transformers,” in Proc. Adv. Neural Inf. Pro- spreading from remote-sensing data,” IEEE Trans. Geosci. Remote
cess. Syst., Red Hook, NY, USA: Curran Associates, 2021, vol. 34, Sens., vol. 60, pp. 1–13, Jul. 2022, doi: 10.1109/TGRS.2022.3192974.
pp. 12,077–12,090. [29] Q. Safder, H. Zhang, and Z. Zheng, “Burnt area segmentation
[16] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolution- with densely layered capsules,” in Proc. IEEE Int. G ­ eosci. Remote
al networks for biomedical image segmentation,” in Medical Sens. Symp., 2022, pp. 2199–2202, doi: 10.1109/IGARSS46834.
­Image Computing and Computer-Assisted Intervention, N. Navab, J. 2022.9884323.
Hornegger, W. Wells, and A. Frangi, Eds. Cham, Switzerland: [30] F. Montello, E. Arnaudo, and C. Rossi, “MMFlood: A multi-
Springer International Publishing, 2015, pp. 234–241. modal dataset for flood delineation from satellite imagery,”
[17] D. Bonafilia, B. Tellman, T. Anderson, and E. Issenberg, “Sen- IEEE Access, vol. 10, pp. 96,774–96,787, Sep. 2022, doi: 10.1109/
1Floods11: A georeferenced dataset to train and test deep learn- ACCESS.2022.3205419.
ing flood algorithms for sentinel-1,” in Proc. IEEE/CVF Conf. [31] D. John and C. Zhang, “An attention-based u-net for detect-
Comput. Vision Pattern Recognit. (CVPR) Workshops, Jun. 2020, pp. ing deforestation within satellite sensor imagery,” Int. J. Appl.
835–845, doi: 10.1109/CVPRW50498.2020.00113. Earth Observ. Geoinf., vol. 107, Mar. 2022, Art. no. 102685,
[18] R. A. Frey, S. A. Ackerman, R. E. Holz, S. Dutcher, and Z. Griffith, doi: 10.1016/j.jag.2022.102685.
“The continuity MODIS-VIIRS cloud mask,” Remote Sens., vol. [32] B. Ekim, T. T. Stomberg, R. Roscher, and M. Schmitt, “MapIn-
12, no. 20, Oct. 2020, Art. no. 3334, doi: 10.3390/rs12203334. Wild: A remote sensing dataset to address the question of what
[19] F. Filipponi, “BAIS2: Burned area index for sentinel-2,” makes nature wild [Software and Data Sets],” IEEE Geosci. Remote
­Proceedings, vol. 2, no. 7, Mar. 2018, Art. no. 364, doi: 10.3390/ Sens. Mag., vol. 11, no. 1, pp. 103–114, Mar. 2023, doi: 10.1109/
ecrs-2-05177. MGRS.2022.3226525.
[20] A. Fisher, N. Flood, and T. Danaher, “Comparing Landsat water [33] C. Yeh et al., “SustainBench: Benchmarks for monitoring the
index methods for automated water classification in eastern Aus- sustainable development goals with machine learning,” in Proc.
tralia,” Remote Sens. Environ., vol. 175, pp. 167–182, Mar. 2016, 35th Conf. Neural Inf. Process. Syst. Datasets Benchmarks Track
doi: 10.1016/j.rse.2015.12.055. (Round 2), Dec. 2021. [Online]. Available: https://openreview.
[21] N. Pettorelli, J. O. Vik, A. Mysterud, J.-M. Gaillard, C. J. Tucker, net/forum?id=5HR3vCylqD
and N. C. Stenseth, “Using the satellite-derived NDVI to assess [34] D. Sykas, M. Sdraka, D. Zografakis, and I. Papoutsis, “A senti-
ecological responses to environmental change,” Trends Ecol- nel-2 multiyear, multicountry benchmark dataset for crop clas-
ogy Evol., vol. 20, no. 9, pp. 503–510, Sep. 2005, doi: 10.1016/ sification and segmentation with deep learning,” IEEE J. Sel.
j.tree.2005.05.011. Topics Appl. Earth Observ. Remote Sens., vol. 15, pp. 3323–3339,
[22] L. Saulino et al., “Detecting burn severity across Mediterranean Apr. 2022, doi: 10.1109/JSTARS.2022.3164771.
forest types by coupling medium-spatial resolution satellite im- [35] C. H. Sudre, W. Li, T. Vercauteren, S. Ourselin, and M. Jorge
agery and field data,” Remote Sens., vol. 12, no. 4, Feb. 2020, Art. Cardoso, “Generalised dice overlap as a deep learning loss func-
no. 741, doi: 10.3390/rs12040741. tion for highly unbalanced segmentations,” in Deep Learning in
[23] W. Bin et al., “A method of automatically extracting forest fire Medical Image Analysis and Multimodal Learning for Clinical Decision
burned areas using GF-1 remote sensing images,” in Proc. IEEE Support. Cham, Switzerland: Springer International Publishing,
Int. Geosci. Remote Sens. Symp. (IGARSS), 2019, pp. 9953–9955, 2017, pp. 240–248.
doi: 10.1109/IGARSS.2019.8900399. [36] K. Simonyan and A. Zisserman, “Very deep convolutional net-
[24] C. A. Cansler and D. McKenzie, “How robust are burn severity works for large-scale image recognition,” 2014, arXiv:1409.1556.
indices when applied in a new region? Evaluation of alternate [37] A. Garcia-Garcia, S. Orts-Escolano, S. Oprea, V. Villena-Marti-
field-based and remote-sensing methods,” Remote Sens., vol. 4, nez, P. Martinez-Gonzalez, and J. Garcia-Rodriguez, “A survey
no. 2, pp. 456–483, Feb. 2012, doi: 10.3390/rs4020456. on deep learning techniques for image and video semantic seg-
[25] L. Knopp, M. Wieland, M. Rättich, and S. Martinis, “A deep mentation,” Appl. Soft Comput., vol. 70, pp. 41–65, Sep. 2018. [On-
learning approach for burned area segmentation with sentinel-2 line]. Available: https://www.sciencedirect.com/science/article/
data,” Remote Sens., vol. 12, no. 15, Jul. 2020, Art. no. 2422, doi: pii/S1568494618302813, doi: 10.1016/j.asoc.2018.05.018.
10.3390/rs12152422. GRS

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 113


CONFERENCE REPORTS
ALBERTO MOREIRA , FRANCESCA BOVOLO , DAVID LONG , AND ANTONIO PLAZA

IGARSS 2023 in Pasadena, California


Impressions of the First Days

A fter four years of online and hybrid conferences, the


43rd IEEE International Geoscience and Remote
Sensing Symposium (IGARSS) was held in person again
Misra (see Figure 2). They gave the welcome address, de-
scribed the logistics for the week as well as the social
program, and highlighted the most important events of
on 16–21 July 2023. The symposium was located at the Pas- the week.
adena Convention Center, Pasadena, in sunny California, The IGARSS 2023 technical program was presented
USA, only 11 miles from Los Angeles (Figure 1). At the base by Dr. Rashmi Shah and Dr. David Kunkee, technical
of the San Gabriel Mountains in the San Gabriel Valley, committee cochairs (see Figure 3). First, Dr. Kunkee pre-
filled with historic architecture and national landmarks, sented some astonishing statistics from the ­conference.
Pasadena is known as the “crown of the valley.” The success of such a big conference is unthinkable with-
To organize such a big conference with thousands out the help of many volunteers: 328 session o ­ rganizers
of attendees from around the world may have posed
a challenge, but it was met most professionally by the
local IGARSS 2023 organizing team, Conference Man-
agement Services, and the IEEE Geoscience and Remote
Sensing Society (GRSS), and it exceeded expectations.
The symposium aimed at providing a platform for shar-
ing knowledge and experience on recent developments
and advancements in geoscience and remote sensing
technologies, particularly in the context of Earth ob-
servation, disaster monitoring, and risk assessment.
A variety of programs was offered, such as keynote
talks, technical sessions, tutorials, exhibitions, a Young
Professionals’ (YP’s) mixer, presentation and writing
workshops, a career panel, a Technology, Industry, and FIGURE 1. City of Pasadena, CA, USA, the venue of IGARSS 2023.
Education (TIE) forum, a technical tour, an awards ban- [Courtesy of Visit Pasadena (visitpasadena.com)]
quet, and also a 3-min thesis competition, a student paper
contest, and a summer school prior to the symposium.
Following are some highlights from the IGARSS
2023 opening and plenary session, held on Monday,
17 July 2023.

WELCOME ADDRESSES AT THE PLENARY SESSION


The plenary ceremony of IGARSS 2023 started on Mon-
day with an introduction of the conference by IGARSS
©IGARSS 2023

2023 general cochairs Shannon Brown and Sidharth

Digital Object Identifier 10.1109/MGRS.2023.3303685


FIGURE 2. Opening remarks of IGARSS 2023 general cochairs
Date of current version: 19 September 2023 Shannon Brown (right) and Sidharth Misra (left).

114 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


TABLE 1. PRESENTATIONS AND ATTENDANCE.

TOTAL PAPERS TOTAL PAPERS ORAL POSTER ORAL POSTER TOTAL


­SUBMITTED ACCEPTED PAPERS PAPERS SESSIONS ­SESSIONS ­REGISTERED STUDENTS
3,688 2,868 1,534 1,334 319 154 2,768 961

and 1,294 reviewers were involved in the


preparations. With about 3,700 attendees,
there was a record number of submissions,
which finally resulted in 2,868 oral or
poster presentations assisted by 475 ses-
sion cochairs and 120 session managers
(see Table 1).
Several changes were made this year to
improve the IGARSS technical program:
Shorter abstracts were used to streamline
the abstracts submission and, instead of
invited sessions “community contributed

©IGARSS 2023
sessions” were introduced to get more of
the community involved. Also, the struc-
ture of the oral session was changed to FIGURE 3. Technical committee cochairs Dr. Rashmi Shah (right) and Dr. David Kunkee
include more discussion among the com- (left) presenting the technical program.
munity members: presentations were
planned to be 12 min long, and within a session there was After Dr. Coughlin’s presentation, GRSS President Dr.
only at the end of all presentations a 15-min slot in which Mariko S. Burgin gave a warm welcoming address to all at-
to ask questions. tendees and reported on the activities of the GRSS (see
Dr. Rashmi recommended the TIE events and YP events. Figure 5). The GRSS is one of 39 IEEE Societies, and it is a truly
She also spoke about the 13 tutorials with more than 300 global community. It has nearly 5,000 members in 144 coun-
participants that were held in the run-up to the conference tries and is organized into 128 Chapters with 12 ambassadors
and were a great success. all over the world. The GRSS is governed by the GRSS AdCom.
After Dr. Rashmi spoke, 2023 IEEE President-Elect
Dr. Tom Coughlin’s introduction followed (see Figure 4).
He sees IEEE as a resource for technology decisions.
As technology of all sorts drives the world’s economy,
this is something that is very important to be aware of.
IEEE is the largest technical professional organization
in the world, its members are involved in all aspects
of technology creation and use, its research powers
patents, and it creates the world’s technical standards.
©IGARSS 2023

IEEE also fosters efforts in future directions, technical


road maps, and tracking megatrends as well as inform-
ing public policy and serving as a resource for techni- FIGURE 4. 2023 IEEE President-Elect Dr. Tom Coughlin’s introduction.
cal discussions.
IEEE has more than 420,000 members in more than 190
countries and sponsors more than 2,000 conferences in 96
countries annually. In addition to its 46 Societies and tech-
nical councils, it provides a lot of volunteer opportunities
that help Members to build networks and learn new con-
cepts. IEEE is the most-cited publisher in new patents from
top-patenting organizations, and IEEE research is increas-
ingly valuable to innovators.
©IGARSS 2023

As part of his IEEE presidency, Dr. Coughlin would like


to increase IEEE’s outreach to younger members and the
broader public, increase engagement with industry, and FIGURE 5. Opening remarks of Dr. Mariko Burgin, 2023 president
make investments in new products and services. of the GRSS, during the plenary session.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 115


The GRSS fosters engagement of its members for the and other opportunities, like GRSS schools, webinar series,
benefit of society through science, engineering, applica- high school and undergraduate student outreach, student
tions, and education as related to the development of the grand challenges, travel grants, and more.
field of geoscience and remote sensing. The GRSS is also All of this speaks in favor of choosing the GRSS as a “pro-
a group of scientists, researchers, and practitioners with fessional home”—a place for technical excellence, where
common interests and a common framework for building you are welcome and where you belong. It is a place that
a community. you can make your own because the GRSS can help interact
Dr. Burgin also highlighted five important GRSS areas. with like-minded researchers, engineers, and developers to
The GRSS disseminates premium science by sponsoring four make a difference in the world through remote sensing.
refereed publications [IEEE Geoscience and Remote Sensing
Magazine (GRSM), IEEE Transactions on Geoscience and Remote MAJOR AWARDS CEREMONY
Sensing (TGRS), IEEE Journal of Selected Topics in Applied Earth Following the opening remarks for IGARSS 2023, Prof. Al-
Observations and Remote Sensing (J-STARS), and IEEE Geosci- berto Moreira, chair of the GRSS Major Awards Commit-
ence and Remote Sensing Letters], cosponsoring IEEE Journal tee, opened the 2023 awards ceremony. As in the past, the
on Miniaturization for Air and opening and plenary session of IGARSS 2023 was chosen
Space Systems and curating an for the recognition of IEEE GRSS members elevated to IEEE
eNewsletter. Eight GRSS tech- Fellow grade and the four major awards of the GRSS. For
ALL OF THIS SPEAKS IN nical committees organize a each award, 2023 GRSS President Dr. Mariko Burgin and
FAVOR OF CHOOSING THE wide variety of technical ac- IEEE President Dr. Tom Coughlin presented the recogni-
GRSS AS A “PROFESSIONAL tivities. The GRSS provides tions and congratulated the awardees.
HOME”—A PLACE FOR TECH- connections and networking
NICAL EXCELLENCE, WHERE opportunities, such as the IEEE FELLOW AWARDS
IGARSS, regional symposia, The grade of IEEE Fellow recognizes unusual distinction in
YOU ARE WELCOME AND
smaller conferences, and co- the profession, and it is conferred only by invitation of the
WHERE YOU BELONG.
sponsored conferences and IEEE Board of Directors upon a person of outstanding and
workshops. The GRSS also or- extraordinary qualifications and experience in IEEE-desig-
ganizes professional activities nated fields. The IEEE bylaws limit the number of Members
with distinguished lecturers, YP events, professional devel- who can be advanced to Fellow grade in any one year to one
opment microgrants, the Women Mentoring Women pro- per mil of the Institute membership, exclusive of students
gram, and much more. Finally, the GRSS provides learning and affiliates. To qualify, the candidate must be a Senior
Member and must be nominated by an individual familiar
with the candidate’s achievements. Endorsements are re-
quired from at least five IEEE Fellows and an IEEE Society
best qualified to judge. The GRSS IEEE Fellow Committee
completes the first evaluation of the nominees. After this, the
IEEE Fellow Committee, comprising approximately 50 IEEE
Fellows, carefully evaluates all nominations, considering the
Society rankings and presents a list of recommended candi-
dates to the IEEE Board of Directors for the final election.
On average, the GRSS performs above the average with
respect to the number of elected Fellows every year. This
year we have four GRSS members who were promoted to
IEEE Fellow.
The first Fellow recognition went to Prof. James Garrison
with the following citation (see Figure 6): “For contribu-
tions to Earth remote sensing using signals of opportunity.”
James Garrison is a professor in the School of Aero-
nautics and Astronautics at Purdue University with a cour-
tesy appointment in the School of Electrical and Computer
Engineering and the Ecological Sciences and Engineering
Interdisciplinary Graduate program. His research interests
include Earth remote sensing using GNSSs and signals of
opportunity. He is the principal investigator for SNOOPI, a
©IGARSS 2023

NASA mission to demonstrate remote sensing with P-band


signals of opportunity. Prior to his academic position, Prof.
FIGURE 6. IEEE Fellow Award recipient Prof. James Garrison. ­Garrison was with NASA. He earned a Ph.D. degree from

116 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


©IGARSS 2023

©IGARSS 2023
FIGURE 7. IEEE Fellow Award recipient Prof. Jonathan Li (middle)
with Dr. Mariko Burgin (right) and IEEE President-Elect Dr. Tom
Coughlin (left). FIGURE 8. IEEE Fellow Award recipient Prof. Gabriele Moser.

the University of Colorado Boulder in 1997 and also holds Engineering and the Engineering Institute of Canada and is
a B.S. degree from the Rensselaer Polytechnic Institute and the recipient of more than 20 prestigious awards.
an M.S. degree from Stanford University. He is a fellow The third Fellow recognition was received by Prof. Ga-
of the Institute of Navigation. From 2018 to 2022 he was briele Moser with the following citation (see Figure 8): “For
editor-in-chief of GRSM. contributions to pattern recognition in remote sensing.”
The second Fellow recognition went to Prof. Jonathan Li Prof. Gabriele Moser is a full professor of telecommu-
with the citation (see Figure 7) “For contribution to point nications at the University of Genoa. His research activ-
cloud analytics in lidar remote sensing.” ity is focused on pattern recognition and image processing
Prof. Jonathan Li received his Ph.D. degree in geomatics methodologies for remote sensing and energy applica-
engineering from the University of Cape Town, South Africa, tions. He served as chair of the GRSS Image Analysis and
in 2000. Prof. Jonathan Li is currently a professor of geomat- Data Fusion Technical Committee (IADF TC) from 2013 to
ics and systems design engineering 2015 and as IADF TC cochair from
with the University of Waterloo, 2015 to 2017. He was publication
Canada. His main research interests cochair of IGARSS 2015, technical
include artificial intelligence (AI)- program cochair of the GRSS Earth-
based 3D geospatial information Vision workshop at the 2015 IEEE
extraction from Earth observation Conference on Computer Vision
images and lidar point clouds, pho- and Pattern Recognition (CV PR),
togrammetry and pointgrammetry and coorganizer of the second edi-
for high-definition map generation, tion of EarthVision at CVPR 2017.
3D vision, and GeoAI for digital Since 2019, he has been the head of
twin cities. He has coauthored more the M.Sc. program in Engineering
than 530 publications, more than for Natural Risk Management at the
330 of which were published in University of Genoa. Since 2021, he
refereed journals. He has also pub- has been a member of the national
lished papers in flagship conferenc- evaluation committee for national
es in computer vision and AI. His scientific qualification (Abilitazione
publications have received more Scientifica Nazionale) as a full pro-
than 15,000 Google citations with fessor in the telecommunications
an h-index of 65. He has supervised field in Italy.
©IGARSS 2023

more than 120 master’s and Ph.D. The fourth Fellow recognition
students as well as postdoctoral fel- was presented to Prof. Ping Yang
lows to completion. He is a fellow FIGURE 9. IEEE Fellow Award recipient Prof. w it h t he c itat ion (see Fig ure 9)
of both the Canadian Academy of Ping Yang. “For s e m i n a l cont r ibut ion s to

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 117


r­ adiative transfer and remote sensing of ice clouds and try Leader Award, and the GRSS Fawwaz Ulaby Distin-
dust aerosols.” guished Achievement Award was posted in 2022 on the
Prof. Ping Yang is a University Distinguished Professor at GRSS website and announced in the eNewsletter of the
Texas A&M University (TAMU), where he currently serves GRSS. The nomination forms are available at http://
as senior ­associate dean for research and graduate studies in www.grss-ieee.org/about/awards/. Any member, with the
the College of Arts and Sciences. He has joint professorship exception of GRSS AdCom members, can make nomi-
appointments with the Department of Physics & Astron- nations to recognize deserving individuals. Typically,
omy and the Department of Oceanography, and he holds the lists of nominated candidates comprise three to five
the ­David Bullock Harris Chair in Geosciences. Prof. Yang names each year. An independent Major Awards Evalua-
has supervised the completion of 29 doctoral dissertations tion Committee makes the selection, which is approved
and 20 master’s degree theses. He has published 359 peer- by the GRSS president and AdCom. The following major
reviewed journal papers and four monographs. His publica- awards were presented:
tions have been cited 23,483 times (Google Scholar) with an ◗◗ Outstanding Service Award
h-index of 78 (Google Scholar). His research focuses on light ◗◗ Education Award
scattering, radiative transfer, and satellite-based remote sens- ◗◗ Industry Leader Award
ing. Prof. Yang is a recipient of the NASA Exceptional Scien- ◗◗ Fawwaz Ulaby Distinguished Achievement Award.
tific Achievement Medal (2017), the American Geophysical
Union Atmospheric Sciences Ascent Award (2013), the David IEEE GRSS OUTSTANDING SERVICE AWARD
and Lucille Atlas Remote Sensing Prize from the American The Outstanding Service Award was established to recog-
Meteorological Society (2020), and the Van de Hulst Light- nize an individual who has given outstanding service for
Scattering Award from Elsevier (2022). Prof. Yang was named the benefit and advancement of the GRSS. The award is
the 2022 Distinguished Texas S­ cientist by the Texas Academy considered annually but will not be presented unless a suit-
of Science. Within TAMU, he received a university-level fac- able candidate is identified. The following factors are sug-
ulty research award bestowed by the Association of Former gested for consideration: leadership innovation, activity,
Students in 2017 and several college-level awards. service, duration, breadth of participation, and coopera-
The IEEE Fellow recognition part of the awards ceremo- tion. GRSS membership is required. The awardee receives a
ny was concluded with a group photo with the four IEEE certificate and a plaque.
Fellows nominated by the GRSS together with GRSS Presi- The 2023 GRSS Outstanding Service Award was pre-
dent Dr. Mariko Burgin and IEEE President-Elect Dr. Tom sented to Dr. Simon Yueh with the following citation (see
Coughlin (Figure 10). Figure 11): “In recognition of his outstanding service for
the benefit and advancement of the Geoscience and Re-
GRSS MAJOR AWARDS AT THE AWARDS CEREMONY mote Sensing Society.”
The call for nominations for the GRSS Education Award, Simon Yueh received his Ph.D. degree in electrical engi-
the GRSS Outstanding Service Award, the GRSS Indus- neering in January 1991 from the Massachusetts I­ nstitute
©IGARSS 2023

FIGURE 10. IEEE Fellow Award recipients Prof. Gabriele Moser (second from left), Prof. Jonathan Li, Prof. James Garrison, and Prof. Ping
Yang (second from right) with GRSS President Dr. Mariko Burgin (right) and IEEE President-Elect Dr. Tom Coughlin (left).

118 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


of Technology. In September 1991,
he joined the Radar Science and
Engineering Section at the Jet Pro-
pulsion Laboratory (JPL). He was
the supervisor of the Radar Sys-
tem Engineering and Algorithm
Development Group during 2002–
2007, the deputy manager of the
Climate, Oceans and Solid Earth
Section from July 2007 to March
2009, and the section manager
from April 2009 to 2013 January.
He served as the project scientist of
the NASA Aquarius mission from
January 2012 to September 2013,

©IGARSS 2023
the deputy project scientist of the
NASA Soil Moisture Active Passive
mission from 2013 January to Sep- FIGURE 11. 2023 IEEE GRSS Outstanding Service Award recipient Dr. Simon Yueh.
tember 2013, and the Soil Moisture
Active Passive Project scientist since October 2013. He has affiliation is required. The awardee receives a certificate
been the principal/coinvestigator of numerous NASA and and a plaque.
U.S. Department of Defense research projects on remote The 2023 GRSS Education Award was presented to
sensing of ocean salinity, ocean wind, terrestrial snow, Prof. Shutao Li with the citation (see Figure 12) “In recog-
and soil moisture. He has authored four book chapters nition of his significant educational contributions to geo-
and published more than 300 publications and presenta- science and remote sensing.”
tions. He received the 2021 J-STARS Prize Paper Award, Shutao Li received his B.S., M.S., and Ph.D. degrees
the 2014, 2010, and 2002 GRSS Transactions Prize Paper from Hunan University, Changsha, China, in 1995,
awards, the 2000 Best Paper Award at IGARSS 2000, and 1997, and 2001, respectively. Prof. Shutao Li has been a
the 1995 GRSS Transactions Prize Paper award for a paper full professor with the College of Electrical and Informa-
on polarimetric radiometry. He received the JPL Lew Allen tion Engineering, Hunan University, since 2004 and is
Award for Excellence in 1998, the Ed Stone Award in 2003, currently the vice rector of Hunan University. Prof. Li’s
the NASA Exceptional Technology Achievement Medal in current research interests include remote sensing image
2014, and the NASA Outstanding Public Leadership Med- processing, pattern recognition, AI, and applications
al in 2017. He was an associate edi- in environmental observation, re-
tor of Radio Science from 2002 to source investigation, and precise
2006 and editor-in-chief of TGRS agriculture. He has authored or co-
from 2018 to 2022. authored more than 300 refereed
journal and international confer-
IEEE GRSS EDUCATION AWARD ence papers. He has received more
The Education Award was estab- than 28,000 citations in Google
lished to recognize an individual S c hola r ( h-i nde x : 79) a nd wa s
who has made significant educa- selected as a Clarivate Analytics’
tional contributions to the field of Global Highly Cited Researcher in
GRSS. In selecting the individual, 2018–2022. For his scientific re-
the factors considered are signifi- search contributions, he received
cance of the educational contribu- two Second-Grade State Scientific
tion in terms of innovation and and Technological Progress Awards
the extent of its overall impact. of China (in 2004 and 2006), a Sec-
The contribution can be at any ond Prize of the National Natural
level, including K-12, undergradu- Science Award by the State Council
ate, and graduate teaching, profes- of China (in 2019), and two First
sional development, and public Prize Hunan Provincial Natural
©IGARSS 2023

outreach. It can also be in any form Science Awards (in 2017 and 2022).
(e.g., textbooks, curriculum devel- Prof. Li is the founder and head of
opment, and educational program FIGURE 12. 2023 IEEE GRSS Education Award the Hunan Provincial Key Labo-
initiatives). GRSS membership or recipient Prof. Shutao Li. rator y of Visual Perception and

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 119


­ rtificial Intelligence. He also founded the GRSS Chang-
A the company’s long-term strategic trajectory and oversees
sha Chapter and cofounded the International Joint Re- Planet’s Space Systems, Corporate Development, and
search Center for Hyperspectral Imaging and Processing. Special Projects functions. He spearheaded Planet’s ac-
He is now an associate editor of TGRS and IEEE Transac- quisition of BlackBridge in 2015, Boundless in 2019, and
tions on Instrumentation and Measurement and a member VanderSat in 2021. Prior to Planet, Mr. Schingler spent nine
of the editorial board of Information Fusion. years at NASA, where he helped build the Small Spacecraft
Office at NASA Ames and served as chief of staff for the
IEEE GRSS INDUSTRY LEADER AWARD Office of the Chief Technologist at NASA headquarters.
The GRSS established the Industry Leader Award to rec- He received an MBA from Georgetown University, an M.S.
ognize an individual who has made significant contribu- degree in space studies from the International Space Uni-
tions over a sustained period of time in an industrial and/ versity, and a B.S. degree in engineering physics from Santa
or a commercial remote sensing discipline. The evaluation Clara University.
awards committee may give preference to an individual
who 1) is a GRSS member, 2) has made significant con- IEEE GRSS FAWWAZ ULABY DISTINGUISHED
tributions to remote sensing system engineering, science, ACHIEVEMENT AWARD
and/or technology, 3) has made significant contributions The Fawwaz Ulaby Distinguished Achievement Award
to dissemination and commercialization of remote sens- was established to recognize an individual who has made
ing products, and 4) has demonstrated leadership to pro- significant technical contributions, within the scope of
mote remote sensing science and technology. Criteria for GRSS, usually over a sustained period. In selecting the in-
selection are significance, quality, and impact of activities dividual, the factors considered are quality, significance,
and contributions and achievements. The award is consid- and impact of the contributions; quantity of the contri-
ered annually and presented if a distinguished candidate butions; duration of significant activity; papers published
is identified. in archival journals; papers presented at conferences and
The 2023 GRSS Industry Leader Award was presented symposia; patents granted; and advancement of the pro-
to Robbie Schingler with the following citation (see Fig- fession. IEEE membership is preferable, but not required.
ure 13): “For co-founding Planet and for outstanding con- The award is considered annually and presented only if
tributions for the commercialization and dissemination of a suitable candidate is identified. The awardee receives a
optical remote sensing data.” plaque and a certificate.
Robbie Schingler is a director, cofounder, and chief The 2023 IEEE GRSS Fawwaz Ulaby Distinguished
strategy officer (CSO) at Planet. As CSO, Mr. Schingler leads Achievement Award was presented to Prof. Howard
©IGARSS 2023

©IGARSS 2023

FIGURE 13. 2023 GRSS Industry Leader Award recipient FIGURE 14. IEEE GRSS Fawwaz Ulaby Distinguished Achievement
Robbie Schingler. Award recipient Prof. Howard Zebker.

120 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


©IGARSS 2023
FIGURE 15. Group photo at the end of the major awards ceremony (from left to right): IEEE President-Elect Dr. Tom Coughlin, Prof. Gabriele
Moser, Prof. James Garrison, Prof. Ping Yang, Prof. Howard Zebker, Dr. Simon Yueh, GRSS President Dr. Mariko Burgin, and Major Awards
Chair Prof. Alberto Moreira.

­ ebker with the citation (see Figure 14) “For sustained


Z KEYNOTE SPEECHES AT THE PLENARY SESSION
­outstanding contributions and leadership in the field of After the awards ceremony, the plenary session started with
radar interferometry.” presentations by three distinguished plenary speakers:
Howard Zebker was born in Ventura, CA, USA, received ◗◗ “Dare Mighty Things Together,” by Dr. Laurie Leshin,
his B.S. degree from Caltech in 1976, his M.S. degree from director of NASA’s JPL
UCLA in 1979, and his Ph.D. degree from Stanford in ◗◗ “Talking Pigs—Math Stories From the Visual Effects In-
1984. Dr. Howard Zebker is currently a professor of geo- dustry,” by Joe Mancewicz, senior software engineer in
physics and electrical engineering at Stanford University, Nvidia’s Omniverse Group
where his research group specializes in interferometric ra- ◗◗ Keynote speech by Elia Saikaly, award-winning adven-
dar remote sensing. Originally a microwave engineer, he ture filmmaker based in Ottawa, Canada.
built support equipment for the Seasat satellite synthetic The first keynote speaker, Dr. Laurie Leshin, director of
aperture radar (SAR) and designed airborne radar systems. NASA’s JPL, gave insights into the work of JPL, the ­c urrent
He later developed imaging radar polarimetry, a technique status of activities in Earth science, and future challenges
for measurement of the radar scattering matrix of a surface. (see Figure 16). As the conference’s participants also had
He is best known for the development of radar interferom-
etry, leading to spaceborne and airborne sensors capable
of measuring topography to meter scale accuracy and sur-
face deformation to millimeter scale. More recently he has
participated in the NASA Cassini mission to Saturn and
currently is concentrating on the upcoming NASA/Indian
Space Research Organisation (ISRO) mission.
The major awards ceremony concluded with a group
photo with all awardees together with GRSS President Dr.
Mariko Burgin, IEEE President-Elect Dr. Tom Coughlin,
©IGARSS 2023

and GRSS Major Awards Chair Prof. Alberto Moreira (see


Figure 15). The deadline for nominations for the 2024 ma-
jor awards and special awards of the GRSS is 15 December FIGURE 16. Dr. Laurie Leshin, Director of NASA’s JPL, provided
2023. A detailed description of the awards is available at a speech on “Dare Mighty Things Together” in the IGARSS
https://www.grss-ieee.org/resources/awards/. plenary session.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 121


the possibility during the week to join a tour of JPL, this In the field of astrophysics, Dr. Leshin described the Nan-
presentation was especially interesting and exciting. cy Grace Roman Space Telescope, which will be launched in
Dr. Leshin related JPL’s history—how everything started 2027. The mission is managed by NASA Goddard, but JPL
as early as back in the 1930s, even before NASA was found- is building the coronagraph instrument, which is one of its
ed. JPL has 6,500 employees and is a unique organization. key instruments. The mission is aiming to image planets
Nationwide and also international teamwork and collabo- in other solar systems. More than 5,000 exoplanets have
ration are ­important parts of JPL’s life. already been detected, but we cannot see them because of
JPL has four big topics in its portfolio: Earth science, the brightness of their stars. To be able to see one of those
planetary science, and astrophysics, and it operates for planets, one has to block out its star very effectively, and
NASA the Deep Space Network with satellite antennas on this is what the coronagraph instrument is designed to do.
three different locations all around the globe. These anten- In the field of planetary science, Dr. Leshin talked about
nas are constantly in contact with JPL’s deep space missions. the mission Europa Clipper, which is to be launched in
In a further presentation, Dr. Leshin presented these topics October 2024. Europa is a moon of Jupiter and is covered
in more detail. by a water–ice shell. Beneath that icy shell, a global ocean
The work of JPL in Earth science focuses on the areas of is suspected, which could harbor life. Europa Clipper will
biodiversity, greenhouse gases, water availability, air qual- go into orbit around Jupiter and have multiple very close
ity, sea level, and natural hazards. JPL is interested not only flybys—some of them go down only tens of kilometers
in the big missions, but also in the applications that can above the surface—with many instruments to study and
improve people’s lives. Dr. Leshin showed a couple of ex- understand the environment on Europa. Another impor-
amples, like the mission EMIT, which is a multispectral sci- tant question to be answered is where future missions could
ence mission originally designed to look at the composition more easily access the planet.
of mineral dust, but that also has the ability to detect meth- Another mission called Psyche is going to be launched in
ane plumes and sees very clearly from space where meth- October 2023. The mission goes to the asteroid called Psyche.
ane emission is happening. As methane is a very powerful The Mars mission with the rover Perseverance and the
greenhouse gas, which has a much shorter residence time helicopter Ingenuity was launched in 2021. They are ex-
in the atmosphere than CO2, providing this information ploring and sampling the inside of a crater that once held
offers an excellent possibility to get ahead of global warm- a lake. By studying the stored small rocks, they want to
ing in an effective way. answer the question of whether life could have started in
The mission Surface Water and Ocean Topography the Mars environment at the same time life was starting on
(SWOT) was launched in 2022 in collaboration between Earth. Finally, Dr. Leshin showed an animation about the
the French Space Agency CNES and NASA. The SWOT Mars Sample Return mission, where all of the rock samples
mission uses radar interferometry, achieving much higher are planned to be transported to Earth. She concluded her
resolution ocean topography and higher spatial resolution speech by referring to Voyager missions 1 and 2 from 1977
and getting closer to the shore than (the spacecraft are still sending sig-
usual. SWOT is going to represent a nals every day), and she expressed
spectacular revolution in our under- her pleasure to be able to do this
standing of the oceans. The mission kind of work for the benefit of the
helps us to understand what is hap- science community.
pening to Earth’s surface water and The second speaker of t he
will measure the height of millions of plenary session was Joe Mance-
lakes and rivers around the world to wicz, senior software engineer in
about a few centimeters, which will Nvidia’s Omniverse Group (see
provide revolutionary insights into Figure 17). After showing an aston-
the distribution of fresh water. ishing video and explaining tech-
JPL’s next big science launch is nical terms, Mr. Mancewicz talked
planned for 2024 in southern India about problems that arise when
and is called the NASA ISRO SAR creating an animation of a moving
(NISAR). The dual-frequency SAR figure. The character’s behavior,
satellite with a 12-m reflector is en- and even its mood and the texture
tering the final testing phase and will of its clothing, must look realistic.
©IGARSS 2023

be able to look at all kinds of land To make the scene imagined by


surface changes, for example, chang- the authors work on the screen, it
es in biomass, ice, earthquakes, or FIGURE 17. The plenary session speech on “Talk- takes a lot of sophisticated work
volcanoes, and as such will be a real ing Pigs—Math Stories From the Visual Effects and a bunch of good ideas. It is a
game changer in understanding land Industry” was given by Joe Mancewicz, senior long way from using nature photos
surface change. software engineer in Nvidia’s Omniverse Group. to several models until you finally

122 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


arrive at a well-done computer animation. Even the rep- in the classroom to world-class expeditions. Expeditions
resentation of the lighting alone needs many iterations. and campaigns have enabled students to raise money for
Mr. Mancewicz demonstrated these challenges on a scene a well in Kathmandu, classrooms in Kenya, an orphanage
from the movie “Night at the Museum: Battle of the Smith- in India, and, most recently, a new school in the village of
sonian,” where Amy Adams and Ben Stiller are looking at a Banakhu, Nepal, after the 2015 earthquake, which tragi-
giant octopus and Ben Stiller throws water onto the animal. cally claimed more than 8,000 lives. FindingLife’s mission
A further example was from the movie “Life of Pi”: is to inspire positive change in youth by bringing the world
after a disastrous shipwreck, Pi Patel, son of an Indian right into their classrooms, and its motto is: Educate. In-
zoo director, is floating in a lifeboat in the middle of the spire. Empower.
ocean—together with a Bengal tiger. Joe Mancewicz talked It was in the framework of this project that Elia Saikaly
about the layers of technical animation: the depiction of successfully climbed Mount Everest. In his speech he talk-
the boat scene required many carefully coordinated steps, ed about the challenges and choices of such a man-trying
from shooting in a swimming pool to simulating the tiger’s enterprise, and he also emphasized that, without the help
muscles, skin, fur, and movements to placing the scene in of the Sherpas, you would not stand a chance. He showed
an ocean environment. photos of himself and his tent at the base of Mount Everest,
Another issue that was explained was the refraction where he connected live on Skype to the schools three times
problem. Mr. Mancewicz demonstrated on a virtual model a week. He let us experience the excitement of the climb.
of an eye the challenge of computing the effects of break- After this first successful expedition, Elia Saikaly
ing light, which is also relevant in the field of Earth ob- planned the next adventure, which was climbing K2, the
servation, if the working and movements of satellites are world’s second highest peak, in the winter season, which
modeled and simulated. was unprecedented until then.
At the end of his truly amazing speech, Mr. Mancewicz On 16 January 2021, a team of 10 alpinists from Ne-
showed that animations can also support the planning and pal, led by Nirmal Purja and Mingma Gyalje Sherpa, suc-
optimization of satellites. As CubeSat picosatellites have ceeded in making the winter ascent of K2. Other alpin-
only a very limited area on their walls for solar cells assem- ists started too, including John Snorri Sigurjónsson from
bling, the available area has to be effectively shared with Iceland, Juan Pablo Mohr Prieto from Chile, and Ali Sad-
other parts. With an animation you can choose the most para from Pakistan, but unfortunately, they went miss-
suitable places for solar panels where the sun’s rays will be ing during the attempt. Elia Saikaly was part of the inter-
the strongest; other parts can be placed in other locations. national team that was searching for the mountaineers.
The third keynote speaker of the plenary session, Elia The search was coordinated together with the Icelandic,
Saikaly (https://eliasaikaly.com/), an award-winning ad- Pakistani, and Chilean governments and also involved
venture filmmaker based in Ottawa, Canada, started his the Icelandic Space Agency. The agency was mapping the
presentation with a short video with breathtaking images mountain and providing satellite imagery, but it could
of the top of the world (see Figure 18). Elia Saikaly has par- not find the missing mountaineers. Six months later, in
ticipated in more than 25 world-class expeditions, includ-
ing 10 to Mount Everest—always with a camera in hand.
Elia Saikaly illustrated through his own story the power of
storytelling, which is also important for researchers when
they try to explain their findings.
From a state of “had been written off,” Elia Saikaly was
able to become a successful documentary filmmaker. In his
early 20s he discovered for himself the video camera and
fell in love with storytelling. Dr. Sean Egan was one of the
key people who had a great influence on him, and Saikaly
fulfilled Dr. Egan’s dream of climbing Mount Everest, which
he unfortunately could not achieve because he died of heart
failure. Dr. Egan’s mission was not about standing on top of
the world but about using the platform of Mount Everest to
spread the message of hope. He wanted people to get fit, get
active, and live healthier, happier, more meaningful lives.
After Dr. Egan’s death, Elia Saikaly wanted to ensure that
his message could be spread and his legacy would live on.
©IGARSS 2023

In 2007 he started a project to create a real-time experience


for pupils from a remote environment and to connect it back
to their curriculum. FindingLife (https://findinglife.ca/) cre- FIGURE 18. Plenary session speech by Elia Saikaly, award-winning
ates immersive educational experiences by connecting students adventure filmmaker based in Ottawa, Canada.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 123


©IGARSS 2023

(a) (b)

FIGURE 19. AdCom participants at the GRSS AdCom meeting prior to IGARSS 2023: (a) AdCom at large. (b) Executive AdCom members.

the summer season, Elia Saikaly put the original team one in the attempt of climbing Mount Everest. The other
back together because the misfortune of their friends was side of the problem is not only the foreigner’s ambitions,
a shock for all of them, and this time they were able to but also the ambitions of the locals, for whom being in
find the bodies of the three rope companions from Paki- business around Mount Everest is a way out of poverty.
stan, Chile, and Iceland. Now local people are trying to fight for insurance and re-
Elia Saikaly was on Mount Everest again this year. He tirement plans.
shared with us reflective thoughts about the reasons why Another issue with Mount Everest is the environmen-
people are dying on Mount Everest. The media oversim- tal disaster caused by the climbers. They leave their trash
plify the reasons and blame traffic as only this season 478 behind, which is very often the result of poor planning
foreign climbers received government permission to climb because no manpower is accounted for to leave no trace
Everest—this is more tourists than usual. Elia Saikaly in- behind. A further issue is the overcommercialization
terviewed mountaineers and families who had lost some- of Mount Everest with clients who are interested only in
speed records and multiple peaks. As people are hovered
by helicopters across the valley, more inexperience and
negligence can be observed. People on the inside of the
country have to change this, but as storytellers, bystanders
and supporters can also help get the word out to finally
find positive solutions.
At the end of his speech, Elia Saikaly addressed the ef-
fects of climate change. The ice and snow layer is also get-
ting thinner on Mount Everest. He compared a photograph
taken by Sir George Mallory in 1921 from a glacier in Tibet
and another from 2009 taken from the same position. Ac-
cording to calculations, the glacier is 300 feet less dense.
Finally, Elia Saikaly reminded us of the responsibility of
the storyteller. Everyone can communicate their message
and share their work with the world, and the possibilities
are unlimited.

FUTURE IGARSS CONFERENCES


©IGARSS 2023

The GRSS AdCom met on 14–15 July 2023, just before


IGARSS. In this meeting all of the Society’s operational and
FIGURE 20. Core members of the IGARSS 2023 organizing and technical issues were discussed and main decisions were
supporting team (from left to right): Mehmet Ogut, Nathan taken. The 2023 members of the GRSS AdCom are shown
Longbotham, Sidharth Misra, Fairouz Stambouli, David Kunkee, in Figure 19. The road map for future IGARSS conferences
Shannon Brown, Joan Francesc Munoz-Martin, Rashmi Shah, Javier was confirmed, and a decision was made for IGARSS 2027:
Bosch-Lluis, Sharmilla Padmanabhan, Mariko Burgin, Maryam ◗◗ IGARSS 2024, Athens, Greece, 7–12 July 2024
Salim, Nereida Rodriguez-Alvarez, and Tianlin Wang. ◗◗ IGARSS 2025, Brisbane, Australia, 3–8 August 2025

124 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


◗◗ IGARSS 2026, Washington, D.C., USA, 19–24 July 2026
◗◗ IGARSS 2027, Iceland, 5–9 July 2027.
You are cordially invited to participate in future We want
to hear
IGARSS conferences, and we look forward to meeting
all of you at IGARSS 2024 in Athens, Greece, on 7–12
July 2024.
IGARSS 2023 in Pasadena was a great success and
from you!
surpassed all expectations. The networking achieved
in the IGARSS week was highly appreciated by the par-
ticipants. Such an outstanding event cannot happen
without the hard work of a large team of volunteers.
Figure 20 shows some of the key organizing team mem-
bers. Not shown in the picture are Chris Ruf, Ronny
Hänsch, Musafa Ustuner, Eric Loria, Alex Akins, Omkar
Pradhan, Mary Morris, Kazem Bakian Dogaheh, Alire- IMAGE LICENSED BY GRAPHIC STOCK

za Tabatabaeenejad, and also the following committee


members at-large: Fabio Pacifici, Paul Rosen, Saibun
Tjuatja, Karen St. Germain, Steve Reising, and Upendra Do you like what you’re reading?
Singh. In total, more than 30 members worked in the Your feedback is important.
core organizing team of IGARSS. Let us know—
send the editor-in-chief an e-mail!
ACKNOWLEDGMENT
The authors would like to thank Klara Antesberger for her
great support in the compilation of this article.
GRS

TAP.
CONNECT.
NETWORK.
SHARE.
Connect to IEEE–no matter where you are–with the IEEE App.
Stay up-to-date Schedule, manage, or Get geo and interest-based
with the latest news join meetups virtually recommendations

Read and download Create a personalized Locate IEEE members by location,


your IEEE magazines experience interests, and affiliations

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 125


TECHNICAL COMMITTEES
VICTORIA VANTHOF , HEATHER MCNAIRN ,
STEPHANIE TUMAMPOS , AND MARIKO BURGIN

Reinforcing Our Commitment:


Why DEI Matters for the IEEE GRSS

S hared beliefs and narratives have played a pivotal


role in fostering cooperation and progress in human
societies. The sense of belonging, common identity, and
sity, equity, and inclusion (DEI) lens and a facilitated
workshop on unconscious bias to the GRSS adminis-
trative governance.
connection among individuals have driven ground-
breaking technological and scientific advancements, DEFINING DIVERSITY, EQUITY, INCLUSION,
among which is the International Space Station (ISS). ACCESSIBILITY, AND BELONGING
Developed by five international space agencies, the DEI are terms often combined and used interchange-
ISS has hosted a wide range of scientific research and ably, but these concepts are distinct and uniquely im-
experimentation and astronauts from more than 19 portant. The concepts of belonging (B) and accessibility
countries. These diverse contributors share a common (A) are increasingly being associated with DEI efforts.
mission of space exploration, united in their pursuit of Diversity (D) is about the dimensions, qualities, and
bettering life on Earth. characteristics that define us [1]. Our identity is typically
Akin to the ISS, the IEEE Geoscience and Remote intersectional given that many traits define who we are
Sensing Society (GRSS) acknowledges the importance (race, ethnicity, religion, gender identity, sexual orienta-
of cultivating diverse and inclusive communities to tion, marital status, disability, age, parental status, etc.).
address complex global challenges that transcend geo- A commitment to equity (E) ensures that access, resourc-
graphical boundaries. Within the fields of geoscience es, and opportunities are provided for all to succeed [2].
and remote sensing, a diverse group of scientists and The terms equity and equality are not interchangeable
practitioners is essential in understanding and evalu- ­(Figure 1). Treating everyone the same (equality) does not
ating our interactions with the environment. This recognize an individual’s needs, in particular for those
diversity paves the way for comprehensive solutions and from underrepresented and historically disadvantaged
advancements that benefit the entire global community. populations. Inclusion (I) provides all with an opportuni-
However, the geoscience community at large faces chal- ty to sit at the table and to cocreate. A culture of inclusion
lenges in terms of diversity and inclusion, with under- embraces, respects, and values diversity [1]. Belonging (B)
represented groups such as women and people of color is the outcome of inclusion and signifies a sense of being
being insufficiently represented. This lack of diversity accepted and welcomed [3]. Inclusion without belonging
limits the range of perspectives and experiences within makes one feel like they must fit in with the dominant
the field. When diversity is lacking, certain individuals culture [3]. Accessibility (A) is the intentional design and
may struggle to find their place and may not feel fully construction of facilities; information and communica-
integrated into the community. tion technologies; programs; services; and policies so that
The GRSS is actively working to create a more equi- all can fully and independently benefit from them [4].
table and welcoming environment in the geosciences Diversity and, to a lesser extent, inclusion, equity, and
and among its members. Two recent advancements are accessibility, can be quantified and/or more easily ex-
the introduction of a committee focused on bringing pressed. Belonging is difficult to measure, but one senses
institutional change within the Society from a diver- almost immediately its presence or absence. For example,
a meeting accessible to attendees from a diverse commu-
Digital Object Identifier 10.1109/MGRS.2023.3303874
nity, where all are invited to the table, is not necessarily a
Date of current version: 19 September 2023 meeting where all feel welcomed, valued, and heard.

126 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


Organizations committed to advancing diversity, equi- ◗◗ Bring broader awareness of DEIAB by providing train-
ty, inclusion, accessibility, and belonging (DEIAB) stand to ing and resources to the GRSS leadership, portfolio
reap significant benefits. Those who view DEIAB efforts as leads, committee chairs, volunteers, and the broader
onerous underestimate the value of these benefits. A diver- GRSS membership.
sity of views, perspectives, talents, and life experiences im- ◗◗ Gather diversity data on the GRSS membership and
proves decision making and client services [3], and in fact, across GRSS committees, publications, initiatives,
research has demonstrated the following [5]: and events to track trends, determine gaps, and iden-
◗◗ Companies with more racial or gender diversity have tify ­opportunities.
more sales revenues, more customers, and greater profits ◗◗ Advance diversity and equity across GRSS committees,
(506 companies evaluated). publications, initiatives, and events and encourage
◗◗ Companies with more female executives were more prof- policies and practices that promote a sense of inclusion
itable (20,000 companies in 91 countries e­ valuated). and belonging.
◗◗ Management teams with a wider range of educa- ◗◗ Communicate the initiatives, innovations, and accom-
tional and work backgrounds produced more innova- plishments of the GRSS in DEIAB excellence.
tive ­products. The committee has initiated a review of GRSS diver-
In contrast, a toxic culture is 10× more important than sity data-gathering activities and a review of the Soci-
compensation in predicting employee turnover [6]. In a ety’s awards processes and procedures. A Diversity and
2021 report by McKinsey & Company on recruitment and
retention [7], 40% of employees (surveyed in five coun-
tries) indicated that they will likely leave their job in the
next three to six months. The top three reasons cited for
leaving included:
◗◗ did not feel valued by their organizations (54%)
◗◗ did not feel valued by their managers (52%)
◗◗ did not feel a sense of belonging at work (51%).

ADVANCING DEIAB IN THE GRSS


For a number of years, the GRSS has delivered important
initiatives to its members—and the broader scientific com-
munity—to advance DEI goals. These are led by our vi-
brant Young Professional (YP); Educational; Chapter; and
Inspire, Develop, Empower, and Advance (IDEA) commit-
tees. IDEA runs a Women Mentoring Women (WMW) pro- FIGURE 1. Equality versus equity. (Source: Interaction Institute for
gram, annually awards professional microgrants to support Social Change. Artist: Angus Maguire: https://interactioninstitute.
young minorities and women pursuing advanced degrees, org/ and http://madewithangus.com/.)
and delivers a Women in Africa webinar series.
In 2023, the GRSS president, Dr. Mariko Burgin, formed
a DEIAB Ad Hoc Committee. This committee compliments
ongoing GRSS activities (Figure 2) but with a mandate to
approach DEIAB efforts more holistically and to identify Educational
gaps and opportunities to inform all GRSS programs in- Young Activities
volved in advancing the Society’s DEIAB goals. This Ad Professional Chapter Activities
Hoc Committee is strategic, working to create institution-
al changes embedded in Society governance, policies, and
procedures. This committee is GRSS’s “agent of change”
for DEIAB and is striving to make advancements that will
live beyond any particular cohort of leadership and vol-
unteer core. IDEA DEIAB
The GRSS is committed to the goals of diversity, equity, Programmatic Strategic
and accessibility and strives to create spaces that are inclu- “Doing” “Agent of Change”
sive and where all feel that they belong. The newly formed
DEIAB Ad Hoc Committee has penned a draft strategic
plan, which lays out five short-term goals.
◗◗ Obtain support and participation of the GRSS leader-
ship, portfolio leads, committee chairs, and volunteers FIGURE 2. GRSS committees promote DEIAB across the Society
in advancing DEIAB across the Society. and the broader geoscience and remote sensing community.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 127


I­nclusion Chair was appointed for the GRSS 2023 flag- ity bias where one makes decisions as a result of personal
ship conference, the International Geoscience and Remote connections. According to feedback collected after the
Sensing Symposium (IGARSS), for the purpose of viewing workshop, some AdCom members felt more comfortable
the conference through a DEIAB lens. IGARSS 2023 pro- speaking about DEI and were eager to have additional in-
vided gender-neutral bathrooms and implemented both a teractive sessions to spot and mitigate unconscious bias. In
family room and a lactation room. These conference facili- addition, some AdCom members suggested that more time
ties are important steps forward in providing a facility and could have been dedicated to the workshop and that future
an experience that are more inclusive and accessible and workshops should include interactive sessions where situa-
where our broader geoscience and remote sensing commu- tions are enacted and interventions are explored. The GRSS
nity feels at home. leadership will consider feedback on this first introductory
workshop as it continues to develop awareness around DEI.
UNCONSCIOUS BIAS WORKSHOP WITHIN As SGO pointed out, the real work commences once the
THE GRSS LEADERSHIP workshop ends. It is important that Society leaders execute
As mentioned previously, companies and organizations good intervention strategies and encourage members to do
that invest in DEI training and workshops can reap benefits the same. In addition, familiarizing oneself with bias and
in terms of productivity and innovation, to name a few. recognizing that we all have biases encourages us to remain
Meanwhile fostering inclusion and belonging broadens open and impartial and to resist falling into assumptions.
the organization’s membership and enables the exchange Of course, the takeaway message from this workshop is to
of ideas and partnerships across the wider scientific com- practice and train ourselves to consciously identify biases.
munity. However, an important topic that needs a deeper Again, this workshop provided an introduction to un-
dive to advance DEI within groups is understanding that conscious bias for GRSS AdCom members. This founda-
our differences also come with biases. tion can be built on by members carrying this knowledge
As a step further toward DEI, the GRSS leadership ini- forward as the GRSS continues to develop strategies to ad-
tiated an Unconscious Bias Workshop that involved the vance the goals of DEIAB.
Society’s Administrative Committee (AdCom). This work-
shop was delivered by Felicia Jadczak of She+ Geeks Out CONCLUSION: THE GRSS DEI
(SGO) on 15 July, the last day of the GRSS AdCom meet- CONTINUES THE JOURNEY
ing in Pasadena, CA, USA. The workshop was attended by Just as the ISS fosters a sense of belonging and shared
45 AdCom members. During the workshop, the founda- purpose among its diverse crew, the GRSS recognizes the
tions of DEI were introduced, followed by a tool kit that vital role of DEIAB in cultivating an environment where
detailed the types of unconscious bias, how to spot these everyone feels welcomed. This sense of belonging will excel
biases, and steps on how to interrupt bias. in innovation to better understand our interactions with
Some of the AdCom members who participated in Earth’s ecosystems.
the workshop already had an understanding of DEI and The journey toward a more diverse and inclusive geosci-
its definition, while others were earlier in their journey. ence community within the GRSS does not end here; it is an
Meanwhile, the topic of unconscious bias was fairly new to ongoing commitment that requires collective effort. Initia-
almost everyone who joined the workshop. The main tives such as the DEIAB Ad Hoc Committee and the Uncon-
message presented in this workshop was that most of the scious Bias Workshop demonstrate ongoing efforts to create
biases we have are involuntary and, as such, are referred to an inclusive and welcoming environment for all members.
as unconscious biases. It is essential to note that unconscious By actively addressing and discussing unconscious biases in
bias is a natural human behavior—whether we like it or not. the workshop, the GRSS is emphasizing the need to open
Unconscious bias can be similarly compared to the well- the dialog and continue to engage on the topic. Moreover,
known idiom, “Birds of the same feather flock together.” the DEIAB Ad Hoc Committee will work to create institu-
This means that we chose to unintentionally or intentionally tional changes embedded in Society governance, policies,
move toward those with whom we have more in common. and procedures.
Because of this, we learn stereotypes that automatically drive If, after reading this, you have any input on how the
and affect our understanding and interactions with other GRSS can serve you better as a member and advance
people. It is important to acknowledge that we can never inclusiveness and belonging in the Society, please e-mail
eliminate bias. If we can recognize and familiarize ourselves president@grss-ieee.org.
with these unintentional biases, it will lead us to initiate
strategies to interrupt such biases. REFERENCES
The 2.5-h Unconscious Bias Workshop ignited an en- [1] “Diversity defined,” Canadian Centre for Diversity and Inclu-
gaging discussion with the AdCom members—from the sion, Toronto, ON, Canada, 2023. [Online]. Available: https://
confirmation bias where one tends to cherry-pick informa- ccdi.ca/our-story/diversity-defined/
tion and interpret it to confirm one’s beliefs; to the gender [2] “Diversity, equity, and inclusion definitions,” Univ. of
bias where one prefers a gender over another; to the affin- ­Washington, Seattle, WA, USA, 2023. [Online]. Available: https://

128 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


www.washington.edu/research/or/office-of-research-diversity [Online]. Available: https://hbr.org/2016/09/diverse-teams-feel
-equity-and-inclusion/dei-definitions/ -less-comfortable-and-thats-why-they-perform-better
[3] “The Advocates’ Journal,” vol. 39, no. 4, pp. 1–40, Spring 2021. [6] “Toxic culture is driving the great resignation,” MIT Sloan Man-
[Online]. Available: https://www.yumpu.com/en/document/ agement Review, Cambridge, MA, USA, Jan. 2022. [Online]. Avail-
read/65310153/the-advocates-journal-spring-2021 able: https://sloanreview.mit.edu/article/toxic-culture-is-driving
[4] “Diversity, equity, inclusion and accessibility: A foundation -the-great-resignation/
for meaningful change,” U.S. Department of Labor, Wash- [7] “Great attrition’ or ‘great attraction’? The choice is yours.”
ington, DC, USA, 2022. [Online]. Available: https://blog.dol. McKinsey. Accessed: Sep. 8, 2021. [Online]. Available: https://
gov/2022/02/22/diversity-equity-inclusion-and-accessibility www.mckinsey.com/capabilities/people-and-organizational
-a-foundation-for-meaningful-change -performance/our-insights/great-attrition-or-great-attraction
[5] “Diverse teams feel less comfortable — And that’s why they per- -the-choice-is-yours
form better,” Harvard Business Review, Boston, MA, USA, 2016. GRS

MANIL MASKEY , GABRIELE CAVALLARO , DORA BLANCO HERAS , PAOLO FRACCARO,


BLAIR EDWARDS , IKSHA GURUNG , BRIAN FREITAG , MUTHUKUMARAN RAMASUBRAMANIAN ,
JOHANNES JAKUBIK , LINSONG CHU, RAGHU GANTI, RAHUL RAMACHANDRAN ,
KOMMY WELDEMARIAM, SUJIT ROY , CARLOS COSTA, ALEX CORVIN , AND ANISH ASTHANA

A Summer School Session on Mastering Geospatial


Artificial Intelligence: From Data Production to Artificial
Intelligence Foundation Model Development and
Downstream Applications

I n collaboration with IBM Research, the NASA Inter-


agency Implementation and Advanced Concepts Team
­(IMPACT) organized a specialized one-day summer school
in HDCRS interdisciplinary research. The session, orga-
nized by IBM Research and NASA IMPACT, was a crucial
component of the four-day summer school program [1], ac-
session focused on exploring the topic of data science at cording to the HDCRS chairs. This article offers a succinct
scale (see Figure 1). This session was a part of the “High overview of the session.
Performance and Disruptive Computing in Remote Sens- As the organizers of the summer school program,
ing” summer school hosted by the University of Iceland we are delighted with our successful collaboration
from 29 May to 1 June 2023 in Reykjavik, Iceland [1]. This with IBM Research and NASA. In the contempo-
marked the third edition of the school organized by the rary era, the unprecedented vast volume of data
High Performance and Disruptive Computing in Remote essential for addressing RS challenges necessitates
Sensing (HDCRS) Working Group of the IEEE Geoscience the application of data science expertise. The spe-
and Remote Sensing Society’s (GRSS’s) Earth Science Infor- cialized data science session offered participants a
matics (ESI) Technical Committee (TC). The school aimed deep understanding of data science at scale, span-
to provide participants with a comprehensive understand- ning the data lifecycle, underlying systems, and
ing of contemporary advancements in high-performance research. This partnership enriched the HDCRS
computing (HPC), machine learning (ML), and quantum program significantly, leaving participants inspired
computing methodologies applied to remote sensing (RS). and informed.
Additionally, it created a networking platform for students —HDCRS chairs.
and young professionals to interact with leading research- Data science has evolved into a diverse field that ex-
ers and professors in RS, thereby promoting collaboration tends beyond the use of ML and artificial intelligence
(AI) algorithms. In practice, it constitutes a unique sci-
Digital Object Identifier 10.1109/MGRS.2023.3302813
entific discipline, marked by complex processes and nu-
Date of current version: 19 September 2023 anced intricacies. Professionals working in data-intensive

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 129


­ omains like Earth science and RS need to grasp these
d ◗◗ to demystify data science at scale and equip participants
layers of data science to excel in their respective fields. with the necessary skills and knowledge to tackle real-
To encourage broader participation in scientific world data challenges.
research, NASA has launched an open science initiative
[2]. This initiative aims to facilitate access to open data, DATA SCIENCE LANDSCAPE
computing platforms, and training opportunities for Data science encompasses a wide-ranging concept that
the next generation of scientific professionals, includ- entails principles and practices spanning data collection,
ing those without advanced HPC programming skills. storage, integration, analysis, inference, communication,
In alignment with the open science visions of both and ethics [3]. These fundamental principles are crucial
NASA and IBM, the summer school seeks to promote for navigating the data-driven landscape of the big data
the broader adoption of data science applications. With era, requiring a diverse range of skills and expertise. Data
a focus on geospatial artificial intelligence, it incorpo- engineers play a crucial role in building and maintaining
rates open science principles. the infrastructure necessary for handling large volumes
A successful data science team requires a diverse range of data. They design and develop data pipelines, ensuring
of expertise that is unlikely to be possessed by a single data quality, reliability, and scalability. Their expertise lies
individual or a group of individuals focused on a specific in managing data storage systems and integrating diverse
aspect of data science. Additionally, the expertise and re- data sources. Data analysts focus on examining datasets,
sources needed for such a team are typically not concen- applying statistical techniques, and identifying patterns
trated within a single institution. Therefore, to conduct and trends. ML engineers specialize in developing and
this session, we fostered collaboration across multiple implementing ML models and algorithms. They train
sectors, including government, industry, and academia. models, fine-tune hyperparameters, and deploy solutions
By leveraging the collective knowledge and resources that can analyze data, make accurate predictions, and fa-
from these different entities, we were able to create a cilitate decision-making tasks. Data visualization experts
comprehensive and impactful learning experience for the have expertise in presenting data in visually appealing and
participants. intuitive ways. They use various visualization techniques
and tools to communicate complex information effectively,
OBJECTIVES OF THE SESSION transforming data into insightful visual representations for
The objectives of the session were better understanding and interpretation. Domain experts
◗◗ to offer participants an in-depth understanding of the bring their in-depth knowledge of specific domains to the
complex data lifecycle and the underlying data systems data science process. They ask relevant questions, identify
◗◗ to explore the research lifecycle and its practical imple- meaningful variables, and provide contextual understand-
mentation ing for data analysis. Their expertise enhances the accuracy
◗◗ to provide participants with comprehensive knowledge and relevance of data-driven solutions. By combining the
about the rapid advancements in AI, with a particular skills and perspectives of these diverse skills, data science
emphasis on the foundational models teams can address complex problems and unlock valuable

Data System
Applications

Landsat Sentinel
Unsupervised Foundation
Fine-Tuning
Learning Models

Harmonization
Analysis Segmentation
(NDVI and Classification (Burn Scars,
so on) Flood) Other
Harmonized
Use Cases
Landsat and
Sentinel

FIGURE 1. The practical data science workflow used during the summer school session (green boxes indicate hands-on experience). NDVI:
Normalized Difference Vegetation Index.

130 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


insights from data. The composition of our organizing The participants were given valuable insights into the
team exemplifies this collaborative approach. pretraining process of the HLS-based GeoFM. Additionally,
they engaged in hands-on exercises focused on fine-tuning
DATA SCIENCE WITHIN GEOSPATIAL DOMAIN the HLS foundation model for specific use cases, specifi-
During the session, a key focus was placed on highlighting cally inland flood detection and burn scar delineation.
the significance of data science within the geospatial do- By actively participating in these activities, participants
main. Given NASA’s extensive collection of Earth-observa- gained practical experience in adapting and customizing
tion data, it plays a pivotal role in comprehending the Earth the GeoFM to address the real-world geospatial challenges
as a cohesive system. The sheer volume of data generated by related to flood detection and burn scar delineation. They
Earth science missions calls for innovative approaches to leveraged the watsonx.ai platform running on NASA’s Sci-
data analysis as conventional manual methods prove insuf- ence Managed Cloud Environment (SMCE) [8] environ-
ficient. To tackle these challenges, the exploration of technol- ment for fine-tuning these models.
ogies like commercial clouds and AI has become essential.
Notably, the session took advantage of these very tech- CHAPTER 4: INTERACTIVE EXPLORATION OF
nologies as integral components of the teaching curricu- FINE-TUNED MODELS
lum. By incorporating commercial clouds and AI, partici-
pants were exposed to cutting-edge tools and techniques PARTICIPANTS AND ORGANIZERS OF THE HDCRS
for effectively analyzing and extracting insights from the SUMMER SCHOOL AT THE UNIVERSITY OF ICELAND
vast and complex Earth-observation datasets. This hands- Chapter 4 of the session provided an interactive experience
on experience equipped them with the necessary skills to for participants, allowing them to delve into the fine-tuned
navigate the evolving landscape of data science within the GeoFM models. This hands-on section offered participants
geospatial domain. the opportunity to actively explore the capabilities of the
fine-tuned models and apply their outputs to new data.
STRUCTURE OF THE SESSION The results obtained were visually displayed in an inter-
The session was divided into four chapters, each focusing on active manner, leveraging map functionality within the
a specific aspect of data science and its practical applications: Jupyter environment.
This final chapter marked a significant milestone in the
CHAPTER 1: DATA PRODUCTION/PROCESSING participants’ data science journey as it allowed them to
In this chapter, participants delved into large-scale data complete a full lifecycle of the data science process. By per-
harmonization and explored the Harmonized Landsat forming inferences on the GeoFM models using new data,
Sentinel-2 (HLS) [4] data as a live case study. The session participants gained valuable experience in deriving mean-
covered the use of cloud computing employed in large-scale ingful insights and developing real-world applications.
data production and processing.
INFRASTRUCTURE AND TOOLS
CHAPTER 2: DATA ANALYSIS The GeoFM was pretrained on IBM Research’s computing
Chapter 2 introduced participants to tools such as the NASA infrastructure. It utilized HLS datasets for the Continental
Fire Information for Resource Management System, which United States in 2017, incorporating six different spectral
utilizes HLS datasets and dynamic tiling services [5]. The bands. A family of the pretrained GeoFMs used consisted of
participants collaboratively developed Jupyter notebooks 100–300 million parameters.
for analysis and visualization of the HLS dataset, enabling Throughout the session, participants utilized the NASA
them to engage in interactive data analysis and explore real- Visualization, Exploration, and Data Analysis (VEDA)
world applications with societal implications. analytics platform [9] for HLS data access and analysis.
SMCE provided the necessary computing resources. The
CHAPTER 3: THEORY AND APPLICATION OF THE SMCE, operating in NASA’s Amazon Web Services envi-
GEOSPATIAL FOUNDATION MODEL ronment, facilitated easy access and rapid onboarding for
Chapter 3 of the session centered around exploring the external partners and collaborators. VEDA JupyterHub
theory and practical application of the geospatial founda- facilitated by the 2i2c [10] was used as a gateway to the
tion model (GeoFM), which was developed through a col- computing platform.
laboration between IBM Research and NASA IMPACT [6].
The theory segment provided participants with a compre- OUTCOMES
hensive understanding of foundation models [7], high- Thirty diverse students from four different continents par-
lighting their emergence as a transformative solution for ticipated in the summer school. This global representation
data-driven science. These models have proven instrumen- added to the richness and diversity of perspectives in the
tal in overcoming challenges such as the requirement for learning environment. The students hailed from various
large-scale labeled datasets and the need for generalization backgrounds and brought unique experiences to the table,
across multiple tasks. fostering a collaborative and multicultural ­atmosphere.

SEPTEMBER 2023 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE 131


FIGURE 2. The participants and organizers of the HDCRS summer school at the University of Iceland.

The international nature of the participants further ing the TiTiler services and Christopher Phillips from the
­emphasized the global relevance and impact of data sci- University of Alabama in Huntsville for putting together
ence in today’s interconnected world. The session adopted training datasets for fine-tuning the GeoFM. The session re-
a practical approach, centering on real-world application sources are available at https://github.com/NASA-IMPACT/
scenarios. The participants actively engaged in hands-on summer-school-2023. The GeoFM, fine-tuned models, and
exercises and practical demonstrations. The session aimed data are openly available at https://huggingface.co/ibm
to foster collaboration and create a platform for exchanging -nasa-geospatial.
ideas among participants. The success of the session high-
lighted the value of collaboration, hands-on experience, REFERENCES
and the application of foundational models in geospatial [1] “HDCRS summer school 2023.” GRSS-IEEE. Accessed: Jul. 16,
analysis. Furthermore, the hands-on interactions enabled 2023. [Online]. Available: https://www.grss-ieee.org/community/
the organizers to document the necessary steps to be taken groups-initiatives/high-performance-and-disruptive
from data science tools, platforms, and compute infrastruc- -computing-in-remote-sensing-hdcrs/hdcrs-summer
ture requirements, enabling large-scale collaborative proj- -school-2023/
ects to analyze Earth-observation data (see Figure 2). [2] “Open-source science initiative.” NASA Science. Accessed:
Jul. 20, 2023. [Online]. Available: https://science.nasa.gov/
CONCLUSION open-science-overview
The specialized session on data science organized by NASA, [3] National Academies of Sciences, Engineering, and Medicine,
IBM Research, and the University of Iceland showcased the Data Science for Undergraduates: Opportunities and Options.
power of collaboration, hands-on experience, and the ap- Washington, DC, USA: The National Academies Press, 2018.
plication of foundational models in geospatial analysis. As [4] “HLS overview,” United States Geological Survey, Valley Drive
data science continues to shape the future, it is crucial to Reston, VA, USA. Accessed: Jul. 16, 2023. [Online]. Available: https://
empower professionals with the necessary skills and knowl- lpdaac.usgs.gov/data/get-started-data/collection-overview/
edge to harness the potential of data-driven decision mak- missions/harmonized-landsat-sentinel-2-hls-overview/
ing. The session stands as an important milestone in this [5] “eoAPI-raster.” Cloudfront.net. Accessed: Jul. 16, 2023. [Online].
journey, reflecting a commitment to enhance the capacity Available: https://d1nzvsko7rbono.cloudfront.net/docs
for data science and push the boundaries of what is achiev- [6] “IBM’s new geospatial foundation model.” IBM. Accessed:
able. Due to the success of the session, we have plans to Jun. 30, 2023. [Online]. Available: https://research.ibm.com/
continue organizing the session with a data science theme blog/geospatial-models-nasa-ai
during future summer schools. [7] R. Bommasani et al., “On the opportunities and risks of
foundation models,” 2021, arXiv:2108.07258.
ACKNOWLEDGMENT [8] Science Managed Cloud Environment. Accessed: Jul. 17,
Our heartfelt gratitude goes to the ESI TC for its generous 2023. [Online]. Available: https://smce.nasa.gov/
sponsorship of the workshop and the GRSS for its ongoing [9] M. Maskey et al., “Dashboard for earth observation,” in Advances
support. Additionally, we would like to express our special in Scalable and Intelligent Geospatial Analytics, 1st ed. Boca Raton,
thanks to the University of Iceland for hosting the event FL, USA: CRC Press, 2023.
and for the invaluable assistance provided by the dedicated [10] “Interactive computing for your community.” 2i2c. Accessed:
volunteers. A special mention goes to Sean Harkins from Jul. 17, 2023. [Online]. Available: https://2i2c.org/
Development Seed for his invaluable support in facilitat- GRS

132 IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE SEPTEMBER 2023


CALL FOR PAPERS
IEEE Geoscience and Remote Sensing Magazine

Special issue on “AI meets Remote


Sensing Image Understanding”
Guest Editors
Prof. Gong Cheng, Northwestern Polytechnical University, China (gcheng@nwpu.edu.cn)
Prof. Jun Li, China University of Geosciences, China (lijuncug@cug.edu.cn)
Prof. Xian Sun, Chinese Academy of Sciences, China (sunxian@aircas.ac.cn)
Ass. Prof. Zhongling Huang, Northwestern Polytechnical University, China (huangzhongling@nwpu.edu.cn)
Prof. Mihai Datcu, University POLITEHNICA of Bucharest (UPB), Romania (mihai.datcu@upb.ro)
Prof. Antonio Plaza, University of Extremadura, Spain (aplaza@unex.es)

Remote sensing (RS) image understanding aims at extracting valuable information and acquiring knowledge from remotely sensed
data, and artificial intelligence (AI) plays a significant role. With the increased data availability and the development of techniques for
data interpretation –particularly, deep learning (DL) techniques– the past few years have witnessed a tremendous growth of
research efforts focused on the visual interpretation of remote sensing images. Such techniques have made significant
breakthroughs in multiple domains, such as scene classification, object detection, feature extraction and recognition, and land-
use/land-cover mapping, to name a few. Nevertheless, there are still several challenges in this field, mostly related to the robustness
and transferability of interpretation approaches, the efficient perception and understanding of RS mages, the effective fusion and
utilization of multi-modal RS data, etc. This special issue that is aimed at investigating new techniques, algorithms and architectures
that can be used to overcome the above-mentioned challenges and bring together the state-of-the-art research in this field.

This special issue accepts review/tutorial papers on the following topics:


• Foundation models for downstream tasks of RS image understanding
• Robust AI architectures for detection, segmentation, and recognition in RS images
• Integration of geographical knowledge and deep neural networks in RS image interpretation
• Development of high-quality and large-scale benchmarks with multi-source data for AI understanding of RS images
• Effective processing using AI of multi-modal/sensor RS data
• Applications, such as smart agriculture, change detection and understanding of RS time series , environmental monitoring and sustainable
development, urban/rural monitoring and assessment, natural disaster warning and management

Articles submitted to this special issue of the IEEE Geoscience and Remote Sensing Magazine must contain significant
relevance to geoscience and remote sensing and should have noteworthy tutorial/review value. Selection of invited papers
will be done on the basis of 4-page White papers, submitted in double-column format. These papers must discuss the
foreseen objectives of the paper, the importance of the addressed topic, the impact of the contribution, and the authors’
expertise and past activities on the topic. Contributors selected on the basis of the White papers will be invited to submit
full manuscripts. Manuscripts should be submitted online at http://mc.manuscriptcentral.com/grsm using the Manuscript Central
interface. Prospective authors should consult the site http://ieeexplore.ieee.org/servlet/opac?punumber=6245518 for guidelines and
information on paper submission. Submitted articles should not have been published or be under review elsewhere. All submissions
will be peer reviewed according to the IEEE and Geoscience and Remote Sensing Society guidelines.

Special Issue tentative schedule:


December 31, 2023 White paper submission deadline
January 31, 2024 Invitation notification
March 31, 2024 Full paper submission deadline
July 31, 2024 Review notification
October 31, 2024 Revised manuscript due
January 31, 2025 Final acceptance notification
February 28, 2025 Final manuscript due
June 2025 Publication date

Digital Object Identifier 10.1109/MGRS.2023.3307828


Share Your
Preprint Research
with the World!

TechRxiv is a free preprint server for unpublished


research in electrical engineering, computer
science, and related technology. Powered by
IEEE, TechRxiv provides researchers across a
broad range of fields the opportunity to share
early results of their work ahead of formal
peer review and publication.

BENEFITS:
• Rapidly disseminate your research findings

• Gather feedback from fellow researchers

• Find potential collaborators in the


scientific community

• Establish the precedence of a discovery

• Document research results in advance


of publication

Upload your unpublished research today!

Follow @TechRxiv_org
Learn more techrxiv.org Powered by IEEE

You might also like