Professional Documents
Culture Documents
A R T I C L E I N F O A B S T R A C T
Keywords: Earth observation (EO) technologies, such as optical imaging and synthetic aperture radar (SAR), provide
Multimodal remote sensing excellent means to continuously monitor ever-growing urban environments. Notably, in the case of large-scale
Disaster damage mapping disasters (e.g., tsunamis and earthquakes), in which a response is highly time-critical, images from both data
Deep convolutional neural network
modalities can complement each other to accurately convey the full damage condition in the disaster aftermath.
However, due to several factors, such as weather and satellite coverage, which data modality will be the first
available for rapid disaster response efforts is often uncertain. Hence, novel methodologies that can utilize all
accessible EO datasets are essential for disaster management. In this study, we developed a global multimodal
and multitemporal dataset for building damage mapping. We included building damage characteristics from
three disaster types, namely, earthquakes, tsunamis, and typhoons, and considered three building damage cat
egories. The global dataset contains high-resolution (HR) optical imagery and high-to-moderate-resolution SAR
data acquired before and after each disaster. Using this comprehensive dataset, we analyzed five data modality
scenarios for damage mapping: single-mode (optical and SAR datasets), cross-modal (pre-disaster optical and
post-disaster SAR datasets), and mode fusion scenarios. We defined a damage mapping framework for semantic
segmentation of damaged buildings based on a deep convolutional neural network (CNN) algorithm. We also
compared our approach to another state-of-the-art model for damage mapping. The results indicated that our
dataset, together with a deep learning network, enabled acceptable predictions for all the data modality sce
narios. We also found that the results from cross-modal mapping were comparable to the results obtained from a
fusion sensor and optical mode analysis.
1. Introduction the damage condition after a large-scale disaster (Bai et al., 2017;
Twumasi et al., 2019a; Ge et al., 2020).
Geophysical disasters such as earthquakes and tsunamis are rare In this study, we construct a novel global multimodal and multi
events that can devastate large urban environments, causing enormous temporal remote sensing dataset from notable earthquakes, tsunamis,
human and economic losses. From 1998–2017 alone, these two types of and typhoon disasters, together with the corresponding reference data of
events were responsible for approximately 750,000 deaths worldwide damaged buildings. The dataset was collected from different optical
(Wallemacq and Below, 2018). Detailed information about the extent sensors as well as diverse microwave ranges of the SAR data. Consid
and level of structural damage is therefore essential to first responders ering that earthquakes and tsunami events are rare and that affected
for adequately conducting rescue and relief operations. In this context, areas often become isolated because of access difficulties, collecting
earth observation (EO) technologies, such as optical imaging and syn reliable ground-truth information is highly expensive. Thus, the
thetic aperture radar (SAR), can provide complementary information on advantage of the proposed framework is its introduction of a unique EO
* Corresponding author.
E-mail address: bruno.adriano@riken.jp (B. Adriano).
https://doi.org/10.1016/j.isprsjprs.2021.02.016
Received 13 July 2020; Received in revised form 4 February 2021; Accepted 19 February 2021
Available online 17 March 2021
0924-2716/© 2021 The Author(s). Published by Elsevier B.V. on behalf of International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS). This is an
open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
building damage dataset (BDD) involving mapping. By using this data the success of the developed techniques mainly depends on i) the quality
set, diverse data availability scenarios can be analyzed, such as single- of the validation data, which is often limited, and ii) an appropriated set
mode, cross-modal, and mode fusion data scenarios, for building dam of pre- and post-disaster images. These factors have influenced the
age recognition. Furthermore, we introduce a damage mapping frame design of previous methods, limiting their applicability to specific
work for the classification of building damage from space using modern characteristics of datasets and affected areas. These facts make field
deep learning algorithms. The main contribution of this work is surveys the gold standard for obtaining precise information on the
threefold: damage condition. Nevertheless, the large EO datasets acquired from
previous events present a valuable resource for developing advanced
• We construct a unique global multitemporal and multimodal EO mapping frameworks to assess future disasters.
dataset together with labeled building footprints from large-scale Recently, deep learning algorithms, such as deep convolutional
earthquake and tsunami events worldwide. neural networks (CNNs), have enabled considerable progress in solving
• We propose a damage mapping framework that integrates remote computer vision problems, such as object classification and image seg
sensing and deep learning to classify the level of building damage mentation (Liu et al., 2019). Due to this success, CNN algorithms have
considering several data availability scenarios. been used for damage recognition using remote sensing datasets. Recent
• We conduct extensive experiments and evaluate the performance of applications verify the potential of this technology (Bai et al., 2018; Ma
the proposed framework with other state-of-the-art deep learning et al., 2019; Miura et al., 2020). However, CNN models require a large
approaches used for damage recognition. number of training images with corresponding high-quality labeled
data. For this reason, research groups have recently collected large
datasets for a variety of tasks, such as common object detection and
1.1. Related work recognition (Lin et al., 2014; Everingham et al., 2015).
In the field of disaster management, the xView2 Challenge (Gupta
Building damage mapping using remote sensing datasets has been et al., 2019), in collaboration with several agencies, introduced one of
extensively studied. We can broadly divide mapping frameworks based the first large-scale multitemporal datasets for building damage map
on the EO data used. The first generation of moderate-resolution optical ping. This dataset contains VHR optical imagery acquired from several
sensors (e.g., https://www.nasa.gov/mission_pages/landsat/main/ disasters, such as floods, wildfires, and earthquakes. Although weather
index.htmlLandsat and the https://asterweb.jpl.nasa.gov/Advanced conditions may constrain the applicability of this dataset to future di
Spaceborne Thermal Emission and Reflection Radiometer (ASTER)) sasters, the research community has provided exceptional contributions
allowed a general interpretation of the structural damage in affected in the forms of CNN architecture design and training strategies through
areas (Yusuf et al., 2001; Altan et al., 2001; Yamazaki and Matsuoka, this competition. Recently, a multimodal optical and SAR dataset was
2007). The follow-up generation of high-resolution (HR) optical sensors also presented in the SpaceNet-6 Challenge (Shermeyer et al., 2020).
enabled detailed damage recognition. Using pixel- or object-based The goal of this competition was to extract only building footprints in a
change detection techniques, several degrees of damage could be cross-modal data scheme. In this paper, we also introduce a novel
adequately classified for a single building (Freire et al., 2014). These multimodal and multitemporal BDD. Here, we seek to satisfy all possible
works were successfully applied to large disasters. For instance, data conditions that first responders face in emergency response
following the 2008 Wenchuan earthquake, Tong et al. (2012) extracted applications.
individual collapsed buildings using 3D geometric changes (building The rest of the paper is organized as follows. Section 2 details the
heights) observed in pre- and post-event IKONOS images. remote sensing dataset and the processing and generation of the labeled
Very-high-resolution (VHR) optical imagery is also used to visually building masks for segmentation of damaged buildings. Section 3 pre
interpret the damage condition after disasters. Although an experienced sents our proposed methodology, in which details of the CNN archi
human interpreter can provide very reliable information, these ap tecture and training settings are described. Sections 4 and 5 show the
proaches are time consuming. Therefore, they are mainly utilized by corresponding experimental results and the discussion, respectively.
large international organizations, such as https://www.unitar.org/ Finally, Section 6 provides the concluding remarks and outlook on our
mapsthe United Nations Institute for Training and Research Operational proposed dataset.
Satellite Applications Programme (UNOSAT) and https://emergency.
copernicus.eu/Copernicus Emergency Management Service (EMS). 2. Materials
Furthermore, analysis using optical imaging is often hampered by
weather conditions when clouds could cause occluded acquisitions. For detailed classification of damaged buildings, HR or moderate-
SAR, which can penetrate clouds, has gained popularity for use in resolution remote sensing imagery is necessary. Optical imaging en
disaster response tasks. Similar to optical imagery, the details of the ables straightforward interpretation of affected areas; however, weather
analysis are dependent on the SAR pixel resolution. The latest SAR and daylight conditions might limit cloudless acquisitions. Recently,
sensors, based on high-frequency X- and C-bands, can detect specific microwave SAR data, with nearly all-weather observation capabilities,
geometric features from built-up areas (Ferro et al., 2013; Rossi and have become an essential tool in rapid disaster response efforts. In this
Eineder, 2015). Polarimetric SAR is also used to extract information context, we have processed optical and SAR imagery from large-scale
with different SAR scattering mechanisms that can be linked to the de earthquakes, tsunamis, and typhoon disasters together with the corre
gree of building damage. For instance, Yamaguchi (2012) demonstrated sponding recorded building damage. This BDD, to the best of the au
that horizontal-horizontal (HH)-polarized double-bounce data provide thors’ knowledge, represents the first multimodal and multitemporal EO
better features for analyzing collapsed structures. A combination of dataset for disaster management research.
optical and SAR data was also extensively studied. Brunner et al. (2010)
used pre-disaster optical images to extract geometric parameters of 2.1. Disaster events
isolated collapsed buildings, and then, damage grading was performed
by comparing a simulated post-event SAR scene, the collapsed structure, For this study, we select and introduce the disasters that have a full
and the actual HR post-disaster SAR data. set of multitemporal and multimodal datasets. Fig. 1 shows the location
Dong and Shan (2013), Plank (2014), Ge et al. (2020), Twumasi et al. of the events considered in this study. Our dataset is composed of one
(2019), Koshimura et al. (2020) presented comprehensive reviews of typhoon, six earthquakes, and two tsunami disasters. Notably, only the
remote sensing for disaster mapping. They concluded that although 2011 Tohoku tsunami and the 2015 Kumamoto earthquake occurred in
considerable progress has been made in damage recognition from space, two distant cities in the same country. The rest of the events impacted
133
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
Fig. 1. Location of the catastrophic earthquake, tsunami, and typhoon events used to construct the multitemporal and multimodal remote sensing dataset for
building damage mapping.
134
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
Fig. 2. Data acquisition times (the time difference with respect to the event origin time is in days).
2.3. Synthetic aperture radar damage classifications were constructed through field survey campaigns
by the Japan-Peru Center for Earthquake Engineering Research and
The almost all-weather acquisition capabilities of SAR sensors Disaster Mitigation (CISMID), the Ministry of Land, Infrastructure,
represent an advantage compared to optical imaging. To complement Transport and Tourism of Japan (MLIT), and the Geospatial Information
the optical dataset, we also collected a set of pre- and post-event SAR Authority of Japan (GSI), respectively (Table 2).
data for all events included in this study. Similar to the optical dataset, Visually interpreting HR optical images also provides adequate in
several commercial sensors provided the SAR information (Table 2). formation on the overall damage (Yamazaki et al., 2005; Koshimura
Moreover, to take advantage of SAR data over built-up areas (Yama et al., 2009; Gokon and Koshimura, 2012; Mas et al., 2015). However,
guchi, 2012; Ferro et al., 2013), we selected the HH polarization scenes the details of damage interpretation are often limited because of the
for all events. The StripMap (SM) acquisition configuration of the almost nadir-looking nature of optical sensors. For this study, we
TerraSAR-X platform (managed by the German Aerospace Center) and downloaded the visual damage interpretation for the 2010 Haiti earth
the COnstellation of small Satellites for the Mediterranean basin quake, 2015 Kathmandu earthquake, 2017 Puebla earthquake, and
Observation (COSMO)-SkyMed platform (managed by the Italian Space 2017 Kermanshah earthquake from UNOSAT. In the case of the 2013
Agency) capture HR X-band data with slant range and azimuth resolu Haiyan typhoon, the building damage interpretation was conducted by
tions of approximately 1.2 m and 3.3 m, respectively. Additionally, the the Japan International Cooperation Agency (JICA) for Tacloban city.
SM model of the Advanced Land Observing Satellite (ALOS)-2 platform Finally, the Copernicus EMS conducted a visual analysis of building
captures L-band data with pixel spacings of 2.5 m and 3.15 m for the damage following the 2018 Palu tsunami (Table 2).
ultrafine and high-sensitivity configurations. In the case of the field survey, vector files of building polygons were
Several preprocessing techniques were applied to the SAR dataset. In available for each event. However, visually interpreted data provide
the case of TerraSAR-X, all the SAR scenes were provided as enhanced only the point location of the damaged building. In such cases, we used
ellipsoid corrected (EEC) products. Accordingly, to obtain the geocoded the building polygon layer of the https://www.openstreetmap.org/
intensity images, radiometric corrections were directly applied using the #map=5/35.588/134.380OpenStreetMap Project. Then, we overlapped
local incident angle provided by the sensor. In the case of COSMO- the polygon and point vector layers and assigned the interpreted damage
SkyMed and ALOS-2 scenes, all the images were acquired as single- using majority voting of the points located within a building polygon
look complex (SLC) products. Thus, we conducted almost identical (Adriano et al., 2019).
preprocessing steps for both sensors. The post-event SAR scenes were set Finally, to facilitate the comparison of the different damage classi
as secondary images and coregistered to the pre-event scenes (primary fication levels, we defined a three-category scale of building damage
images). In the multilook process, we used the minimum number of (Table 3). Our damage definition is based on the building structural
looks to obtain the highest pixel resolution for each sensor. Next, to condition after the disaster. As such, the highest degree of damage
suppress the SAR speckle noise, we applied the enhanced Lee filter (Park Destroyed is assigned when the structure is destroyed (i.e., collapsed) or
et al., 1999) to the radiometrically corrected intensity images using a washed away. Conversely, the Survived class is set when the building
moving window of 3 × 3 pixels. Finally, we used the 1-arcsecond structure appears to be undisturbed or when there is no visible damage
Shuttle Radar Topography Mission (SRTM) digital elevation model to the building rooftop. Notably, buildings could be apparently undis
(DEM) for orthorectification (terrain correction) and geocoding of all the turbed but severely damaged both inside and along the sidewalls. The
SAR scenes. Survived class is likely the most uncertain of the chosen three. This is a
standard limitation for visual and automatic interpretation of satellite
imagery. Finally, the middle damage category, moderately damaged
2.4. Generation of labels for damage categories (labeled as Moderate), corresponds to buildings showing visible changes
in their structure or surroundings. (see Table 4).
Collecting ground-truth data on the damage condition after a large- The European Macroseismic Scale (EMS-98) (Grünthal et al., 1998)
scale disaster strikes an urban environment is costly and time defines five-grades of structural damage for masonry and RC buildings
consuming. However, it is an essential task, particularly if we consider following a structural and nonstructural failure point of view. These
that detailed labeling of building units is necessary for adequately degrees range from minor cosmetic cracks to total or near collapse of the
planning rescue missions. Traditionally, two main approaches are
employed to gather adequate information for labeling building damage.
Table 3
The first method is field surveys conducted by experts, which often re
Descriptions of the damage classes defined in this study.
quires a large amount of economic and human resources (Masi et al.,
Damage level Buildings Description
2017; Monfort et al., 2019). This approach, however, can provide the
most reliable information by categorizing several degrees of building Destroyed 16,542 Completely collapsed or washed away
damage. In the cases of the 2007 Pisco earthquake, the 2011 Tohoku Moderate 28,112 Visible changes in and around the building
Survived 78,799 The building appears undisturbed
tsunami, and the 2016 Kumamoto earthquake, the initial building
135
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
Table 4 the optical and SAR images and corresponding labeled building foot
Comparison of the main features of the proposed BDD and other multitemporal prints included in the proposed BDD.
and multimodal datasets.
Dataset Multitemporal Multimodal Goal Target 3. Method
areas
BDD Yes Optical + Damage 9 We propose a framework for building damage mapping using CNNs
SAR assessment (Lecun et al., 1998). Given that adequate building locations are often
xView2 Yes Optical Damage 19 unknown in immediate emergency response, we use a CNN model for
assessment
multiclass semantic segmentation of damaged buildings. Fig. 5 depicts
SpaceNet6 No Optical + Footprint 1
SAR extraction the workflow of our framework. Here, the model extracts high-
dimensional features from each temporal dataset separately. Then, the
extracted feature vectors are used to map and grade the damaged
structure. Similarly, the Architectural Institute of Japan (AIJ) (Okada buildings in the affected area.
and Takai, 2000) defines seven categories, including “no damage” and
“complete collapse” as D0 and D6 scales, respectively. This detailed
3.1. Convolutional neural network model
damage categorization is often possible only through field surveys or
using high-level information collected from the ground, such as ground-
The CNN architecture is based on a U-Net model (Ronneberger et al.,
shot photographs. In contrast, the damage classes in the present study
2015). This architecture consists of an encoder-decoder design for se
are intended to align and summarize different criteria for categorizing
mantic segmentation. In these types of networks, high-dimensional
damage degrees according to the building overall integrity. Fig. 4 shows
feature vectors are extracted from input images by the decoder using
the comparison of the EMS-98 and AIJ scales and damage classes defined
successive blocks. In this work, we modify the encoder design by
in this study.
adopting two encoder streams to derive features from the pre- and post-
For instance, in the case of tsunami impact, debris or remaining
disaster datasets separately. By setting a change detection approach, the
water could be visible around buildings, and in earthquake-induced
encoders share their extracted features through concatenation and 2D
damage scenarios, part of the roof or a sidewall of a house could be
convolution operations (Fig. 5).
damaged. Table 3 summarizes the damage classes used in this study and
Each encoder stream is composed of five blocks. We use a set of two
the number of buildings included in each class. Fig. 3 shows examples of
3 × 3 2D convolutions, batch normalization, and the rectified linear
unit (ReLU) activation function in each block. A 2 × 2 max-pooling
downsampling operation (kernel = 2) follows each encoder block. The
output number of feature channels is doubled starting from 64 at the end
of each block.
The decoder part follows a mirror design of the encoders, where the
max-pooling operation is replaced by a 4 × 4 transpose convolution to
sequentially upsample the extracted feature vectors. Furthermore, to
recover the original pixel resolution of the input image and share the
information learned by the encoder with the decoder (Ronneberger
et al., 2015), the U-Net model uses skip connections that concatenate the
feature vectors of two corresponding encoder and decoder blocks.
As shown in Fig. 5, in our proposed network, the encoder part shares
the combined pre- and post-disaster features through the skip connec
tions. Moreover, considering that our BDD contains diverse building
structures and urban layouts, we also incorporate an additional atten
tion gate operation, which automatically learns to focus on different
target shapes by suppressing trivial regions in the input images Oktay
et al. (2018).
To obtain the desired number of classes N (Eq. (1)) from the last
decoder block, we apply a softmax activation function to the output
feature vector z. As a result, the network outputs N-channel vectors with
a predicted probability p(z) for each class i (Eq. (2)). Then, we compute
the final categorical output y for the given input images by maximizing
p(z) (Eq. (3)).
ezi
σ (z)i = (1)
∑
N
ezn
n=1
U-Net-based models are known for their ability to work with rela
Fig. 3. Examples from the multitemporal and multimodal BDD. The first, sec tively small training datasets (Ronneberger et al., 2015). Compared with
ond and third columns show damage from earthquake, tsunami, and typhoon other CNN-based architectures (e.g., fully convolutional networks), U-
disasters, respectively. Net is a lightweight design involving approximately 8 × 106 trainable
136
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
Fig. 4. Comparison of the building damage scales of the Architectural Institute of Japan (AIJ) Scale, the European Macroseismic Scale (EMS-98), and the damage
classes defined in this study.
Fig. 5. Overview of the building damage mapping framework based on the Attention U-Net segmentation model. Nint and Nout denote the number of channels of the
input and output images, respectively. In the case of this diagram, the input images correspond to the data Mode 2 scenario (multitemporal SAR imagery).
137
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
Table 5 the disaster or when cloud-free post-event optical images are not
Comparison of the damage classes defined in this available.
study and the xView2 Challenge. • The fifth mode is also a cross-modal scenario using pre-event optical
This study xView2 and SAR images and a single post-event SAR dataset. Here, in addi
Destroyed Destroyed
tion to the conditions of the fourth mode, we evaluate the strategy of
Moderate Minor damage using all accessible pre-disaster datasets. Thus, the scenario is an
Major damage extension of the third mode.
Survived No damage
Note that we do not include a cross-modal scenario of pre-disaster
SAR data and post-disaster optical images. The continuous acquisition
Table 5, to compare the performance of the winning solution with that of
of visual imagery (also SAR data) by several spaceborne platforms
our BDD, we re-arranged the xView2 damage classes into three cate
guarantees the accessibility of pre-disaster images for emergency
gories. Then, we used the average value of the weights for the Minor and
response in the case of future disasters. Thus, cross-modal scenarios of
Major damage classes for our BDD Moderate class. Finally, the weight
pre-event SAR and post-event optical images can quickly become the
values for the xView2 Destroyed and No damage were assigned to the
fourth and fifth modes.
BDD Destroyed and Survived classes, respectively.
The winning solution from the xView2 Challenge is also based on a
U-Net architecture. However, it follows a Siamese design, where two 4. Results
separate networks are used for each pre- and post-disaster dataset. Then,
the output features from the last decoder block are combined for the In this section, we describe the settings of our numerical experiments
building damage grading task. Further details regarding the optimiza and report the results from the proposed data modality scenarios
tion algorithm and loss function settings can be found at https://github. (Table 6) using our proposed mapping framework and the xView2-based
com/DIUx-xView/xView2_first_place. model.
138
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
Fig. 6. Error matrices computed over the test tiles for each data mode. (a) shows the results of using the winning solution of the xView2 Challenge, and (b) shows the
results of our proposed framework.
the xView2 winning model. This figure indicates that Mode 1 (fusion of general, our approach outperforms the winning solution from the
multitemporal optical and SAR imagery) and Mode 2 (only the optical xView2 Challenge. In this mode, our network produces a more balanced
images, similar to the xView-2 Challenge) achieve good performance for classification distribution for all classes, which is comparable to the
multiclass building damage grading. In particular, the accuracy ach results obtained in Mode 2 by the xView2-based model. Finally, the
ieved for detecting the Survived class is approximately 0.70 for these results of Mode 5 also indicate the superior performance of our frame
data modality scenarios. However, the performance decreases to work. Here, the accuracies for the Survived and Moderate classes are
average values of 0.50 and 0.35 for the Moderate and Destroyed classes, comparable to those obtained in Mode 4. In this last mode, however, our
respectively. Here, notable misclassifications emerge for these classes, network shows slightly better accuracy in predicting the Destroyed class.
where the CNN model confuses Destroyed as the Moderate class and Table 7 shows the overall quantitative evaluation (Fscore values and
Moderate as the Survived class. Additionally, the results for Mode 3 (only corresponding standard deviations) computed on the held-out test
the multitemporal SAR dataset) indicate a very low performance of the dataset for the two previous approaches for building damage mapping. It
winning solution. In this case, the network fails to classify all three shows that the two models, the xView2 model and our approach, are
damage categories accurately. In the case of Mode 4 (cross-modal relatively stable across all experiments. Here, we can see that our
mapping using pre-disaster optical images and post-disaster SAR data),
the results show that only the Survived class achieves a moderate per
Table 7
formance, with an accuracy of approximately 0.58. Finally, the fifth data
Accuracy assessment (Fscore ) calculated on the held-out test dataset.
mode also shows poor performance. In this case, only the Survived class
is slightly identified; however, the network also confuses the other Mode Method Classes Mean
damage classes with the Survived class, providing an overall incorrect Destroyed Moderate Survived
result for building damage mapping. 1 xView2 0.3912 ± 0.4396 ± 0.6197 ± 0.4835 ±
The performance of our mapping framework for all data modes is 0.0016 0.0079 0.0048 0.0048
summarized by normalized error matrices (Fig. 6(b)). Similar to the This 0.4173 ± 0.4567 ± 0.6143 ± 0.4961 ±
xView2 model results, these results also indicate that Mode 1 and Mode study 0.0053 0.0092 0.0092 0.0295
2 achieve the highest performance. Here, although the accuracy values 2 xView2 0.3795 ± 0.4229 ± 0.6173 ± 0.4732 ±
for the Survived class are slightly lower than those of the xView2 0.0114 0.0126 0.0067 0.0102
implementation, the misclassification in the three damage classes is This 0.4137 ± 0.4714 ± 0.6178 ± 0.5010 ±
study 0.0165 0.0026 0.0026 0.0300
considerably reduced. Notably, the use of multitemporal optical datasets
3 xView2 0.1403 ± 0.1768 ± 0.2799 ± 0.1990 ±
produces an accuracy of approximately 0.63 for classifying the inter
0.0247 0.0198 0.0351 0.0265
mediate damage level. This value represents the maximum accuracy
This 0.2044 ± 0.2242 ± 0.3114 ± 0.2467 ±
among all the classes. However, false classification also occurs in this study 0.0145 0.0271 0.0271 0.0156
data mode, where the network confuses the Destroyed class as the 4 xView2 0.1439 ± 0.2941 ± 0.5584 ± 0.3321 ±
Moderate class. The latter behavior is slightly reduced when the optical 0.0274 0.0714 0.0074 0.0354
and SAR datasets are fused (Mode 1), where the accuracy for the This 0.3215 ± 0.3936 ± 0.5775 ± 0.4309 ±
Destroyed class is marginally higher than that for Mode 2. study 0.0293 0.0101 0.0101 0.0371
The network trained on Mode 3, similar to the xView2 solution, 5 xView2 0.1013 ± 0.0845 ± 0.4242 ± 0.2033 ±
cannot correctly distinguish all the damage classes. However, the overall 0.0894 0.0344 0.0344 0.0392
accuracy values for all the damage levels are slightly higher than that of This 0.3333 ± 0.3774 ± 0.5692 ± 0.4266 ±
study 0.0043 0.0140 0.0186 0.0348
the xView2-based model. The results from Mode 4 indicate that, in
139
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
mapping framework achieves superior results for almost all the data characteristics as well as landscape configurations. Here, the reference
modality scenarios in comparison to the winning solution from the building damage masks were predominantly constructed from visual
xView2 Challenge. The proposed framework, which simultaneously interpretation analysis. Hence, the results obtained in this study repre
extracts and classifies building damage, gives the highest average scores sent a relative approximation of the human visual interpretation ability.
when optical datasets are involved. The same efficiency is also observed This is particularly the case for tsunami and typhoon events, where
for a cross-modal dataset (Mode 4). In the case of the SAR dataset (Mode damage to sidewalls, affected by water waves or wind speeds, is typi
3), the scores demonstrate that the xView2-based model and our model cally difficult to observe using optical imagery. This latter condition
cannot produce satisfactory results. However, the results for Mode 5 highlights the importance of including SAR data, which are character
(considered an extension of Mode 3) show a considerable improvement ized by their side-looking observation nature. Second, given that the
when a pre-disaster optical dataset is used for the input images. percentage of earthquake events is approximately 60%, our results may
be biased, achieving higher performance for this kind of disaster.
5. Discussion Nevertheless, we tried to reduce this bias by randomly splitting the
training, validation, and testing datasets for our experiments.
In this study, we created a unique multitemporal and multimodal Sudden disasters and weather conditions can limit our options for
BDD for three different types of large-scale disasters, namely, earth using an optimal set of remote sensing images for emergency response.
quakes, tsunamis, and typhoons. We put great effort into collecting For such circumstances, our BDD enables extensive analysis by consid
satellite images with a short time difference between the acquisitions ering several data availability scenarios, such as single-mode, cross-
before and after disasters. However, in areas with rapid urban de modal, and fusion-mode optical and SAR data. We presented a frame
velopments, such as demolition and construction of buildings, these work for building damage mapping based on a modern Attention U-Net
changes might have affected the overall assessment of the proposed architecture, showing that our approach satisfactorily classified building
damage mapping framework. We considered worldwide locations to damage into three levels for most dataset scenarios. However, damage
provide our BDD with high heterogeneity in terms of building classification using only SAR data remains challenging using both CNN
Fig. 7. Visual comparison of the damage classification results. (a) shows the results using the winning solution from the xView2 Challenge, and (b) shows the results
of our proposed framework.
140
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
141
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=6565347. doi: Mas, E., Bricker, J., Kure, S., Adriano, B., Yi, C., Suppasri, A., Koshimura, S., 2015.
10.1109/TGRS.2013.2271564. Survey and satellite damage interpretation of the 2013 Super Typhoon Haiyan in the
Brunner, D., Lemoine, G., Bruzzone, L., 2010. Earthquake Damage Assessment of Philippines. Nat. Hazards Earth Syst. Sci. 15, 805–816. https://doi.org/10.5194/
Buildings Using VHR Optical and SAR Imagery. IEEE Trans. Geosci. Remote Sens., nhess-15-805-2015.
48, 2403–2420. URL: http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm? Masi, A., Chiauzzi, L., Santarsiero, G., Liuzzi, M., Tramutoli, V., 2017. Seismic damage
arnumber=5411791 http://ieeexplore.ieee.org/document/5411791/. doi:10.1109/ recognition based on field survey and remote sensing: general remarks and examples
TGRS.2009.2038274. from the 2016 Central Italy earthquake. Nat. Hazards 86, 193–195. https://doi.org/
Celebi, M., Sahakian, V.J., Melgar, D., Quintanar, L., 2018. The 19 September 2017 M 10.1007/s11069-017-2776-8. URL: http://link.springer.com/10.1007/s11069-017-
7.1 Puebla-Morelos Earthquake: Spectral Ratios Confirm Mexico City Zoning. Bull. 2776-8.
Seismol. Soc. Am. 108, 3289–3299. https://doi.org/10.1785/0120180100 arXiv: Matsuoka, M., Estrada, M., 2013. Development of Earthquake-Induced Building Damage
https://pubs.geoscienceworld.org/bssa/article-pdf/108/6/3289/4562782/bssa- Estimation Model Based on ALOS/PALSAR Observing the 2007 Peru Earthquake.
2018100.1.pdf. J. Disast. Res. 8, 346–355. https://doi.org/10.20965/jdr.2013.p0346.
Dong, L., Shan, J., 2013. A comprehensive review of earthquake-induced building Miura, H., Aridome, T., Matsuoka, M., 2020. Deep Learning-Based Identification of
damage detection with remote sensing techniques. ISPRS J. Photogram. Remote Collapsed, Non-Collapsed and Blue Tarp-Covered Buildings from Post-Disaster Aerial
Sens. 84, 85–99. https://doi.org/10.1016/j.isprsjprs.2013.06.011. Images. Remote Sens. 12 https://doi.org/10.3390/rs12121924.
(, 2007EERI, 2007. The Pisco, Peru, Earthquake of August 15, 2007. Technical Report Miura, H., Midorikawa, S., Matsuoka, M., 2016. Building Damage Assessment Using
Earthquake Engineering Research Institute (EERI) Oakland, CA 94612, United States. High-Resolution Satellite SAR Images of the 2010 Haiti Earthquake. Earthquake
Everingham, M., Eslami, S.M.A., Van Gool, L., Williams, C.K.I., Winn, J., Zisserman, A., Spectra 32, 591–610. https://doi.org/10.1193/033014EQS042M.
2015. The Pascal Visual Object Classes Challenge: A Retrospective. Int. J. Comput. Monfort, D., Negulescu, C., Belvaux, M., 2019. Remote sensing vs. field survey data in a
Vision 111, 98–136. https://doi.org/10.1007/s11263-014-0733-5. post-earthquake context: Potentialities and limits of damaged building assessment
Ferro, A., Brunner, D., Bruzzone, L., 2013. Automatic Detection and Reconstruction of datasets. Remote Sensing Applications: Society and Environment, 14, 46–59. URL:
Building Radar Footprints From Single VHR SAR Images. IEEE Trans. Geosci. Remote doi: 10.1016/j.rsase.2019.02.003 https://linkinghub.elsevier.com/retrieve/pii/
Sens., 51, 935–952. URL: http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm? S2352938518303161. doi:10.1016/j.rsase.2019.02.003.
arnumber=6256716. doi:10.1109/TGRS.2012.2205156. Mori, N., Takahashi, T., GROUP, T. T.E.T.J.S., 2012. Nationwide Post Event Survey and
Freire, S., Santos, T., Navarro, A., Soares, F., Silva, J., Afonso, N., Fonseca, A., Tenedório, Analysis of the 2011 Tohoku Earthquake Tsunami. Coast. Eng. J. 54, 1250001-
J., 2014. Introducing mapping standards in the quality assessment of buildings 1–1250001-27. https://doi.org/10.1142/S0578563412500015.
extracted from very high resolution satellite imagery. ISPRS J. Photogram. Remote Mori, N., Takahashi, T., Yasuda, T., Yanagisawa, H., 2011. Survey of 2011 Tohoku
Sens., 90, 1–9. URL: doi: 10.1016/j.isprsjprs.2013.12.009 http://linkinghub. earthquake tsunami inundation and run-up. Geophys. Res. Lett. 38 https://doi.org/
elsevier.com/retrieve/pii/S0924271613003031. doi:10.1016/j. 10.1029/2011GL049210. URL: https://agupubs.onlinelibrary.wiley.com/doi/abs/
isprsjprs.2013.12.009. 10.1029/2011GL049210.
Ge, P., Gokon, H., Meguro, K., 2020. A review on synthetic aperture radar-based building Naito, S., Tomozawa, H., Mori, Y., Nagata, T., Monma, N., Nakamura, H., Fujiwara, H.,
damage assessment in disasters. Remote Sens. Environ. 240, 111693. https://doi. Shoji, G., 2020. Building-damage detection method based on machine learning
org/10.1016/j.rse.2020.111693. utilizing aerial photographs of the Kumamoto earthquake. Earthquake Spectra.
Geiß, C., Aravena Pelizari, P., Marconcini, M., Sengara, W., Edwards, M., Lakes, T., https://doi.org/10.1177/8755293019901309, 8755293019901309.
Taubenböck, H., 2015. Estimation of seismic building structural types using multi- Okada, S., Takai, N., 2000. CClassifications of structural types and damage patterns of
sensor remote sensing and machine learning techniques. ISPRS J. Photogram. buildings for earthquake field investigation. In: the 12th World Conference on
Remote Sens. 104, 175–188. https://doi.org/10.1016/j.isprsjprs.2014.07.016. Earthquake Engineering.
Ghosh, S., Huyck, C.K., Greene, M., Gill, S.P., Bevington, J., Svekla, W., DesRoches, R., Okamura, M., Bhandary, N.P., Mori, S., Marasini, N., Hazarika, H., 2015. Report on a
Eguchi, R.T., 2011. Crowdsourcing for Rapid Damage Assessment: The Global Earth reconnaissance survey of damage in Kathmandu caused by the 2015 Gorkha Nepal
Observation Catastrophe Assessment Network (GEO-CAN). Earthquake Spectra 27, earthquake. Soils Found., 55, 1015–1029. URL: http://www.sciencedirect.com/
179–198. https://doi.org/10.1193/1.3636416. science/article/pii/S0038080615001079. 2015, doi: 10.1016/j.sandf.2015.09.005.
Goda, K., Kiyota, T., Pokhrel, R.M., Chiaro, G., Katagiri, T., Sharma, K., Wilkinson, S., Special Issue on the Six International Symposium on Deformation Characteristics of
2015. The 2015 Gorkha Nepal Earthquake: Insights from Earthquake Damage Geomaterials IS-Buenos Aires.
Survey. Frontiers. Built Environ. 1, 8. https://doi.org/10.3389/fbuil.2015.00008. Oktay, O., Schlemper, J., Folgoc, L.L., Lee, M.C.H., Heinrich, M.P., Misawa, K., Mori, K.,
URL: https://www.frontiersin.org/article/10.3389/fbuil.2015.00008. McDonagh, S.G., Hammerla, N.Y., Kainz, B., Glocker, B., Rueckert, D., 2018.
Gokon, H., Koshimura, S., 2012. Mapping of Building Damage of the 2011 Tohoku Attention U-Net: Learning where to look for the pancreas. CoRR, abs/1804.03999.
Earthquake Tsunami in Miyagi Prefecture. Coast. Eng. J., 54, 1250006–1. URL: URL: http://arxiv.org/abs/1804.03999. arXiv:1804.03999.
http://www.worldscinet.com/cej/54/5401/S0578563412500064.html. doi: Park, J., Song, W.J., Pearlman, W.A., 1999. Speckle filtering of SAR images based on
10.1142/S0578563412500064. adaptive windowing. IEE Proc. - Vision, Image Signal Process. 146, 191–197.
Gokon, H., Post, J., Stein, E., Martinis, S., Twele, A., Mück, M., Geiß, C., Koshimura, S., Paulik, R., Gusman, A., Williams, J.H., Pratama, G.M., Lin, S.-L., Prawirabhakti, A.,
Matsuoka, M., 2015. A Method for Detecting Buildings Destroyed by the 2011 Sulendra, K., Zachari, M.Y., Fortuna, Z.E.D., Layuk, N.B.P., Suwarni, N.W.I., 2019.
Tohoku Earthquake and Tsunami Using Multitemporal TerraSAR-X Data. IEEE Tsunami Hazard and Built Environment Damage Observations from Palu City after
Geosci. Remote Sens. Lett. 12, 1277–1281. the September 28 2018 Sulawesi Earthquake and Tsunami. Pure Appl. Geophys.,
Grünthal, G., Musson, R., Schwarz, J., Stucchi, M., 1998. European Macroseismic Scale 176, 3305–3321. URL: doi: 10.1007/s00024-019-02254-9. doi:10.1007/s00024-
1998. Technical Report 101 Centre Européen de Géodynamique et de Séismologie. 019-02254-9.
Gupta, R., Hosfelt, R., Sajeev, S., Patel, N., Goodman, B., Doshi, J., Heim, E., Choset, H., Plank, S., 2014. Rapid Damage Assessment by Means of Multi-Temporal SAR – A
Gaston, M., 2019. Creating xBD: A Dataset for Assessing Building Damage from Comprehensive Review and Outlook to Sentinel-1. Remote Sens. 6, 4870–4906.
Satellite Imagery. In: The IEEE Conference on Computer Vision and Pattern https://doi.org/10.3390/rs6064870. URL: https://www.mdpi.com/2072-4292/6/
Recognition (CVPR) Workshops. 6/4870.
Karimzadeh, S., Matsuoka, M., Miyajima, M., Adriano, B., Fallahi, A., Karashi, J., 2018. Roeber, V., Bricker, J.D., 2015. Destructive tsunami-like wave generated by surf beat
Sequential SAR Coherence Method for the Monitoring of Buildings in Sarpole-Zahab. over a coral reef during Typhoon Haiyan. Nat. Commun. 6, 7854. https://doi.org/
Iran. Remote Sens. 10 https://doi.org/10.3390/rs10081255. URL: https://www. 10.1038/ncomms8854.
mdpi.com/2072-4292/10/8/1255. Roeslin, S., Ma, Q.T.M., García, H.J., 2018. Damage Assessment on Buildings Following
Kingma, D.P., Ba, J., 2014. Adam: A Method for Stochastic Optimization. arXiv: the 19th September 2017 Puebla, Mexico Earthquake. Frontiers. Built Environ. 4, 72.
1412.6980. https://doi.org/10.3389/fbuil.2018.00072. URL: https://www.frontiersin.org/
Koshimura, S., Moya, L., Mas, E., Bai, Y., 2020. Tsunami Damage Detection with Remote article/10.3389/fbuil.2018.00072.
Sensing: A Review. Geosciences 10. https://doi.org/10.3390/geosciences10050177. Rogan, J., Franklin, J., Stow, D., Miller, J., Woodcock, C., Roberts, D., 2008. Mapping
Koshimura, S., Oie, T., Yanagisawa, H., Imamura, F., 2009. Developing Fragility land-cover modifications over large areas: A comparison of machine learning
Functions for Tsunami Damage Estimation Using Numerical Model and Post- algorithms. Remote Sens. Environ. 112, 2272–2283. https://doi.org/10.1016/j.
Tsunami Data from Banda Aceh, Indonesia. Coast. Eng. J. 51, 243–273. https://doi. rse.2007.10.004.
org/10.1142/S0578563409002004. Ronneberger, O., Fischer, P., Brox, T., 2015. U-Net: Convolutional Networks for
Lecun, Y., Bottou, L., Bengio, Y., Haffner, P., 1998. Gradient-based learning applied to Biomedical Image Segmentation. CoRR, abs/1505.04597. URL: http://arxiv.org/
document recognition. Proc. IEEE 86, 2278–2324. abs/1505.04597. arXiv:1505.04597.
Li, Q., Shi, Y., Huang, X., Zhu, X.X., 2020. Building Footprint Generation by Integrating Rossi, C., Eineder, M., 2015. High-Resolution InSAR Building Layovers Detection and
Convolution Neural Network With Feature Pairwise Conditional Random Field Exploitation. IEEE Trans. Geosci. Remote Sens. 53, 6457–6468.
(FPCRF). In: IEEE Trans. Geosci. Remote Sens., pp. 1–18. Shahzad, M., Maurer, M., Fraundorfer, F., Wang, Y., Zhu, X.X., 2019. Buildings Detection
Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., in VHR SAR Images Using Fully Convolution Neural Networks. IEEE Trans. Geosci.
Zitnick, C.L., 2014. Microsoft coco: Common objects in context. In: Fleet, D., Remote Sens. 57, 1100–1116. https://doi.org/10.1109/TGRS.2018.2864716.
Pajdla, T., Schiele, B., Tuytelaars, T. (Eds.), Computer Vision – ECCV 2014. Springer Sharma, K., Deng, L., Noguez, C.C., 2016. Field investigation on the performance of
International Publishing, Cham, pp. 740–755. building structures during the April 25, 2015, Gorkha earthquake in Nepal. Eng.
Liu, X., Deng, Z., Yang, Y., 2019. Recent progress in semantic image segmentation. Artif. Struct., 121, 61–74. URL: http://www.sciencedirect.com/science/article/pii/
Intell. Rev. 52, 1089–1106. https://doi.org/10.1007/s10462-018-9641-3. S0141029616301626. doi: 10.1016/j.engstruct.2016.04.043.
Ma, H., Liu, Y., Ren, Y., Yu, J., 2019. Detection of Collapsed Buildings in Post-Earthquake Shermeyer, J., Hogan, D., Brown, J., Etten, A.V., Weir, N., Pacifici, F., Haensch, R.,
Remote Sensing Images Based on the Improved YOLOv3. Remote Sens. 12, 44. Bastidas, A., Soenen, S., Bacastow, T., Lewis, R., 2020. Spacenet 6: Multi-sensor all
https://doi.org/10.3390/rs12010044. weather mapping dataset. arXiv:2004.06500.
142
B. Adriano et al. ISPRS Journal of Photogrammetry and Remote Sensing 175 (2021) 132–143
Shi, Y., Li, Q., Zhu, X.X., 2020. Building segmentation through a gated graph Wei, Y., Zhang, K., Ji, S., 2020. Simultaneous Road Surface and Centerline Extraction
convolutional neural network with deep structured feature embedding. ISPRS J. From Large-Scale Remote Sensing Images Using CNN-Based Segmentation and
Photogram. Remote Sens. 159, 184–197. https://doi.org/10.1016/j. Tracing. In: IEEE Trans. Geosci. Remote Sens., pp. 1–13. https://doi.org/10.1109/
isprsjprs.2019.11.004. tgrs.2020.2991733.
Smith, L.N., 2017. Cyclical Learning Rates for Training Neural Networks. In: 2017 IEEE Widiyanto, W., Santoso, P.B., Hsiao, S.-C., Imananta, R.T., 2019. Post-event field survey
Winter Conference on Applications of Computer Vision (WACV), pp. 464–472. of 28 September 2018 Sulawesi earthquake and tsunami. Nat. Hazards Earth Syst.
Tajima, Y., Yasuda, T., Pacheco, B.M., Cruz, E.C., Kawasaki, K., Nobuoka, H., Sci. 19, 2781–2794. https://doi.org/10.5194/nhess-19-2781-2019. URL: https://
Miyamoto, M., Asano, Y., Arikawa, T., Ortigas, N.M., Aquino, R., Mata, W., www.nat-hazards-earth-syst-sci.net/19/2781/2019/.
Valdez, J., Briones, F., 2014. Initial Report of JSCE-PICE Joint Survey on the Storm Xie, S., Girshick, R., Dollár, P., Tu, Z., He, K., 2017. Aggregated Residual Transformations
Surge Disaster Caused by Typhoon Haiyan. Coast. Eng. J. 56, 1450006. https://doi. for Deep Neural Networks. In: 2017 IEEE Conference on Computer Vision and
org/10.1142/S0578563414500065. Pattern Recognition (CVPR), pp. 5987–5995.
Taucer, F., Alarcon, J.E., So, E., 2009. 2007 August 15 magnitude 7.9 earthquake near Xie, Y., Dai, W., Hu, Z., Liu, Y., Li, C., Pu, X., 2019. A Novel Convolutional Neural
the coast of Central Peru: analysis and field mission report. Bull. Earthq. Eng. 7, Network Architecture for SAR Target Recognition. J. Sensors, 2019, 1246548. URL:
1–70. https://doi.org/10.1007/s10518-008-9092-3. doi: 10.1155/2019/1246548. doi:10.1155/2019/1246548.
Tong, X., Hong, Z., Liu, S., Zhang, X., Xie, H., Li, Z., Yang, S., Wang, W., Bao, F., 2012. Yamada, M., Ohmura, J., Goto, H., 2017. Wooden Building Damage Analysis in Mashiki
Building-damage detection using pre- and post-seismic high-resolution satellite Town for the 2016 Kumamoto Earthquakes on April 14 and 16. Earthquake Spectra
stereo imagery: A case study of the May 2008 Wenchuan earthquake. ISPRS J. 33, 1555–1572. https://doi.org/10.1193/090816eqs144m.
Photogram. Remote Sens. 68, 13–27. https://doi.org/10.1016/j. Yamaguchi, Y., 2012. Disaster Monitoring by Fully Polarimetric SAR Data Acquired With
isprsjprs.2011.12.004. ALOS-PALSAR. Proceedings of the IEEE, 100, 2851–2860. URL: http://ieeexplore.
Twumasi, N.Y.D., Shao, Z., Orhan, A., 2019a. Remote Sensing and GIS Methods in Urban ieee.org/document/6205771/. doi:10.1109/JPROC.2012.2195469.
Disaster Monitoring and Management - An Overview. Int. J. Trend Scient. Res. Yamanaka, H., Chimoto, K., Miyake, H., Tsuno, S., Yamada, N., 2016. Observation of
Develop. (IJTSRD), 3, 918–926. URL: https://doi.org/10.31142/ijtsrd23976. doi: earthquake ground motion due to aftershocks of the 2016 Kumamoto earthquake in
ijtsrd23976. arXiv:https://doi.org/10.31142/ijtsrd23976. damaged areas. Earth, Planets and Space 68, 197. https://doi.org/10.1186/s40623-
Twumasi, N.Y.D., Shao, Z., Orhan, A., 2019. Remote sensing and gis methods in urban 016-0574-2.
disaster monitoring and management – an overview. Int. J. Trend Scient. Res. Yamazaki, F., Matsuoka, M., 2007. Remote Sensing Technologies in Post-disaster
Develop. (IJTSRD) 3, 918–926. https://doi.org/10.31142/ijtsrd23976. Damage Assessment. J. Earthquake Tsunami 01, 193–210. https://doi.org/10.1142/
Vetr, M.G., Saeidian, M., Naserpour, A., 2018. Sequential SAR Coherence Method for the S1793431107000122.
Monitoring of Buildings in Sarpole-Zahab, Iran The Main Reasons for Great Damages Yamazaki, F., Yano, Y., Matsuoka, M., 2005. Visual Damage Interpretation of Buildings
of Reinforced Concrete Buildings on 12th November 2017, Sarpol-e Zahab in Bam City using QuickBird Images following the 2003 Bam, Iran, Earthquake.
Earthquake. Journal of Seismology and Earthquake Engineering, 10. URL: http:// Earthquake Spectra 21, 329–336. https://doi.org/10.1193/1.2101807.
www.jsee.ir/index.php/jsee/article/view/499. Yusuf, Y., Matsuoka, M., Yamazaki, F., 2001. Damage assessment after 2001 Gujarat
Wallemacq, P., Below, R., 2018. Economic losses, poverty and disasters: 1998–2017. earthquake using Landsat-7 satellite images. J. Indian Soc. Remote Sens. 29, 17–22.
Technical Report United Nations Office for Disaster Risk Reduction 9–11 Rue de https://doi.org/10.1007/BF02989909.
Varembee CH 1202, Geneva Switzerland.
143