Professional Documents
Culture Documents
Abstract: Risk assessment for large engineered systems involving geotechnical uncertainties—such as dams, coastal works, industrial
Downloaded from ascelibrary.org by University of Central Florida on 11/03/23. Copyright ASCE. For personal use only; all rights reserved.
facilities, and offshore structures—has matured in recent decades. This capability is becoming an ever more pressing concern considering
increasing natural hazards and the effects of climate change. This progress and the current state of practice are reviewed by considering four
projects. An underlying theme is what we have learned about how geoprofessionals should think about uncertainty, judgment, and risk.
The organizing key is Bayesian thinking, which is neither new nor unintuitive to most geoprofessionals. However, this way of thinking
and analyzing data is powerful, even when dealing with the sparse data of geotechnical projects. The presentation builds on the past Terzaghi
Lectures of Casagrande, Whitman, and Christian. It uses examples of infrastructure systems for which geotechnical considerations were im-
portant. Reliability concepts developed in this earlier practice provide a window into the processes of integrating site characterization data,
judgment, engineering models, and predictions to achieve tolerable levels of risk. Along the way, arguments are made on the nature of uncertainty
and on the benefits of processing data, judgment, and inferences through a consistent lens. DOI: 10.1061/JGGEFK.GTENG-10201. This work
is made available under the terms of the Creative Commons Attribution 4.0 International license, https://creativecommons.org/licenses/by/4.0/.
Author keywords: Risk; Reliability; Uncertainty; Systems engineering; Bayesian.
Introduction risk analyses of large geotechnical systems and at the same time
to examine the philosophy behind that approach.
The topic is risk analysis of large geotechnical systems. Such
systems include flood and coastal protection works, hydropower
dams, large industrial facilities, offshore installations, and similar Tonen Reprised
infrastructures. All of these involve complicated geotechnical
uncertainties. The approach to the risk analysis of such complex We start with a retrospective look at the Tonen Refinery project
systems involving geotechnical uncertainties has matured in recent (1978–1982) discussed by Whitman. This was among the first risk
decades but the need for future development remains given the analyses of a large geoengineered project. Whitman asked two ques-
intensifying of natural hazards and a changing climate (Allan tions of that example: (1) is it possible to quantify risk, and
et al. 2021). (2) can geotechnical reliability be applied in practice? He adapted
There have been at least three Terzaghi Lectures related to Casagrande’s title of 20 years earlier to ask, “can we evaluate a cal-
uncertainty and risk in geotechnical practice. Casagrande (1965) culated risk?” It is important to consider Whitman’s discussion
introduced a qualitative way of thinking about “the calculated risk in the context of its time. This was an era when risks associated
in earthworks.” Whitman (1984) proposed quantifying that ap- with nuclear power were being debated, WASH-1400, US Nuclear
proach and suggested that quantification was indeed possible. Regulatory Commission (USNRC)’s Reactor Safety Study: The
Christian (2004) expanded and critiqued the quantitative approach Introduction of Risk Assessment to the Regulation Of Nuclear Re-
with later developments in reliability theory. actors had just appeared, and the National Aeronautics and Space
The present discussion focuses on the application of geotechni- Administration was trying to quantify the risks of manned space
cal risk analysis to increasingly complicated infrastructure systems. flight (NRC 1988; USNRC 2016). The years between Casagrande’s
It uses four projects to illustrate lessons learned in that evolution: lecture and Whitman’s had seen the incidents at Teton, Lower San
(1) a reprisal of the Tonen Refinery project discussed by Whitman Fernando, Buffalo Creek, Canyon Lake, Laurel Run, Toccoa Falls,
(1984), (2) the Interagency Performance Evaluation Taskforce and others (Gee 2017). He answered each question in the affirmative.
(IPET) following Hurricane Katrina, (3) the Panama Canal Natural Several lessons were learned from this early project that have be-
and Chronic Risks Study, and (4) the joint-industry study on spillway come routine in current risk analysis methods. Among these were:
systems reliability. The engineering methodologies of these exam- 1. The use of event trees for organizing uncertainties,
ples are important, but equally are the concepts underlying them. 2. A taxonomy of uncertainty for geotechnical data,
The purpose is to layout the rudiments of the current approach to 3. The importance of spatial variation of geotechnical properties,
and
1
Glenn L. Martin Institute Professor of Engineering, Dept. of Civil 4. Methods for communicating risk to stakeholders.
and Environmental Engineering, Univ. of Maryland, College Park, MD The project involved more than 100 storage tanks founded on
20742. ORCID: https://orcid.org/0000-0002-9571-9282. Email: gbaecher@ loose sandy bay fill in a seismically active harbor (Fig. 1). To protect
umd.edu
against offsite loss of product, the facilities were subdivided by fire-
Note. This manuscript was submitted on June 22, 2021; approved on
September 28, 2022; published online on January 10, 2023. Discussion walls into patios and those patios were contained within larger re-
period open until June 10, 2023; separate discussions must be submitted vetments. Many of the patios were interconnected by underground
for individual papers. This paper is part of the Journal of Geotechnical drainage systems. While the direct cost of foundation failures
and Geoenvironmental Engineering, © ASCE, ISSN 1090-0241. could be large, they would be exceeded by environmental costs.
Fig. 1. (Color) Tonen refinery and tank farm, Kawasaki. (Image © 2021 Google, Landsat/Copernicus; Data SIO, NOAA, U. S. Navy, NGA, GEBCO;
Data Japan Hydrographic Association; Data LDEO-Columbia, NSF, NOAA.)
Usefulness of Event Trees in Organizing Uncertainties calculated by multiplying the successive conditional probabilities
at each node
Many of the lessons learned on the Tonen project seem obvious
today, but they were less so then. The first was how to build a PðpathÞ ¼ Pðx1 ÞPðx2 jx1 Þ · · · Pðxn jxn−1 ; : : : ; x1 Þ ð1Þ
systems model of the facility. The then-recent WASH-1400 study
used fault trees to combine the performances of the many compo- In which PðpathÞ = joint probability of the events along an
nents in a nuclear plant to calculate the probability of accident individual path through the event tree from its beginning to an out-
sequences. A fault tree is a logic diagram that reasons backward come; and fx1 ; : : : ; xn g = individual uncertainties in the tree. The
from an adverse performance to the faults necessary to cause it vertical line indicates a conditional probability of the uncertainty on
(Modarres 1993). That worked well with mechanical-electrical sys- the left given the occurrence of the uncertainties on the right. The
tems that could be fully enumerated. Yet, that was not the case for sum or integral of the uncertainties at each node must equal 1.0
the tank farm which was a more open system. The tank farm had because each node is a probability distribution of an uncertainty.
some components that could be discretely enumerated, like the Commercial software applications are widely available to do these
tanks themselves and drainage pipes and firewalls, but the perfor- calculations.
mance also depended on soil deposits under the site, some natural
and some manmade. It also depended on external hazards, specifi-
A Taxonomy of Uncertainty for Geotechnical Data
cally earthquakes, that at the time had not been part of the WASH-
1400 analysis. Event tree models were used because they were Most of the subsurface data for the Tonen project came from
flexible and could be used to combine uncertainties of many types, standard penetration tests, which were highly variable (Fig. 3).
including those based on judgment (Fig. 2). An event tree is a logic The geometry of the soil strata differed systematically across
diagram complementary to a fault tree that reasons forward from an the site, so a set of soil models was created. Yegian and Whitman
initiating event to subsequent adverse performance. (1978) recently developed a reliability model for relating blow
Event trees have since become the standard representation for count data to the probability of liquefaction triggering, and that
uncertainties in almost all dam and levee safety studies (Hartford was the approach used. Since the late 1970s, there has been a great
and Baecher 2004; FEMA 2015). They are widely used in offshore deal of work on liquefaction triggering and residual strength, but
applications (Rausand and Haugen 2020) and in mining (Brown that came later and is tangential to the present discussion (NRC
2012). In an event tree, the many uncertainties entering an analysis 2016).
are laid out graphically, with each node in the tree representing the The question arose as to what the large scatter in the in situ
probability distribution of an uncertainty. The probability of pro- data meant for engineering predictions. Some of that scatter was
ceeding down a chain of uncertain outcomes through the tree is presumably due to real variability and some due to random testing
Fig. 2. Map and event tree for a Patio 4. Each of the nodes at the fourth level had a sub-tree as shown for the single case of two tank failures.
(Reprinted from Whitman 1984, © ASCE.)
errors. However, in interpreting the data one also had to consider to some extent average out, but statistical errors and measurement
systematic biases in the measurements and biases due to statistical bias create systematic errors that do not. The latter are more concern-
and testing errors. Not each of these errors affects prediction un- ing because they create correlations of error among the predictions of
certainty in the same way. Spatial variation and measurement noise performance.
Fig. 5. Spatial averaging and correlation of soil properties beneath tanks of various sizes and separations. The left-hand side (a) shows variance
reduction as a function of tank radius. The right-hand side (b) shows the covariance between the averages of soil properties under adjacent tanks as a
function of their radius and separation. (Data from Lambe and Associates 1982.)
defines risk, and their focus is largely on probabilities. The obvious frequency. A tradeoff between how much of the spatial variation
way of communicating risk is by multiplying probability and con- is modeled as aleatory and how much as epistemic can be made
sequence to get an expected value. One sometimes sees this as a by changing the granularity of the trend model. This is conceptually
formal definition of risk, but in the field of risk science, the concept similar to changing the flexibility of a regression model.
is more complex (Kaplan and Garrick 1981). In principle, one The separation of uncertainty between aleatory and epistemic
ought to be willing to pay up to this expectation to eliminate the is mostly a convenience of modeling (Faber 2005; Kiureghian and
risk, but this can be a difficult proposition for the owner. Ditlevsen 2009). Thus, while it is sometimes said that epistemic
Instead, data were collected on comparative risks for similar fa- uncertainty is reducible (with more knowledge) but aleatory un-
cilities and used to compare site-specific risks (Fig. 6). The concept certainty is not, that is at most true within the context of a set of
was that this comparative information could be used with regula- modeling assumptions (Vrouwenvelder 2003). For example, it is
tory authorities to demonstrate good faith in taking actions to re- common in probabilistic seismic hazard analysis to presume that
duce risk. The approach worked well in communicating with the earthquakes occur randomly and independently within the con-
client and in turn with regulators. Whitman published this figure in straints of a source model. In fact, the geomechanics controlling
his Terzaghi Lecture, and it has become well-known. It is fortunate the occurrence and size of the earthquakes are likely determinis-
that he did so, because otherwise, it would have quietly disap- tic, but we do not know enough to characterize the process as
peared: at the time, no one on the team thought it all that interesting. such, so we model the uncertainty as aleatory. That is a modeling
choice.
Geotechnical Uncertainty I: Probability and
Uncertainty Geotechnical Uncertainties Are (Mostly) Epistemic
The discussion to this point has dealt with modeling geotechnical Geotechnical uncertainties are mostly epistemic. They have to do
uncertainty; however, the Tonen project also led to conceptual in- with what we know, not about randomness in nature. As a result,
sights which took longer to be recognized. These are likely of more they are almost always subjective. The probability theory appropri-
consequence than the modeling tools themselves. ate to these sorts of uncertainty has to do with degrees-of-belief,
not with frequencies (Vick 2002). The lectures of Casagrande,
Whitman, and Christian call out the importance of judgment in
Probability Is in the Model, Not in the Ground geotechnical practice, which is simply another name for the sub-
Georges Matheron, the founder of geostatistics said, “probability is jectivity of most geotechnical uncertainties.
in the model, not in the ground” (Matheron 1989). There is nothing
random about in situ soil properties: they are what they are, but we
just do not know what they are. In modern risk and reliability mod- Interagency Performance Evaluation Taskforce
eling, uncertainties are commonly divided between aleatory and
epistemic. Aleatory uncertainties are those modeled as variability The second project was the IPET (2005–2009). The IPET was
(or randomness as in the toss of a coin) in nature and describable by organized under the Corps of Engineers to understand the perfor-
frequencies. Thus, one may model the spatial variations of soil mance of the hurricane protection system in New Orleans during
properties across a site as aleatory, but not the properties at a par- Hurricane Katrina (Link and Harris 2007). This, too, involved risk
ticular point. Epistemic uncertainties are those due to a lack of analysis, but now of an even larger geoengineered system. This sys-
knowledge and are described by degrees-of-belief. Soil properties tem in addition, however, protected human life (Figs. 7 and 8). The
at a particular point are unique although possibly uncertain. They system was subsequently renamed, the Hurricane and Storm
may be unknown but are not variable and not describable by Damage Risk Reduction System (HSDRRS).
Fig. 6. Generic probability versus consequence chart prepared for the Tonen project. (Reprinted from Whitman 1984, © ASCE.)
Fig. 7. (Color) Flooding of New Orleans after Hurricane Katrina in 2005. (Image courtesy of NOAA Corps/Lt. Cmdr. Mark Moran.)
Fig. 8. (Color) Subbasin definitions for the calculation of stage-storage relations. (Reprinted from USACE 2008, courtesy of US Army Corps of
Engineers.)
TVC Approach
Fragility ≡ PðperformancejloadÞ ð5Þ
The TVC approach is widely used for natural hazards. Threats
to the system are characterized by probabilities of load levels. The probability of adverse performance as a function of load
These might be surge or flood elevations, peak ground accelera- is commonly called a fragility curve. The probability of the load is
tions, wind speeds, malicious intent, or the like. The performance combined with the fragility to generate probabilities of adverse
of the system under these loads is characterized by fragility, which performance. Consequences are then associated with adverse
is the probability of adverse performance conditioned on load performance
fined as a structural transition, gate, pump station, drainage con- the reliability analysis: (1) geological and geotechnical uncertain-
trol structure, or other local discontinuity. A portion of the system ties involving the spatial distribution of soils and soil properties
definition near Lakefront Airport in New Orleans East is shown in within and beneath the reaches; (2) engineering mechanics uncer-
Fig. 10 (USACE 2008). This offered the opportunity to examine tainties, involving the performance of manmade systems such as
risk at the census block level or aggregated to subbasins, basins levees, floodwalls, and point features such as drainage pipes; and
(parishes), or system-wide levels. the engineering modeling of that performance, including geotech-
The frequencies of occurrence of surge and waves differ around nical performance modeling; (3) erosion uncertainties, involving
the HSDRRS. The IPET Hurricane Hazard Team developed new the performance of levees and fills around floodwalls during
statistical techniques based on numerical surge and wave models overtopping, and at points of transition between levees and flood-
to quantify these hazards. The model of hurricane occurrence used wall, in some cases leading to loss of grade or loss of support, and
a joint probability distribution of principal hurricane parameters to consequently to breaching; and (4) mechanical equipment uncer-
characterize the frequencies of storm events based on historical tainties, including gates, pumps, and other operating systems, and
storms in the Gulf of Mexico. An ensemble of about 2000 possible human operator factors affecting the performance of mechanical
hurricane tracks and characteristics (Fig. 11) was assembled and equipment.
combined with respect to individual storm probabilities to generate
hydrographs at individual reaches or point features around the pro-
Logic Trees
tection system (Fig. 12) (USACE 2009). This was judged to re-
present the range of storms that might strike New Orleans and An event tree approach was used on IPET (Fig. 14), as at Tonen, but
was used to generate the local threat at each reach expressed as now incorporating logic trees to simplify the separation of episte-
the annual probability of surge height. mic from aleatory uncertainties (Bommer and Scherbaum 2008;
Bommer 2012). The idea of logic trees is to separate out the epi-
stemic uncertainties due to lack of knowledge into their own tree
Fragility Curves and then to condition an event tree containing only aleatory uncer-
Geotechnical reliability was described by fragility curves that char- tainties on the respective branches of that epistemic logic tree
acterized the predicted performance of individual reaches or struc- (Fig. 15). This greatly simplifies the event tree calculations by
tures as a function of surge and wave levels. The fragility curves implicitly capturing the correlation structure among the aleatory
were generated from an evaluation of subsurface conditions, the events, which in the Tonen project had to be manually calculated
character of the structures, and relevant failure mechanisms defined and entered into the tree, an approach that may also overlook
previously by the IPET Performance Analysis Team. subtle correlations. The logic tree approach had the benefit of sepa-
There was neither time nor resources to reanalyze the many rating those uncertainties that can be reduced by collecting more
reaches and features. On the other hand, the New Orleans District data or improving models from those that cannot (Kiureghian and
US Army Corps of Engineers (USACE) maintained comprehensive Ditlevsen 2009).
historical design records and drawings, with some extending to the In hindsight, using event trees was not the best approach. The
creation of the District in 1803 (Fig. 13). These were used to number of end-nodes in the event trees was in the millions for each
approximate fragility curves for individual reaches based on herit- of the 76 storm scenarios used in the analysis. This caused com-
age data and engineering designs, often based on practical but ro- putation problems, reduced the number of scenarios that could be
bust methods such as first-order second-moment reliability (Duncan computed, and precluded interim results from being easily ex-
2000) or expert opinion. This provided what the team judged to tracted. Direct simulation using Monte Carlo might have been more
be reasonable fragility curves in a timely way, but obviously with powerful and cost-effective. In this case, a nested simulation ap-
less fidelity than a comprehensive reanalysis might have provided proach could have been used in which a simulation containing only
using a more modern analytical tool, which in any case was not the aleatory uncertainties is nested within an outer simulation con-
practicable. taining the epistemic uncertainties, as discussed below.
Failure was defined as breaching, which allowed water to enter a
protected polder. This failure occurred in two ways: (1) loss of
Putting Numbers on Engineering Judgment
levee or wall stability, or (2) overtopping causing the protected side
of the levee or wall to erode and breach (USACE 2007a, b; see The subjective probabilities of experts were critical to the analysis.
Table 1). The loading at which failure occurs varies along the reach Geotechnical engineers are true believers in the value of engineer-
and is not precisely known. Part of the uncertainty is due to episte- ing judgment to fill in for gaps in information or inadequacies
mic uncertainty in the average soil strength or average permeability of modeling and analysis. Peck (1980) famously championed this
along the reach, but another part is due to spatial variability within view. Yet human judgment is flawed, and a host of cognitive biases
the reach that was modeled as an aleatory uncertainty. The spatial are recognized as affecting how and how well people put numbers
Fig. 10. (Color) (a) NOLA East basin; and (b) corresponding reaches and point features as defined in the IPET8 study. (Reprinted from USACE 2008,
courtesy of US Army Corps of Engineers.)
Fig. 11. (Color) Ensemble of 76 hurricane scenarios. (Data from USACE 2008.)
Fig. 12. (Color) Modeled hydrographs for a particular hurricane at the 135 reaches around the hurricane protection system (USACE 2008). Surge
elevation is shown on the y-axis against time in days.
on their subjective uncertainty (Baecher 2019; Marr 2019, 2020; 2007 reconstruction conditions. Again, these were portrayed
Vick 2002). on probability-consequence charts, sometimes referred to today
as frequency-number (FN) charts (Fig. 16). The FN chart is a
Communicating Risk to the Public modern formalization of the concept behind Fig. 6. Here, it displays
Consequence data (property and life loss) were generated for each annual exceedance probabilities of risk to life for the preHurricane
storm scenario. The risk model generated probabilistic estimates Katrina and postreconstruction conditions. Similar results from the
of the extent and depth of inundation in each of the 27 polders. California Delta Risk Management Strategy (URS 2009) are shown
Inundation depths were used to interpolate property and life for comparison. The latter results tail off more quickly because the
loss estimates for prehurricane Katrina conditions and post population at risk is smaller.
Fig. 13. (Color) Sheet pile failure analysis by deep wedge instability from historical design files of the New Orleans District, USACE, for a specific
levee reach. (Reprinted from USACE 1972, courtesy of US Army Corps of Engineers.)
Fig. 14. (Color) Risk analysis logic diagram. (Adapted from USACE 2008.)
uncertainties.
Uncertainty about States of Nature
Two characteristics of Bayesian statistics make it fit for geotech-
nical use: it allows probabilities to be defined on states of nature
In recent years, the use of FN charts has become common
such as site conditions or engineering parameters, and it admits
in dam safety (ANCOLD 2003; FEMA 2015), industrial risk
subjective probabilities. Frequentist statistics allows neither.
(Jongejan 2008), and offshore engineering (Lacasse et al. 2019),
Frequentist and degree-of-belief statistics are not in conflict, they
among other fields. Criteria for tolerable societal risk have been
are just different. They use the same mathematical theory of
set by various organizations. Following recommendations by the
probability (Kolmogorov 1950) but differ in ways of defining un-
Australian-New Zealand Committee on Large Dams (ANCOLD
certainty (Baecher 2017). The distinction is philosophical but im-
2003) and others, these FN charts have sometimes been adopted
portant because it dictates how one draws inferences and makes
as guidelines for tolerable risk, although some organizations such
decisions.
as the Bureau of Reclamation explicitly do not use them for that
The distinction is illustrated in Fig. 17. A program of site char-
purpose (Reclamation 2011). These guidelines are typically orders
acterization results in observed data on soil engineering properties,
of magnitude lower than the risks existing along the Gulf coast or in
riverine flood hazard areas elsewhere in the United States (Fig. 16). the geometry of the subsurface, or other states of nature to be used
Riverine and coastal flood criteria in the United States have histor- in analysis and design. Frequentist statistics treats these states of
ically used a 1% annual exceedance probability as a criterion for nature as deterministic but unknown: in that theory, it is meaning-
insurance and risk mitigation. less to define probabilities on states of nature because those states
do not have frequencies. In frequentist statistics, it is the data that
are assigned probabilities, not the states of nature. Inferences are
Geotechnical Uncertainty II: Bayesian Thinking described as confidence bounds, p-values, and the like. These char-
acterize probabilities on what data might have been observed, not
The statistics course many of us took in college is mostly useless on the states of nature (Diaconis and Skyrms 2018; O’Hagan 2004).
for geotechnical practice—in the author’s opinion. That course These are direct probabilities.
Fig. 16. (Color) Calculated FN curves for IPET and in the California Delta Risk Management Project compared with ANCOLD (2003) criteria for
dam safety (USACE 2008).
In contrast, Bayesian statistics takes the data as known (i.e., they PðdatajHÞ = conditional probability of not finding the geological
have been observed) and calculates probabilities on the states of detail if it actually exists. There are many things to be learned from
nature that might have generated those data. Because probability this seemingly simple equation:
is interpreted as a degree-of-belief and not as a frequency, this is 1. First, data never speak for themselves. They only tell us how to
acceptable. These are inverse probabilities (Dale 1999). This is update from a prior probability (what we thought before seeing
neither new nor unintuitive to geoprofessionals, and most of us the data) to a posterior probability (what we should logically
are unwittingly Bayesians. de Mello (1975, 1977) was among think afterward). Lindley (1985) said, “today’s posterior is
the first to champion Bayesian ideas, although many others fol- tomorrow’s prior.”
lowed (Contreras 2020; Feng 2015; Gambino and Gilbert 1999; 2. Second, updating probabilities from priors to posterior is a
Garbulewski et al. 2009; Jiang et al. 2018, 2020; Juang and Zhang straightforward matter of multiplying the prior by the weight
2017; Kelly and Huang 2015; Papaioannou and Straub 2017; of evidence in the data. The weight of evidence is entirely con-
Phoon 2020; Schweckendiek 2014; Zhou et al. 2021). tained in this conditional probability of the data given the hy-
pothesis. The term of art for this conditional probability is the
likelihood.
Updating Probabilities 3. Third, at some point in history there are no data upon which to
Little is known about the eponymous Rev. Thomas Bayes (1701?– base a prior; therefore, the priors are ultimately a matter of
61) (Dale 2003). Even his likeness, which is pervasive in print and judgment and necessarily subjective.
on the web, is thought not to be him (IMS 1988). The person to 4. Fourth, this is the mathematics of learning by experience. It is
whom we owe the modern and more complete derivation of the the logic of Terzaghi’s (1961) observational approach.
theorem is Laplace (1749–1827). In the 20th century, the geophysi- Typically, the prior probability distributions in the Bayesian
cist Sir Harold Jeffreys (1891–1989) produced a long treatise on the equation are updated by site characterization, construction, and per-
inverse probability that remains in print today. This way of thinking formance data. The messiness of the mathematics of doing this up-
has a firm geoscience pedigree. dating through complicated engineering models for a long time
A simple modern representation of Bayes’ Theorem is impeded practical applications; however, the advent of numerical
methods like Markov chain Monte Carlo (MCMC) has facilitated
PðHÞPðdatajHÞ practical use (Robert and Casella 2010).
PðHjdataÞ ¼ ð7Þ
PðdataÞ Terzaghi proposed the “learn-as-you-go” approach. Peck (1969)
gave it the name, observational method. It is now an essential fea-
in which, H = some hypothesis (which could have to do with an ture of geotechnical practice. The environmental community calls
event, a probability distribution, or anything else about which one it, adaptive management. Einstein (1991) and Wu (2011) devel-
is trying to learn; PðHÞ = prior probability of the hypothesis; oped a quantitative Bayesian approach to the observational method,
PðHjdataÞ = posterior probability of the hypothesis, given the ob- as de Mello (1975) had also suggested.
served data; PðdatajHÞ = Likelihood of hypothesis based on the
data (i.e., the conditional probability of the data were the hypothesis
true); and PðdataÞ = marginal probability of the data irrespective of Site Characterization and Sampling
H (a normalizing constant).
The uncertainties of site characterization and their probabilities
The marginal probability of the data is readily found from the
are inverse probabilities. They are probabilities defined on states
total probability theorem as the weighted sum of the probability of
of nature, not probabilities on the data which are observed from
the data under the two respective hypotheses
those states. Hoek (1999) in the 2nd Glossop Lecture commented
PðdataÞ ¼ PðHÞPðdatajHÞ þ PðH̄ÞPðdatajH̄Þ ð8Þ that a geological model—conceptual, hand-drawn, or computer
rendered—is the building block upon which the design of any ma-
in which H̄ = complement, i.e., “not-H.” In more complicated jor construction project is based. It is principally the geological
applications, the normalizing constant is often determined model that generates the prior probabilities in Eq. (7). A recent ex-
numerically. ample of the importance of the geological model in interpreting the
Perhaps the simplest example is the well-studied question of statistics of deep-driven pile performance is provided by McGillivray
whether some geological detail exists at a site although undetected and Baecher (2021).
in exploration (Christian 2004). Presume that one believes that the One reads that site characterization is so uncertain because we
probability of such a detail exiting at the site before exploration is sample but a small fraction of the subsurface. This ignores statis-
performed was p ¼ 0.5. This might be a subjective judgment or tical science. Political polls in US elections sample but a tiny frac-
might be based on geological information. Perhaps a geophysical tion of the electorate, maybe 1,000 out of the 100 million who vote
survey has been undertaken with a p ¼ 0.3 chance of finding an nationally, or about 0.001% (FiveThirtyEight 2022). The inferential
are what statisticians call purposive. That means the locations program for the Gaillard Cut (Fig. 19). This was but a few years after
are chosen on purpose for one of many reasons: spatial coverage, Whitman’s Terzaghi Lecture. The chart was used to communicate
least favorable places, or places that have the most influence on what the impact of a landslide control program could be on risks
design. In any event, a purposive sample is not a random sample. faced by the Authority and thus to justify a significant investment.
Why does random sampling matter? It matters because the pre- The natural and chronic risk assessment began with a qualitative
sumption of all statistical methods is that the data are random. risk register. The risk register categorized risks by type of asset at
If they are not random, then the statistical theory does not apply. the Canal (dams, locks, cuts, gates, power stations, water plants, and
The infamous Literary Digest poll in the 1932 election between others) and among various hazards (seismic, hydrologic, meteoro-
Landon and Roosevelt is a painful example (Lusinchi 2012). It logical, operational). The annual risks across the various asset types
predicted a Landon victory, while in the event, Landon won only and failure modes were compiled. Given the magnitude of these
Vermont and Maine. The nonrandom sample was taken from the risks, management considered two questions: (1) is a particular risk
magazine’s subscribers who were mostly wealthy. acceptable, and if not, and (2) how much must it be reduced or how
can it be managed? FN curves were used as a qualitative means of
Judgment versus Statistical Methods making these judgments.
Scientific and operations data were compiled to characterize
Casagrande avoided quantitative probabilities as did most of his risk, while reliability models were developed to translate those
contemporaries, choosing verbal descriptions such as “grave risk” data into actionable assessments. Risks were categorized as cata-
and “great uncertainties” instead. We know today that verbal strophic, significant, or moderate. The first set was modeled in
descriptions can be deceiving (Baecher 2019), and psychological detail; the others were managed operationally. The resulting prob-
research has demonstrated that intuition is a poor substitute for abilities and consequences were tracked to understand where risk
the laws of probability when evaluating uncertainty. Experts are remediation was called for. This comprehensive risk assessment
systematically overconfident in reporting uncertainty, ignore base allowed ACP to reduce risk while controlling costs.
rates (i.e., prior probabilities), disregard the variability of small
samples, believe in the Prosecutor’s Fallacy of reversed condition-
als (Aitken and Taroni 2004), and exhibit many other systematic Qualitative Risk
errors in reasoning about probabilities. These can all be dealt with
Beginning in 2011, a qualitative risk analysis was undertaken. The
but need to be understood.
project was divided into phases. The initial phase was the develop-
In a long set of experiments in the social sciences and medicine,
ment of a risk register. The risk register is a list of threats, facilities,
the American psychiatrist Paul Meehl (1954) and later the psy-
and facility components that might be affected by those threats
chologist Robyn Dawes (Dawes et al. 1989) demonstrated that even
and possible consequences if those threats occurred (Table 2).
simple statistical models like regression analysis significantly out-
The risk register has become a common tool in geotechnical engi-
perform clinical judgments. This was a precursor to the later work
neering, especially in underground construction. The risk registers
of Tversky and Kahneman. Recently, Morgenstern (2018) has
at ACP ran to 600 items across all asset classes, but these were
made a related suggestion concerning dam safety and risk, specifi-
cally in the relationship between empirically-driven and purely sub- reduced to a few dozen highly important risks that were then mod-
jective appraisals: “Whether precautionary or performance-based, eled and analyzed in detail.
or even utilizing subjective judgments based on experience, it is es- The purpose of the risk register was to identify as many signifi-
sential that the risk assessment process be constrained by evidence cant risks to the Canal infrastructure as possible and to rank those
and its evaluation to a higher degree than is currently the case.” risks for further analysis. The rank-ordering categorized risks into
three sets: (1) those that were thought to require further analysis and
possibly modeling to obtain quantitative assessments, (2) those
Panama Canal Natural and Chronic Risk that were significant and needed to be monitored but were thought
not deserving of detailed analysis, and (3) those that could be man-
The third project was the natural and chronic risk study—an enter- aged as part of normal operations. A qualitative risk rank was used
prise risk analysis—for the Panama Canal Authority (ACP) (2011– to compare the structures and components within the portfolio. The
2016), an attempt to quantify enterprise risk. The Panama Canal, development of a qualitative risk assessment required many work-
commissioned in 1914, is among the world’s iconic engineering ing sessions to develop categories of hazards and an inventory of
achievements (Fig. 18). The Canal transits 14,000 vessels a year critical infrastructure. ACP used its own subject matter experts in
carrying more than three percent of world maritime trade. In the multidisciplinary teams to assess the likelihoods and consequence
early 1900s, the Panama site was thought free of natural hazards and the partitioning and ordering.
and therefore favored over other crossings. History has changed In the qualitative phase, simple rankings were assigned to prob-
that appraisal, and it is now understood that seismic, hydrologic, abilities and consequences. Both the probability of the hazard and
Fig. 18. (Color) Critical infrastructure of the Panama Canal. (Image courtesy of Wikimedia Commons/Thoroe.)
Table 2. Part of the ACP risk registers for engineering and chronic risks
Type ID Event Description Component Failure mode Mode Failure cause P C
DAM Gatun Floods PMF Crest Erosion Erosion Overtopping 2 1
Downloaded from ascelibrary.org by University of Central Florida on 11/03/23. Copyright ASCE. For personal use only; all rights reserved.
3 Impede operations for a short period (<1 year) or Moderate loss of toll revenue. Direct repair costs greater $500m-1b
moderate direct or indirect economic cost than $500m
4 Damages that affect canal capacity and revenues More than $100 m loss of toll revenue. Direct repair costs $100m–500m
greater than $100 m
5 Economic damages but the canal continues to Less than $100 m loss of toll revenue. Direct repair costs <100 m
operate less than $100 m. No impact on ACP reputation
Source: Data from Alfaro et al. (2016).
Scenario Risk
System simulation rather than detailed event trees was used to an earthquake would have cascading effects downstream at the
analyze risks. As had been learned in the IPET project, full- Pedro Miguel and Miraflores Locks, and the capacities of which
enumeration techniques like event trees become unwieldy for large might also have been compromised by the same earthquake. There
systems with many components and many uncertainties. Another may be nonlinear and even feedback effects within the larger sys-
benefit was also recognized, which was the utility of simulation for tem of assets. These interactions may be obvious in hindsight, but
modeling and understanding scenarios of risk simultaneously the system simulation may allow their identification ex ante by data
affecting large parts of the system. A large distributed system like mining of simulation results. This has been demonstrated in recent
the Canal comprises hundreds of assets including water retaining investigations including large-scale dam safety studies on the Göte
dams and dikes, navigation infrastructure, pumping facilities, elec- River system in Sweden (Ascila et al. 2015), the Cheakamus
trical generating stations, water treatment plants, and many others. System in British Columbia (King 2020), and the Lower Mattagami
These are mostly interconnected by the hydraulic system of water- River Project in Ontario (Zielinski 2021).
ways and structures. How does this large number of interacting
assets react to a unique natural hazard event—say, a large
earthquake—which simultaneously affects them all? An engineer- Geotechnical Uncertainty III: Safety
ing analysis of individual structures or components is clearly of
interest, but the adverse performance or failure of one may have How safe is safe enough? In Panama, the consequences of adverse
cascading effects on others (F. Nadim, personal communication, performance were mostly financial. This is not always the case with
2014). For example, the loss of a pool at Madden Dam during geotechnical risks. For example, in dam safety, life safety risk can be
Fig. 21. (Color) Systems approach to spillway safety. (Adapted from Hartford et al. 2016; Leveson 2012; Rasmussen and Svedung 2000.)
simulation to study complex systems is now common in many sci- systematic uncertainties to be separately considered and the pos-
entific disciplines, especially when exploring questions that are sible correlations they generated to be accommodated implicitly.
poorly suited to traditional reliability analysis. Today, simulation A similar approach is possible with simulation. First, a simulation
approaches are used in fields as disparate as oil and gas exploration, model is developed containing only the aleatory uncertainties of
medicine, materials science, urban planning, and aerospace (NSF interest to the problem (Fig. 22). The probabilities of the aleatory
2006). Over the past decade, the US National Research Council has uncertainties are taken to be conditional on the values of the epi-
published more than a dozen reports on the use of simulation in stemic uncertainties. A second simulation loop of just the epistemic
engineering science (e.g., NRC 2002, 2008, 2010b). uncertainties is nested around the first aleatory loop. For each reali-
zation of the uncertainties in the epistemic loop, the aleatory loop is
iterated many times. Then, the epidemic loop is itself iterated many
Nested Simulation times. Finally, the two sets of results are combined. A similar ap-
A common way of approaching emergent behavior is by simula- proach was used by USACE in its risk analysis of the Herbert
tion. This might be using a Monte Carlo model, systems dynamics, Hoover Dikes (USACE 2014).
agent-based models, or one of many other similar approaches
(Gianni et al. 2018). Monte Carlo simulation of a rudimentary kind Human Factors
had been used on both Tonen and IPET. On Tonen, it was used to
A third thing that was learned on the SSRP was the importance of
extrapolate the relatively small event trees for individual patios to
human factors to operational safety. This has been emphasized in
the larger interactions among the many components of the entire
recent forensic investigations of dam incidents and failures (FERC
facility. On IPET, it was used to understand length effects on levee
2018, 2022) and has been discussed in specific detail by Alvi
reliability and to investigate combinations of failures (e.g., gate
(2013). Humans are tasked with making decisions and implement-
openings, levee penetrations, and structural discontinuities).
ing actions to operate a system in a safe and efficient way. This
The simulation approach goes beyond traditional reliability that
involves both cognitive and physical responses. Depending on the
considers ultimate geotechnical, structural, hydraulic, and other
event sequence, human errors can lead to consequences ranging
capacities treated separately to one that considers many simultane-
from minor to the shutdown of a facility or to failure. It is rea-
ous ways that a system can perform adversely (Hartford et al. sonable to incorporate human reliability in systems models as is
2016). There are good reasons for the increasing popularity of sim- routinely done in nuclear, oil and gas, and other hazardous indus-
ulation in risk analysis: simulation allows complex systems to be tries. This topic is beyond the scope of the present discussion, but
modeled easily and cheaply, computer speeds are increasing, dis- increasing attention is being paid to its importance.
crete events are readily included, interactions among differing types
of system response (e.g., physics of failure, communication and
control, and human reliability) can be represented, and the numeri- Path Forward
cal precision of simulation is often independent of the complexity
of the system being modeled. Geotechnical practice has come a long way in its use of risk and
It was observed on IPET and later in Panama that separating reliability thinking since Casagrande’s lecture in 1965. What started
epistemic and aleatory uncertainties in event tree analysis by using in the 1970s as relatively simple probabilistic versions of traditional
logic trees was powerful. This allowed model, parameter, and other design equations has evolved into risk analyses of large engineered
Fig. 22. (Color) Nested simulation for spillway safety. (Adapted from Komey 2018.)
systems, like those used in other engineering disciplines. These policy are needed. It seems inevitable that the observational
methods are now routinely applied to dam safety, large industrial approach, adaptive planning, and Bayesian methods of learning
projects, flood and coastal protection projects, offshore facilities, from experience will be a central part of that thinking.
and other large geotechnical works. They are used to inform infra- Another challenge is how society judges the tolerability of risk.
structure decisions, develop public policies, and ensure public safety. Today, life and safety risks posed by critical infrastructures vary by
The earlier Terzaghi Lectures of Casagrande, Whitman, and orders of magnitude. Regulatory agencies rely on the VSL, whereas
Christian each call out the importance of evaluated judgment in operating agencies rely on tolerable risks. Safety, however, is a so-
geotechnical practice. Data may seem objective, but they are al- cial and political question; engineering practice needs to develop
ways filtered through the lens of geology and experience. The cur- consistent and rational guidance.
rent generation of risk-informed tools that have evolved since those
lectures are consistent with that way of thinking, placing it on a
coherent conceptual footing. Data Availability Statement
Nonetheless, challenges remain. Among the most obvious chal-
lenges continues to be education. There remains confusion in prac- No data, models, or code were generated or used during the study.
tice over the nature of probability in geotechnical engineering and
the meaning of statements made about uncertainty. Whereas a ru-
dimentary understanding of probability and statistics has become Acknowledgments
necessary in most technical professions, it has languished in geo- This paper is dedicated to the memory of John T. Christian who died
technical engineering. Most practitioners appear to accept the fun- during its final editing. The author is indebted to a career-long col-
damentally subjective character of uncertainty in geotechnical laboration. The author is also indebted to the early workers in geo-
practice, and most apply some form of Bayesian thinking, inverse technical risk and reliability, many of whom he had the privilege of
probability, and observational methods whether or not done so working with and learning from. Among these are the MIT cadre
explicitly or even knowingly. from the 1970s and 80s (alphabetically): C. Allin Cornell, Herbert
We also face conceptual challenges. One is the increasing non- H. Einstein, Ronald C. Hischfeld, Charles C. Ladd, W. Allen Marr,
stationary of natural hazards. Whereas in the past the historical re- Erik Vanmarcke, Daniele Veneziano, and Robert V. Whitman. He is
cord could be extrapolated to predict the future; in a nonstationary similarly indebted to the many others who have influenced the sub-
world of climate change, the past may no longer be a prelude. sequent years and provided a review of this paper, among whom are:
Our forecasts of future hazards increasingly are functions of the Luis Alfaro, Tony Bennett, Karl Dise, Charles Dowding, Jerry
epistemic uncertainties of models and scenarios rather than the Foster, Gerald E. Galloway, Robert B. Gilbert, Fernando Guerra,
aleatory uncertainties of the historic record. Risk methodology Desmond N. D. Hartford, Suzanne Lacasse, Lewis (Ed) Link,
is up to the task, but new conceptual approaches to design and B. K. Low, Martin W. McCann, Ross McGillivray, Samuel G.
-7535(97)00052-0. actor safety study: The introduction of risk assessment to the regulation
Rasmussen, J., and I. Svedung. 2000. Proactive risk management in a dy- of nuclear reactors. Washington, DC: USNRC.
namic society. Karlstad, Sweden: Swedish Rescue Services Agency. Vanmarcke, E. H. 1977a. “Probabilistic modeling of soil profiles.” J. Geo-
Rausand, M., and S. Haugen. 2020. Risk assessment: Theory, methods, and tech. Eng. Div. 103 (11): 1227–1246. https://doi.org/10.1061/AJGEB6
applications. Wiley series in statistics in practice. Hoboken, NJ: Wiley. .0000517.
Reason, J. T. 1997. Managing the risks of organizational accidents.
Vanmarcke, E. H. 1977b. “Reliability of earth slopes.” J. Geotech. Eng.
Aldershot, UK: Ashgate Publishing Company.
Div. 103 (11): 1247–1265. https://doi.org/10.1061/AJGEB6.0000518.
Reclamation. 2011. Dam safety public protection guidelines (interim).
Vick, S. G. 2002. Degrees of belief: Subjective probability and engineering
Denver: US Bureau of Reclamation.
judgment. Reston, VA: ASCE.
Regan, P. 2010. Dams as systems—A holistic approach to dam safety.
Vrouwenvelder, A. C. W. M. 2003. “Uncertainty analysis for flood defense
Sacramento, CA: US Society on Dams.
systems in the Netherland.” In Proc., ESREL, European Safety and
Robert, C. P., and G. Casella. 2010. Monte Carlo statistical methods.
Reliability Conf. 2003, 11–18. Maastricht, Netherlands: A.A. Balkema
New York: Springer.
Schweckendiek, T. 2014. On reducing piping uncertainties—A Bayesian Publishers.
decision approach. Delft, Netherlands: Delft Univ. of Technology. Whitman, R. V. 1984. “Evaluating calculated risk in geotechnical engi-
Stevens, S. S. 1946. “On the theory of scales of measurement.” Science neering.” J. Geotech. Eng. 110 (2): 143–188. https://doi.org/10.1061
103 (2684): 677–680. /(ASCE)0733-9410(1984)110:2(143).
Terzaghi, K. 1961. “Past and future of applied soil mechanics.” J. Boston Wu, T. H. 1974. “Uncertainty, safety, and decision in soil engineering.”
Soc. Civ. Eng. 48: 110–139. J. Geotech. Eng. Div. 100 (3): 329–348. https://doi.org/10.1061
Tversky, A., and D. Kahneman. 1971. “Belief in the law of small numbers.” /AJGEB6.0000026.
Psychol. Bull. 76 (2): 105–110. https://doi.org/10.1037/h0031322. Wu, T. H. 2011. “2008 peck lecture: The observational method: Case his-
URS. 2009. Delta risk management strategy executive summary. tory and models.” J. Geotech. Geoenviron. Eng. 137 (10): 862–873.
Sacramento, CA: California Department of Water Resources. https://doi.org/10.1061/(ASCE)GT.1943-5606.0000509.
USACE. 1972. New Orleans east lakefront levee Paris road to south point, Yegian, M. K., and R. V. Whitman. 1978. “Risk analysis for ground failure
Lake Pontchartrain. Washington, DC: USACE. by liquefaction.” J. Geotech. Eng. Div. 104 (7): 921–938. https://doi.org
USACE. 2007a. Volume IV: The storm in Performance evaluation of the /10.1061/AJGEB6.0000672.
New Orleans and southeast Louisiana hurricane protection system: Zhou, W.-H., Z.-Y. Yin, and K.-V. Yuen. 2021. Practice of Bayesian prob-
Final report of the interagency performance evaluation task force. ability theory in geotechnical engineering. Singapore: Springer.
IPET4. Washington, DC: USACE. Zielinski, P. A. 2021. “Dams as complex systems.” In Proc., Annual Meet-
USACE. 2007b. Volume V: The performance levees and floodwalls in ing of the Canadian Dam Association, 12. Markham, ON, Canada:
Performance evaluation of the New Orleans and southeast Louisiana Canadian Dam Association.