You are on page 1of 36

C H A P T E R

2
Contemporary Methods for Statistical
Design and Analysis
D.R. Fox1,2
1
Environmetrics Australia, Beaumaris, VIC, Australia; 2University of Melbourne,
Parkville, VIC, Australia

2.1 INTRODUCTION This chapter deals with a number of facets


associated with the statistical aspects of
The present chapter deliberately avoids modern-day ecotoxicology. It is neither complete
rehashing the familiar statistical concepts nor comprehensive and invariably reflects the
covered in any university “STAT101” course. author’s biases and interests. For example,
Not only are there numerous texts on all aspects models of contaminant uptake, transport, and
of “classical” statistical theory and practice (eg, fate have not been discussed nor is there any dis-
Cox and Snell, 2000; Bickel and Doksum, 2016; cussion of stochastic models used to describe
Kutner et al., 2016), but there now exists a sub- population dynamics. However, the topics that
stantial body of “guideline” material, including are covered are hopefully sufficiently representa-
statistical guidance documents, that has been tive of the main statistical issues and challenges
amassed by environmental jurisdictions around facing ecotoxicologists in their pursuit of sustain-
the world including Australia, New Zealand, able ecosystem protection.
the European Union, Canada, and the United
States. Accordingly, this chapter attempts to
shift the focus away from where we have been 2.2 TOXICITY MEASURES
and more toward where we need to go. For
example, redeploying the scarce experimental To monitor and manage contaminants (ie, po-
resources that are liberated by abandoning tential toxicants) in a marine environment, it is
wasteful analysis of variance (ANOVA) proce- necessary to measure both their concentration
dures for computing no observed effect concen- in receiving waters and their “effects” on pre-
trations (NOECs) and actually modelling the scribed components of the ecosystem. Effects
concentrationeresponse (C-R) phenomenon can be either acute or chronicdthe difference
rather than reducing it to a trite statement of primarily a function of dose which is the product
“no observed effect at concentration x.” of concentration and time. Acute effects are those

Marine Ecotoxicology
http://dx.doi.org/10.1016/B978-0-12-803371-5.00002-3 35 Copyright © 2016 Elsevier Inc. All rights reserved.
36 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

that result from short-term exposure (relative to • ECx (effect concentration)dthe concentration
the organism’s life span) to high concentrations; at which x% of organisms will exhibit an
chronic effects result from long-term exposure “effect”;
(relative to the organism’s life span) to relatively • ICx (inhibitory concentration)dthe
low concentrations. concentration at which x% “impairment”
The ability of a substance to cause acute and/ occurs;
or chronic effects in an organism is referred to as • NEC (no effect concentration)dthe highest
its toxicity. What constitutes an “effect” is a concentration below which no effect occurs;
critical question and one not always easily • NOEC/NOAEL [no observed (adverse) effect
answered; however, ecotoxicologists use a vari- concentration]dthe largest of a small number
ety of endpoints ranging from sublethal (for of discrete test concentrations for which the
example, compromised reproductive capability) mean response is not statistically different
to lethal (death), which define the dependent vari- from the control response;
ables in C-R experiments (discussed in greater • LOEC/LOAEL [lowest observed (adverse)
detail in Section 2.6). effect concentration]dthe smallest of a small
The quantification of toxicity is most number of discrete test concentrations for
commonly assessed from an examination of an which the mean response is statistically
organism’s response (with respect to a particular different from the control response;
endpoint) to changes in toxicant concentration • MATC (maximum acceptable toxicant
giving rise to the so-called C-R curve. concentration)ddefined as the geometric
The development of toxicity metrics is rooted mean of the NOEC and LOEC;
in human toxicity studies that date back to the • BECx (bounded effect concentration)dthe
turn of the 20th century. In 1908 Theodore Cash highest concentration for which one may
published a paper in the British Medical Journal, claim (with 95% confidence) that its effect
in which he described early doseeresponse ex- does not exceed x% (Hoekstra and van
periments (Cash, 1908). In his paper he talks of Ewijk, 1993)
“the nearest fulfillment of a mathematical rela-
Not surprisingly, this array of metrics has
tionship seemed to be achieved by working up-
motivated numerous studies to compare their
wards from that amount of any drug which
efficacy as regulatory instruments as well as to
produced the minimum of appreciable action.”
reconcile the implications of their use for
This dose, he notes, was variously referred to as
ecosystem protection (Crane and Newman,
the “Grenzdose” or “limit dose.” Cash (1908) also
2000; Hose and Van den Brink, 2004; Payet,
introduced the additional terms “minimal effective
2004; Shieh et al., 2001; van der Hoeven et al.,
dose” and “maximal ineffective dose,” although it
1997). This is an important issue that is far
was later suggested these terms be replaced with
from resolved. The choice of which particular
the median lethal dose or LD50 (Trevan, 1927). How-
toxicity measure to use in practice is largely
ever, note that an LC refers to a concentration
dictated by the requirements of standard testing
outside of biological tissues, while an LD refers
protocols. Where flexibility of choice exists, it is
to a concentration inside biological tissues.
unfortunately true that practitioners often treat
What is interesting is the plethora of toxicity
these measures as either exchangeable or
metrics that have since emerged. Those
related via a simple scaling. The most common
commonly used in ecotoxicology include:
example of the latter is the arbitrary scaling of
• LDx or LCxdthe dose/concentration that is measures of acute toxicity by some order of
lethal to x% of organisms; magnitude in an attempt to harmonize the
2.2 TOXICITY MEASURES 37
results with measures of chronic toxicity. Statis- In the event there are insufficient chronic
tically, there is no justification for this although toxicity data the Guidelines (OECD, 2014)
Fox (2006) did attempt to at least remove the recommend that acute EC50 and/or IC50 and/
arbitrariness of the scaling factordotherwise or LC50 be converted to “chronic equivalent
known as the acute-to-chronic ratio or ACRdby data” for the derivation of a protective
allowing the data to determine an “optimal” concentration.
value for the ACR. While the science of toxicity metrics con-
The regulatory use of toxicity measures is tinues to evolve, for example, the use of
confused and confusing. For many years NOECs Bayesian methods (Link and Albers, 2007; Fox,
and LOECs were the preferred measures of 2010; Cliffroy et al., 2013; Zhang et al., 2012;
toxicity although more recently their use has Grist et al., 2006; Jaworska et al., 2010), the chal-
attracted strong criticism (Chapman et al., 1996; lenge for regulators and practitioners is to
Jager, 2012; Fox, 2008; Fox et al., 2012) and ensure that toxicity data and the statistical
even calls to ban their use altogether (van Dam methods used to generate them are (1) ecologi-
et al., 2012; Warne and van Dam, 2008; Landis cally relevant, (2) fit-for-purpose, (3) robust to
and Chapman, 2011). However, such views moderate violations of assumptions, (4) scientif-
have themselves been criticized with others ically defensible, and (5) statistically credible.
arguing that calls to ban the NOEC are misin- We shall return to these themes in Section 2.7
formed and the result of simplistic statistical when we consider statistical issues associated
thinking (Green et al., 2013). with Species Sensitivity Distributions (SSDs).
Jurisdictions around the world have pub- At this stage, a number of important statistical
lished various “Guideline” documents and considerations arise:
while most articulate preferred general strate-
1. The nature of the endpoint determines the
gies, many leave the detail up to the analyst.
type of random variable and associated
For example, the Organization for Economic
statistical treatment of toxicity data. Random
Cooperation and Development (OECD) Guide-
variables are either discrete or continuous. As a
lines recommend the use of model-based
rule-of-thumb, if the endpoint is the result of
toxicity measures although “the choice is left
observation (eg, mortality status) rather than
to the reader” as far as the statistical detail is
measurement, then the random variable is
concerned (OECD, 2014). At the other end of
discrete and the data are typically obtained
the spectrum, the recently revised Australian
as a result of counting (number of dead
and New Zealand Guidelines for toxicants in
organisms). Alternatively, if the endpoint
fresh and marine waters takes a more prescrip-
requires physical measurement of some
tive approach (Warne et al., 2013). With respect
property (eg, length of fish), then the
to chronic toxicity measures, these guidelines
random variable is continuous (eg, growth
give preference in the following order:
rate). This distinction is not simply
1. NEC; academicdit is a necessary (but not
2. ECx/ICx/LCx where x  10 (NB: all three sufficient) requirement to satisfy (2) above.
equally ranked); Unfortunately insufficient attention is paid to
3. BEC10; this point in many ecotoxicological studies
4. ECx or LCx where 15  x  20; with the result that inappropriate statistical
5. NOEC or NOEC estimated from MATC, tests and/or models are used with
LOEC, or LC50 values. incompatible data constructs, for example,
38 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

t-tests and ANOVA used to analyze binary more specifically (Environment Canada, 2005;
(eg, dead/alive) data obtained from small CCME, 2007; European Commission, 2011;
(eg, <20) samples. Newman, 2012; ANZECC/ARMCANZ, 2000a,b;
2. The arbitrary scaling and pooling of toxicity OECD, 2012, 2014) although most of this is based
metrics to generate “sufficient” data from an on and repeats standard frequentist principles
assumed common distribution of toxicity (as that are taught in all introductory statistics
is done in SSD modeling) is difficult to justify courses (Sparks, 2000). The OECD (2012) and
ecologically and impossible to justify Environment Canada (2005) documents exem-
statistically. The (implicit) assumption that plify this point. Topics covered in both docu-
the sample percentiles of toxicity data ments include hypothesis testing, Type I/II
corresponding to either a single endpoint errors, statistical power, randomization, replica-
across multiple species or multiple endpoints tion, outliers, and data transformations. While
across a single species have the same these statistical concepts are important, it could
underlying statistical distribution is surely be argued that the emphasis placed on clas-
false. One only needs to consider the sical/frequentist statistics has stifled the devel-
dichotomy of discrete/continuous opment of strategies and procedures that are
distributions arising from different endpoints better equipped to handle the many and varied
to make this assertion. perturbations of assumed conditions encoun-
3. The statistical properties of a toxicity metric tered in ecotoxicology. The assumption of
need careful elicitation and evaluation before randomness is a case in point. It is safe to say
contemplation for use. It is insufficient to that the majority of statistical methods are pred-
appeal to heuristics or intuition when arguing icated on the joint notions of randomness and in-
the case for the adoption of a “new” metric, dependence. Indeed statistical theory demands
procedure, or estimation strategy. It may well that toxicity data used in SSD modeling be ob-
be that the use of an arbitrary ACR of 100 is tained from a randomly selected sample of spe-
expected to yield a conservative estimate of a cies. The well-accepted reality is that this is
“protective concentration,” but such an never the case (Fox, 2015 and references therein).
approach completely ignores the competing The irony is that while guideline documents
risks and objectives in any environmental stress the importance of randomness (eg,
assessment. There are at least two ways we “randomization should prevail in all aspects of
“can get it wrong” in setting environmental the design and procedures for a toxicity test,”
standards (cf. Type I and Type II errors in Environment Canada, 2005), they simulta-
statistical hypothesis testing). Using a limit neously recommend procedures that ensure
which fails to adequately protect the samples are biased. For example, the revised
environment is one way. Using a limit which Australian and New Zealand Water Quality
inappropriately denies human activity is Guidelines recommend using toxicity data
another. from at least eight species from at least four taxo-
nomic groups (Batley et al., 2014). Such purposive
sampling is the antithesis of randomness and
2.3 DESIGN CONSIDERATIONS while the distinction between probability sampling
and judgmental sampling for ecotoxicological
Much has been written about the design and studies is not always made clear, good advice
analysis of environmental data in general does exist albeit in the broader context of envi-
(USEPA, 2002, 2006) and ecotoxicological data ronmental monitoring (USEPA, 2002, 2006).
2.3 DESIGN CONSIDERATIONS 39
So while not diminishing the importance of documents, the remainder of this section is
sound statistical design to guide all aspects of devoted to the exploration of some more
the data collection and analysis process, the “contemporary” aspects of experimental design
reality is that ecotoxicological studies tend to in ecotoxicology.
be severely constrained by (1) high cost of ANOVA techniques have (and continue) to
data acquisition; (2) inability (or compromised play a significant role in the analysis of ecotoxi-
ability) to invoke the core statistical principles cological data. Even though the use of this tech-
of randomness, replication, and blocking; and nique is expected to diminish as scientists move
(3) nonconformity. The high cost of data acqui- away from generating NOEC data, ANOVA
sition is a function of the logistics of field sam- methods still have an important role to play in
pling coupled with the expensive laboratory testing hypotheses concerning the toxic effects
analyses required to generate toxicity data. of chemicals in the environment. The identifica-
The strict definition of randomness means that tion of an appropriate experimental design is a
every “unit” in the target “population” under critical first step in the use of ANOVA and
investigation has an equal chance of being related tools of statistical inference. At the very
included in the sample and we have already least the experimental design should be such
seen that protocols exist (eg, Australian Water that it:
Quality Guidelines, ANZECC/ARMCANZ,
• allows for the unbiased and efficient estimation
2000a,b) which ensure this cannot happen. In
of all effects of interest;
addition, standardized protocols for
• controls (to the extent possible) sources of
laboratory-based toxicity tests tend to be avail-
extraneous variation likely to affect the
able for only a relatively small number of ani-
measured response (for example, a
mals or organisms thus ensuring another layer
temperature or salinity effect in a C-R
of nonrandom selection. Replication improves
experiment); and
the quality of estimation and inference but is a
• makes minimal use of limited resources.
casualty of (1), while control through the use
of “blocking” where experimental units are Reducing bias, improving precision, and con-
organized according to some other exogenous trolling extraneous variation tend to result in a
variable(s) is only an option if the major sources greater number of treatment combinations
of extraneous variation are known in advance. and/or increased replicationdboth of which in-
“Nonconformity” refers to the propensity of crease the cost of the experiment. It is therefore
ecotoxicological data to violate many of the surprising that orthogonal fractional factorial de-
prerequisites or assumptions required by most signs have not been used more widely in ecotox-
statistical tests and procedures described in icology (Dey, 1985). While it is not possible to
the various guideline documents. These provide a comprehensive treatment of this
include, but are not limited to, violations of as- important topic in this book, we illustrate the
sumptions concerning: independence; distribu- potential benefits with the use of a simple
tional form; variance structures; sample size; example.
outliers; censoring; and response-generating In a recent paper, Webb et al. (2014) described
mechanism. a toxicology experiment that “did not succumb to
Rather than summarizing standard statistical standard experimental design.” The challenge
design theory, which is readily available in text- was to satisfy the requirements of the three dot-
books (eg, Hinkelmann and Kempthorne, 2008; points above in a way that accommodated unique
Gad, 2006) and the aforementioned guideline physical and logistical constraints. Their solution
40 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

relied upon advanced mathematical and compu- with the use of a fraction of the treatments in
tational skillsdthe detail of which is beyond the the full designdhence the name fractional facto-
scope of this book. The situation described in rial design. In addition, if the treatments
Webb et al. (2014) motivates the following hypo- comprising this fraction are carefully selected,
thetical example illustrating the use of a fractional it is possible to estimate the main effects inde-
factorial design. pendently of each otherda property that is
clearly desirable but not guaranteed by either
a random or subjective selection. In statistics,
2.3.1 Example the independence of two random variables has
A study into the potential impacts associ- the geometrical interpretation of orthogonality
ated with the discharge of waste water from a (ie, being at right angles to each other). Hence,
proposed desalination plant relied on toxicity fractional factorial designs which permit the in-
testing. For one of these tests, researchers dependent estimation of effects are referred to
were interested in whether or not the hypersa- as orthogonal fractional factorial designs. These
line effluent was toxic to marine organisms. are not new and date back to the work of
Other factors thought to be important were Adelman (1961), Bose and Bush (1952), Rao
the time of day (TOD) when exposure to the (1950), Kempthorne (1947), and others. As one
toxicant commenced as well as the temperature might expect, the method of identifying how
(temp) and salinity (salin) of the waste stream. many and which treatments to include in the
As in the Webb et al. (2014) study, the manner fractional design such that the orthogonality
in which test samples were stored was poten- requirement is met is far from simple and re-
tially another source of variation that needed quires a good understanding of advanced math-
to be controlled for. In this case, beakers could ematical concepts such as linear algebra,
be placed on shelves that were arranged in three Hadamard matrices, and Galois Field theory.
racks each having four shelves. The positioning Thankfully, statistical software tools make this
of beakers (representing different combinations task easier although more complex designs
of dose, TOD, temp, and salin) on shelves was tend not to be included. R (R Development
important due to the potential influences of Core Team, 2004) is the only free package of
light levels, proximity to the door, and thermal such tools; it has a large and rapidly increasing
stratification. The basic experiment involved the library of user-contributed functions including
direct manipulation of the four factors: dose packages for creating and analyzing fractional
(present/absent); TOD (am/pm); temp (15 C/ factorial designs. Readers interested in learning
25 C); and salin (ambient/elevated) coupled more should consult the R website (CRAN,
with the two factors determining beaker posi- 2015) and the texts by Lawson (2015) and
tion (racks and shelves). A “full factorial” exper- Gad (2006).
iment would require all 384 combinations of Returning to the present example, an illus-
these factors to be tested at least once. Not tration of the savings in experimental effort
only is such an experiment time-consuming is indicated by the experimental design
and expensive, it may be unnecessary. If infor- represented by the allocations in Table 2.1
mation is only sought on the “main effects” which has reduced the number of treatments
(ie, the effect of each factor separately) and from 384 to a mere 25. This design is what is
higher-order interaction effects are (or can be referred to as Resolution III, meaning that main
assumed to be) negligible, then significant sav- effects are estimated independently of each
ings in experimental effort can be realized other but not independently of interactions
2.4 DATA PROCESSING AND HANDLING 41
TABLE 2.1 Orthogonal Fractional Factorial Design resources (typically in the form of more treat-
for Effluent Toxicity Study ment combinations).
Run Shelf Rack Dose TOD Temp Salin
More will be said about “optimal” designs
in the context of planning a C-R experiment in
1 1 3 Absent am 15 Ambient Section 2.6.
2 2 3 Absent am 15 Ambient
3 3 3 Absent am 15 Ambient
2.4 DATA PROCESSING AND
4 3 3 Present pm 25 Elevated HANDLING
5 4 3 Present pm 25 Elevated
The rapid rise of the phenomenon referred to
6 1 1 Absent am 25 Elevated
as “big data” has rejuvenated interest in what
7 2 1 Absent pm 25 Ambient scientists regard as the necessary, but mundane,
8 3 1 Present pm 15 Ambient task of data storage, handling, and manipulation
(referred to by some as “statistical janitorial
9 3 1 Present am 15 Ambient
work,” New York Times, 2014).
10 4 1 Absent am 15 Elevated According to one definition “big data” is
11 1 1 Absent pm 15 Elevated characterized by “data sets with sizes beyond
the ability of commonly used software tools to
12 2 1 Present am 15 Elevated
capture, curate, manage, and process data
13 3 1 Present am 25 Ambient within a tolerable elapsed time” (Wikipedia,
14 3 1 Absent am 25 Ambient 2015). Areas of science generating big data
include astronomy, telecommunications, geno-
15 4 1 Absent pm 15 Ambient
mics, and natural resource management. While
16 1 2 Present am 25 Ambient these are worthy of investments in R&D effort,
17 2 2 Present am 15 Elevated concern has been expressed that the rush to
“get on board” with the big data push may be
18 3 2 Absent pm 15 Elevated
diverting attention away from the equally
19 3 2 Absent pm 15 Ambient important issue of “little data” (Environmetrics
20 4 2 Absent am 25 Ambient Australia, 2014a).
While ecotoxicological data sets may never
21 1 2 Present pm 15 Ambient
fall within the realm of big data, they can span
22 2 2 Absent pm 25 Ambient orders of magnitude in size. For example users
23 3 2 Absent am 25 Elevated of the USEPA ECOTOX database (USEPA,
2015) can download up to 10,000 records on sin-
24 3 2 Absent am 15 Elevated
gle chemical toxicity data, whereas SSDs are usu-
25 4 2 Present am 15 Ambient ally modeled using no more than 20
observations and often as few as 5 or 6.
As with design considerations (Section 2.3),
the various guideline documents (or accompa-
(hence the need to know or assume that inter- nying documents) published by jurisdictions
action effects are negligible). Other reso- around the world generally provide comprehen-
lution designs may be available which are less sive treatment of commonly used data analysis
restrictive but require greater experimental techniques. While most of this is useful, some
42 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

is outdated. For example Appendix I of the Ca- and interrogation. ggplot2 is a powerful graphics
nadian Guidelines (Environment Canada, 2005) package that allows users to interactively
is a blank sheet of logarithmic paper for plotting explore and display data by modifying a base
results of C-R experiments by hand. plot by adding and removing layers. Space re-
Another feature of all of these guideline doc- strictions preclude a more comprehensive dis-
uments is their strong reliance on flowcharts for cussion of these packages, however some of
the statistical treatment of data. While the statis- their capabilities are briefly examined next.
tically naïve will find these both useful and
comforting, rigid adherence to a highly struc-
tured approach to preliminary data analysis 2.4.1 Manipulating Data: The R Package
tidyr
runs counter to the objectives of exploratory
data analysis (EDA). Indeed the essence of EDA The tidyr and dplyr packages work hand-in-
is to uncover patterns, trends, anomalies, corre- hand. Both dplyr and tidyr are designed to
lations, outliers, and other important features of clean up or “mung” messy data (Quora, 2014).
a data set via a fairly unstructured approach us- tidyr works on the simple and consistent phi-
ing powerful computer graphics and special- losophy that columns of a dataframe (R’s termi-
ized software tools designed to tease out nology for a rectangular array of data) represent
hidden structure. There is no roadmap for this variables and rows represent observations. To
process. see how this works, consider the data in
The task of preparing, organizing, and manip- Table 2.2, which give the number of surviving
ulating data is both necessary and time- marine organisms (out of 100) for each of four
consuming, with some claiming these activities replicates as a function of toxicant concentra-
constitute up to 80% of the data analysis effort tion. Although this tabular presentation of re-
(Dasu and Johnson, 2003). The lack of consistent sults is compact, it is not well suited to further
advice or a standard code of practice is largely statistical analysis.
responsible for this situation. Wickham (2014) The function gather() fixes this. The R code
has identified the following common problems below illustrates how the tabular data stored
with digital data sets: in a CSV file called example_1.csv is read into
• Column headers are values, not variable
names; TABLE 2.2 Number of Surviving Organisms (out of
• Multiple variables are stored in one column; 100) for Each of Four Replicates as a
• Variables are stored in both rows and Function of Toxicant Concentration (%)
columns;
CONCEN (%) 1 2 3 4
• Multiple types of observational units are
stored in the same table; and 0 89 92 88 93
• A single observational unit is stored in 3.1 87 93 97 91
multiple tables.
6.3 96 91 90 94
To address the lack of guidance, Wickham 12.5 93 89 95 95
(2009) has developed the R packages tidyr,
dplyr, and ggplot2. The first two are used for 25 76 67 73 85
“data wrangling”dthat is the often tedious pro- 50 0 0 0 0
cedures associated with getting data into a
100 0 0 0 0
format that is amenable for further analysis
2.4 DATA PROCESSING AND HANDLING 43
R and then converted into the “standard”
format:

dat<-read.csv(“example_1.csv”) # read in tabular data


names(dat)<-c(“concen”,”1”,”2”,”3”,”4”) # assign column names
dat1<-gather(dat,rep,surv,2:5) # convert to standard format
head(dat1) # display first 6 rows of converted data

concen rep surv


1 0.0 1 89
2 3.1 1 87
3 6.3 1 96
4 12.5 1 93
5 25.0 1 76
6 50.0 1 0

An example of a data set in which variables There are a total of six variables in Table 2.3
appear in both rows and columns is shown (concentration, length, pH, salinity, dissolved ox-
in Table 2.3. This is a portion of data from a ygen, and replicate); these appear in the rows
larger experiment to assess the toxicity of and columns of the table. The function gather()
effluent from a desalination plant by exam- is first used to separate out the variables fol-
ining the effect of different effluent concentra- lowed by function spread() which creates multi-
tions (as a percentage of undiluted effluent) on ple columns according to levels of other factor(s).
macroalgal growth (as measured by gameto- Again, the R code below illustrates how the
phyte length). Information on the covariates tabular data stored in a CSV file called
pH, salinity, and dissolved oxygen was also example_2.csv are read into R and then con-
recorded. verted into the “standard” format.

Dat<-read.csv(“example_2.csv”) # read in tabular data


dat2<-gather(dat,type,rep,A:D) # stacks all reps into single col
names(dat2)[3:4]<-c(“rep”,”value”) # assign names to cols
dat3<-spread(dat2,type) # creates separate cols from ‘type’
str(dat3) # get details of structure of new dataframe

‘data.frame’: 28 obs. Of 6 variables:


$ concen: num 0 0 0 0 3.1 3.1 3.1 3.1 6.3 6.3 .
$ rep : Factor w/ 4 leve“s”"”"”"”"”"”"”"D": 1 2 3 4 1 2 3 4 1 2 ...
$ DO : num 98 98 98 98 103 ...
$ length: num 19 23.5 20.9 20.6 21 ...
$ pH : num 8 8 8 8 7.98 7.98 7.98 7.98 8.01 8.01 ...
$ salin : num 37.3 37.3 37.3 37.3 37.3 37.3 37.3 37.3 36.7 36.7 ...
44 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

TABLE 2.3 Portion of Gametophyte Length Data for a Macroalgal Growth Test.
Columns are Effluent Concentration, Measurement Type, Replicates AeD.

Concentration (%) Type A B C D

0 Length 19.027 23.476 20.908 20.559


0 pH 8 8 8 8
0 Salin 37.3 37.3 37.3 37.3
0 DO 98 98 98 98
3.1 Length 20.98 20.581 21.867 19.663

3.1 pH 7.98 7.98 7.98 7.98


3.1 Salin 37.3 37.3 37.3 37.3
3.1 DO 103.2 103.2 103.2 103.2

The result is dataframe dat3, which is now in create publication-quality graphics by interac-
“standard” format with six columns (variables) tively adding “layers.” These might consist of a
and 28 rows (observations). plot of the same variables but from a different
data set or “objects” such as a smooth fit gener-
ated by one of R’s packages.
2.4.2 Visualizing Data: The R Package The following example from Eduard Sz€ ocs’
ggplot2
website (Sz€ocs, 2015a) shows how ggplot2 can
According to its developer ggplot2 “is unlike be used to prepare an annotated SSD for chlor-
most other graphics packages because it has a pyrifos data taken from the USEPA ECOTOX
deep underlying grammar” (Wickham, 2009). database (USEPA, 2015). The data are stored in
The main strength of ggplot2 is its ability to an R dataframe called df (Fig. 2.1A and B).

df <- df[order(df$val), ] # rearrange toxicity data in ascending order


df$frac <- ppoints(df$val, 0.5) # use intrinsic function ppoints to compute
# empirical estimates of cumulative
# probabilities
#
# Next use ggplot2 to build up SSD
#

require(ggplot2) # load the ggplot2 package


p<-ggplot(data ¼ df) # sets up base layer of plot and stores as
# an R object

p<-pþgeom_point(aes(x ¼ val, y ¼ frac), size ¼ 5)


# adds a layer of points to base layer and
# stores back into object p

p # plot the object simply by naming it


# see Figure 2.1A
2.4 DATA PROCESSING AND HANDLING 45
# Next use log-scale for concen, add species and axis labels,
# and change background theme.

p<-pþgeom_text(aes(x ¼ val, y ¼ frac, label ¼ species), hjust ¼ 1.1,


size ¼ 4) þ theme_bw() þ scale_x_log10(limits¼c(0.0075, max(df$val))) þ
labs(x ¼ expression(pas‘e(’Concentration of Chlorpyrifos‘[ ’, m‘, ‘g
’, L^-‘, ’ ]’)), y‘¼ ’Fraction of species affec’ed’)

p # plot the object simply by naming it


# see Figure 2.1B

(A) 1.00

0.75
frac

0.50

0.25

0.00
0 10 20 30 40 50
val
(B)
1.00 Chironomus plumosus
Pteronarcys californica
Plecoptera
Parapoynx stratiotata
Xanthocnemis zealandica
Notonecta maculata
Molanna angustata
Heptageniidae
0.75 Ranatra linearis
Corixa punctata
Fraction of species affected

Plea minutissima
Anax imperator
Sigara arguta
Neoplea striola
Sialis lutaria
Claassenia sabulosa
Anopheles quadrimaculatus
0.50 Anisops sardeus
Peltodytes sp.
Pteronarcella badia
Leptoceridae
Chaoborus obscuripes
Caenis horaria
Culex pipiens ssp. molestus
Chironomus dilutus
Cloeon dipterum
0.25 Chironomus tentans
Ephemerella sp.
Culex quinquefasciatus
Simulium vittatum
Atalophlebia australis
Aedes taeniorhynchus
Chironomus riparius
Procloeon sp.
0.00 Deleatidium sp.

0.1 10.0
Concentration of Chlorpyrifos [µg L–1]

FIGURE 2.1 Using ggplot2 to create empirical SSD: (A) base layer; (B) annotated plot.
46 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

2.5 ESTIMATION AND INFERENCE parameter(s), while in Bayesian statistics, data


are used to update the prior probability distribu-
Statistical inference is concerned with the dual tions using Bayes’ formula. The updated distri-
problems of estimation and hypothesis testing. butions are known as posterior densities and are
While these are related, there are nevertheless the basis for subsequent parameter inference.
subtle differences and objectives (Fig. 2.2). The two modes of estimation and inference
Estimation is concerned with quantifying un- have proven highly divisive among the statisti-
known model parameters in a manner that is in cal community with much journal space and
some way “optimal.” Hypothesis testing proce- many conferences devoted to arguments about
dures, on the other hand, commence with a state- the relative merits of the two approaches. Thank-
ment concerning the value of an unknown fully those tensions have largely dissipated with
parameter and use the information contained many scientists now acknowledging the legiti-
in a sample to assess the plausibility of this state- macy of both approaches. The adoption of either
ment. Two statistical paradigms for problems of a Bayesian or Frequentist framework for any
inference have emerged: Frequentist statistics (ie, given study is likely to reflect considerations of
the “classical” statistics referred to in Section personal preference, ease-of-use, and efficacy.
2.1) and Bayesian statistics. Most of the statistical As the use of Bayesian techniques in ecotoxi-
procedures used in ecotoxicology are frequentist cology is expected to steadily increase, the chal-
based although Bayesian methods are gaining lenge posed for regulators is how to
popularity (Evans et al., 2010; Fox, 2010; Billoir accommodate the element of subjectivity that
et al., 2008). invariably accompanies this mode of inference.
The main difference between the two para- Indeed, much of the “controversy” referred to
digms is that Frequentist statistics treat model above has centered on this issue. It is perhaps
parameters as fixed but unknown constants, because the Frequentist framework is “data-
while Bayesian statistics treat the model param- driven” and treats subjective probability as inad-
eters as random variables characterized by proba- missible that it has enjoyed such a prominent
bility distributions (referred to as a prior place in setting environmental standards. On
densities). Data are used in a Frequentist mode the other hand, it is precisely the element of
to find an “optimal” estimate of the unknown subjectivity in a Bayesian approach that many
argue provides for a richer, more informative
analysis. An attractive feature of the Bayesian
framework is that the prior density can be used
as the vehicle by which expert opinion is intro-
duced into the analysis. However, as noted by
Barnett and O’Hagan (1997), the process of elic-
iting information from one or more experts “is
fraught with technical difficulties.” More recent
studies using structured elicitation techniques
have sought to reduce the sources of expert
bias and error in the presence of epistemic uncer-
tainty (McBride et al., 2012).
FIGURE 2.2 Conceptual representation of statistical However, concerns remain about the use of
inference in an ecotoxicological setting. Taken from Fox,
D.R., Billoir, E., Charles, S., Delignette-Muller, M.L., Lopes, C.,
subjective information in the derivation of
2012. What to do with NOECs/NOELs e prohibition or innova- “safe” concentrations of toxicants in marine
tion? Integr. Environ. Assess. Manag. 8, 764e766. and fresh waters although, as noted by Fox
2.5 ESTIMATION AND INFERENCE 47
(2015), approaches using Frequentist statistical zero and variance s2ε . The simplest version of
methods are not as objective as many would Eq. (2.1) corresponds to the simple regression
think. It is also difficult to see how a “protective” of Y on x, that is:
concentration derived using Bayesian methods
Yi ¼ a þ bxi þ ei (2.2)
would stand up to scrutiny in a court of lawd  
even if the underlying methodology is sound. In this case Y ¼ [y1, y2,.,yn]; X ¼ 1
T
xT ; T

One can foresee the credibility of a Bayesian- bT ¼ [a b] where 1 is a (1  n) vector of ones


derived figure being challenged on a number and x is a (1  n) vector containing the values
of fronts. For example, it might be argued that of the covariate x. When some or all of the cova-
the selection of “experts” responsible for eliciting riates are factors, then the corresponding x is a
priors in the Bayesian analysis was biased or that vector of “dummy” codes. For example, x might
the inevitable comparison with the result from a simply indicate the presence or absence of a toxi-
more traditional Frequentist analysis yields cant in which case the values assigned to x could
inconsistent and/or irreconcilable differences. be {0, 1}.
An example of the latter is provided in Cliffroy Irrespective of the whether X contains all
et al. (2013) who compared HC5 estimates (the measured values, all dummy codes, or a combi-
5% hazardous concentrations derived from nation of both, the parameters of the model are
SSDs) using both Frequentist and Bayesian ap- estimated using the same equation. That is:
proaches. While they found that their Bayesian  1
method tended to underestimate the reference b
b ¼ X T X XT Y (2.3)
value, discrepancies between the two ranged
Furthermore, the varianceecovariance matrix
from a factor of 0.2 to a factor of nearly 6. While
of the estimated parameters is:
the propensity for underestimation was
h i  1
regarded by the authors as a positive feature of
Cov bb ¼ s2ε XT X (2.4)
the approach due to the increased level of protec-
tion it afforded, it is a view unlikely to be shared A key understanding is how the statistical
by all stakeholders in an environmental planning design influences the quality of inference. In the
and assessment process. case of the simple linear regression above, it is
Although it is not possible to give a compre- readily verified that (XTX) is given by Eq. (2.5)
hensive account of statistical estimation proce- and its inverse given by Eq. (2.6).
dures here, it is relevant to mention the general 2 3
linear model as it is the foundation of many com- P
n
6 n xi 7
mon methods such as regression, ANOVA, and 6 i¼1 7
XT X ¼ 6 n 7 (2.5)
analysis of covariance (ANCOVA) (Graybill, 4P Pn
25
1976). The general model common to all of these xi xi
i¼1 i¼1
techniques is given by Eq. (2.1).
 1 1
Y¼Xbþε (2.1) XT X ¼ !2
Pn Pn
where Y is a (n  1) vector of responses; X is a n 2
i¼1 xi  i¼1 xi
(n  p) “design” matrix that contains coding to 2 P 3
describe the experimental factors and/or values n P
n (2.6)
x2  xi
of covariates {x1, x2,.}; b is a (p  1) vector of 6 i¼1 i 7
6 i¼1 7
parameters to be estimated; and ε is a (n  1) 6 7
4 P n 5
vector of random error terms assumed to be  xi n
independently normally distributed with mean i¼1
48 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

Thus we see from Eqs. (2.4) and (2.6) that the in the diagonal entries all being equal to 25. In
variances and covariances of the estimated deciding how good our design is, it would be
parameters are entirely a function of the xi’s useful to have a measure of the overall efficiency
(for a given s2ε ) and not of the responses, yi. of the proposed design with the ideal design.
Furthermore, that the off-diagonal entry in One way of doing this is to examine the determi-
Eq. (2.6) is nonzero tells us that the parameters nant (Wikipedia, 2016) of the matrix (XTX)1.
{a, b} are not estimated independently (unless Thus, one criterion for selecting a “good” frac-
Pn
i¼1 xi ¼ 0).
tional factorial design is one which minimizes
Returning to the fractional factorial design this determinant. This is the basis of D-
given by Table 2.1 in Section 2.3 the factors and optimality, which is discussed further in Section
number of levels are shelf (4), rack (3), dose (2), 2.6. For the present example, it can be verified
TOD (2), temp (2), and salin (2). Without going that the design given in Table 2.1 is approxi-
into the details, the number of parameters p mately 70% efficient when compared to a full
required in the design matrix X in Eq. (2.1) is factorial design.
P As mentioned at the beginning of this section,
p ¼ kj Lj  k þ 1 where k is the number of fac-
estimation and hypothesis testing procedures
tors and Lj is the number of levels for factor j. So
overlap, although the emphasis is different. Esti-
for the design of Table 2.1, p ¼ 15  6 þ 1 ¼ 10
mating parameters of SSDs or complex C-R
resulting in a (25  10) design matrix X. For the
models often involves the use of sophisticated
coding of X (not given here) corresponding to
mathematical and statistical tools. Even more
the data in Table 2.1, the following XTX matrix
challenging is the determination of standard er-
was obtained:
rors of quantities derived from the fitted model.
2 3
25 0 0 0 0 0 0 0 0 0 As an example Eq. (2.7) is a four-parameter logis-
6 0 15 0 0 0 0 0 0 0 0 7 tic function used to model C-R data.
6 7
6 7
6 0 0 15 0 0 0 0 0 0 0 7
6
6 0
7 b1  b0
6 0 0 15 0 0 0 0 0 0 77 y ¼ b0 þ (2.7)
6 7 1 þ expfb2 ½ln x  ln b3 g
6 0 0 0 0 10 0 0 0 0 0 7
X X¼6
T
6 0
7
6 0 0 0 0 10 0 0 0 0 77 This is a complex, nonlinear model. Not only
6 7 it is nonlinear in concentration (x), but it is also
6 0 0 0 0 0 0 25 0 0 0 7
6 7
6 0 0 7 nonlinear in the parameters {b0, b1, b2, b3}. There
6 0 0 0 0 0 0 25 0 7
6 7 is no simple way of estimating the parameters;
4 0 0 0 0 0 0 0 0 25 0 5
this will invariably require the use of specialized
0 0 0 0 0 0 0 0 0 25
computer software such as the R package drc
The orthogonality of the fractional factorial discussed in Section 2.9. Once fitted, the model
design is immediately apparent from the fact is generally used to estimate the response b y 0 cor-
that all off-diagonal entries in XTX are zero. As responding to a concentration x0 by replacing
mentioned earlier, this is desirable since all parameters
n in
o Eq. (2.1) by their estimates
main effects will be estimated independently of b b1; b
b0; b b2; b
b 3 and/or estimate the concentra-
each other. Furthermore, the diagonal structure
tion b
x 0 for a given response y0 using Eq. (2.8).
of XTX means that the inverse is readily obtained
8   9
by replacing the diagonal elements by their re- > bb 1 y0 >
>
>ln þ b 2 ln b 3 >
b b >
ciprocals. Given that the diagonal entries are < bb 0 y0 =
not all identical means that the model parame- b
x 0 ¼ exp (2.8)
>
> b
b2 >
>
ters will be estimated with varying precision. >
: >
;
Ideally our coding scheme would have resulted
2.6 CONCENTRATIONeRESPONSE MODELING 49
Given values for the parameter estimates, use variety of shapes, ranging from a simple linear
of Eqs. (2.7) and (2.8) could be performed on a relationship to more complex curves displaying
calculator. However, the computation of the hormesis (a stimulatory effect at low concentra-
standard error of the resulting estimate has no tions) and hysteresis (where the trace of the
“closed-form” expression and more sophisti- response as a function of increasing concentra-
cated methods are required. Statistical resampling tion is not the reverse of the trace of the response
techniques such as the bootstrap or jack-knife are as a function of decreasing concentration). Two
often used for this purpose. These involve examples of idealized C-R curves are shown in
repeatedly sampling (with replacement) from Fig. 2.3.
the sample data. Parameter estimates are ob- Also indicated in Fig. 2.3 is the NEC and two
tained for each sample and the quantity of inter- estimates of an EC5 (ie, the concentration at
est is computed using Eq. (2.7) or Eq. (2.8). The which a 5% effect is expected. While hypotheti-
variation among the collection of estimates cal, Fig. 2.3 does highlight a real and commonly
thus obtained provides an estimate of the stan- occurring issue in C-R modeling, which is the
dard error for that quantity. Clearly, this is a sensitivity of the derived metric(s) to model
complicated and repetitive task that is ideally choice and parameterization.
suited to computer implementation. If the response in Fig. 2.3 represents the frac-
tion of unaffected organisms, then a plot of the
affected fraction assumes the shape shown in
2.6 CONCENTRATIONeRESPONSE Fig. 2.4. This elongated “S” shape arises not
MODELING only in C-R modeling but also in the context of
population growth models.
This section provides an overview of some Early investigations into population dy-
of the statistical issues associated with C-R namics by the British scholar, the Reverend
modeling as well as indicating opportunities Thomas Malthus (Malthus, 1978) used simple
for future development. It is not intended to exponential models to describe population
provide an exhaustive treatment of all aspects growth. However, these lacked realism because
of C-R methodologiesdfor example, the use of biological populations do not grow indefinitely.
toxicokineticetoxicodynamic (TKTD) models More realistic growth models displayed asymp-
to evaluate toxicity at the level of individual totic behavior; one of the most important of these
organisms. is the logistic function which was published in
In the present context, C-R experiments yield 1838 by Francois Verhulst (1838). The logistic
the data that are used to set “safe” exposure to equation remained a relatively obscure mathe-
concentrations of a toxicant, which is the subject matical result until it was rediscovered in 1920
of Section 2.7 on SSDs. We note that the terms by Raymond Pearl and Lowell Reed at John
“C-R” and “doseeresponse” tend to be used Hopkins University (Pearl and Reed, 1920). In
interchangeably, although strictly speaking these the 1930s prominent statisticians Chester Bliss
are not the same since dose is a function of concen- and Ronald Fisher took a slightly different
tration, frequency, and duration (of exposure). approach to bioassay modeling by treating the
C-R experiments and the mathematical and stimulus (dose) as the covariate and, because of
statistical modeling tools developed to analyze variability in individual tolerance levels, treated
the data they generate have a long and distin- the response as a random variable (Bliss, 1935). In
guished history which we briefly trace next. this formulation a probability model is associ-
A plot of an animal or organism’s response to ated with the response, which is often a normal
increasing levels of a toxicant can assume a wide distribution, but this is not a requirement. Bliss
50 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

FIGURE 2.3 Idealized C-R curves with threshold effect [blue (dark gray in print versions)] having EC5 ¼ 16.3; and without
[red (gray in print versions)] having EC5 ¼ 8.9. The no effect concentration (NEC) is 15.5.

1.0

0.8

0.6
response

0.4

0.2

0.0
0 20 40 60 80 100
concentration units

FIGURE 2.4 Typical elongated S-shape curve for fraction of affected organisms as a function of toxicant concentration.

was responsible for coining the term probit as explanatory variables. In 1944 US statistician
shorthand for probability unit (Bliss, 1935). Joseph Berkson advocated the use of the logistic
Probit analysis quickly established itself as the function as an alternative to probit analysis
de facto modeling approach to describe any rela- (Berkson, 1944) and introduced the term logit as
tion of a discrete binary outcome to one or more shorthand for the mathematical transformation
2.6 CONCENTRATIONeRESPONSE MODELING 51
of proportions using the logarithm of the odds change has been slow and hindered by pockets
ratio (the odds ratio is loosely defined as of inertia (Green et al., 2013; Green, 2016) even
 in the face of extensive advocacy (Landis and
probability ð“success”Þ
probability ð“failure”Þ
.
Chapman, 2011; Warne and van Dam, 2008;
As mentioned in Section 2.2, ANOVA tech- van Dam et al., 2012; Fox and Landis, 2016b).
niques have been used for many years in To highlight the fundamental difficulty,
conjunction with C-R experiments to derive the Fig. 2.5 shows the results of an experiment to
now widely discredited NOEC/NOEL toxicity investigate the toxicity of a herbicide to cucum-
metrics. While not wishing to revisit the long ber growth. When plotted on a log-scale using
list of objections here, it is nonetheless construc- the measured concentration data, the nature of
tive to briefly explore the critical flaw with these the relationship between response and concen-
measures in our quest for better designed C-R tration is revealed (Fig. 2.5A); however,
experiments. ANOVA only uses it as a label to distinguish
In the context of C-R experiments, hypothesis different dose groups (Fig. 2.5B). Furthermore,
tests should be used to assess the validity of while Fig. 2.5A indicates an “effect” in the region
statements about a toxic effect while estimation between 10e4 and 10e3, use of ANOVA methods
techniques should be used to derive a measure identifies dose groups 1e7 in Fig. 2.5B as not be-
of a toxic effect (Fox et al., 2012). Indeed, it is a ing statistically significant from the control
gross waste of experimental resources to derive group. Taking the “largest” of these results in
a toxicity measure using ANOVA methods since dose group 7 being the estimate of the effect
it makes no use whatsoever of the store of infor- level. Dose group 7 corresponds to an original
mation contained in the response-generating mech- concentration of 0.0055 units or 10e2.26 which is
anism (ie, the C-R model). Additionally, ANOVA well into the effect range of Fig. 2.5A.
requires replication at each concentration, If we suspect (or better still, know) that the re-
whereas this is not a requirement for C-R sponses for dose group 7 had been corrupted,
modeling. It has been argued that hypothesis they could be legitimately removed and the anal-
testing is preferable to modeling if the relation- ysis re-performed. Doing so results in the NOEC
ship between the effect and the toxicant concen- increasing to 0.0065 units, while a model-based
tration is unknown (Newman and Clements, analysis (not shown here) showed the estimate
2008) or poorly defined (Green, 2016). However, of the NEC remained relatively unaffected by
this is precisely the role of EDA techniques dis- the removal of the offending data and actually
cussed in Section 2.4. Ultimately, the problem decreased slightly. Finally, while a statement of
with ANOVA methods and the associated multi- precision or confidence for model-based esti-
ple comparison procedures used to derive mates can be readily obtained, such measures
toxicity measures is that concentration is treated are not only meaningless for the ANOVA-
as a factor not a covariate. This means that concen- based analysis but simply cannot be obtained
trations, which are carefully measured in a labo- (Fox et al., 2012).
ratory, are stripped of quantitative information Clearly, the use of models to generate
and are simply treated as labels for which esti- toxicity estimates provides for a richer and
mates of uncertainty and statements of precision more informative mode of inference. However
are inadmissible (Fox et al., 2012; Fox and the “cost” of these gains is usually an increase
Landis, 2016a). in computational complexity; although this
Although recommendations to move to may have posed problems in the past, the
model-based estimation procedures date back availability of sophisticated tools such as R
almost 20 years (for example, OECD, 1998), and special purpose software such as the drc
52 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

(A) 1.2

1.0

0.8
fraction
0.6

0.4

0.2

0.0
–6 –5 –4 –3 –2 –1 0
concentration units
(B)
1.5

1.0
fraction

0.5

0.0

–0.5
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
dose group

FIGURE 2.5 Cucumber shoot weight (as a fraction of control weight) versus: (A) herbicide concentration on log-scale (6
representing control group), solid line is smooth and (B) as a nominal dose group (0 representing control group), vertical lines
show 95% confidence intervals for mean response for each group. Adapted from Moore, D.R.J., Warren-Hicks, W.J., Qian, S.,
Fairbrother, A., Aldenberg, T., Barry, T., Luttik, R., Ratte, H.-T., 2010. Uncertainty analysis using classical and Bayesian hierarchical
models. In: Warren-Hicks, W.J., Hart, A. (Eds), Application of Uncertainty Analysis to Ecological Risks of Pesticides, CRC Press,
Boca Raton FL, USA.

package means that estimation and inference more prevalent; while the issue of how to
for complex C-R models can be undertaken formally accommodate subjective assessment
with relative ease. into a regulatory framework remains an open
As mentioned in Section 2.5, the use of question, there is little doubt that this mode of sta-
Bayesian methods in ecotoxicology is becoming tistical thinking and analysis will only continue to
2.6 CONCENTRATIONeRESPONSE MODELING 53
increase. By way of example, and following the A point of divergence between Frequentist
procedure in Fox (2010), Bayesian methods may and Bayesian methods of inference is that Fre-
be used to fit the exponential-threshold model quentists use point/interval estimates and/or
given by Eqs. (2.9a) and (2.9b) to the data appear- hypothesis testing, while Bayesians base all
ing in Fig. 2.5A. inference on the posterior density. Hypothesis
d
testing is a Frequentist concept that has no
Yi wgY ð,Þ (2.9a) Bayesian analogue. As an example, Fig. 2.6 sum-
E½Yi jxi  ¼ mi ¼ a exp½  bðxi  gÞIðxi  gÞ marizes the results of sampling from relevant
posterior distributions of a, b, g as well as that
(2.9b)
of the EC5. Conventional summary statistics
where Yi denotes the response at concentration xi such as means and standard deviations can be

1 z>0 obtained for these distributions; however, the
and IðzÞ ¼ . preferred Bayesian point estimate is either the
0 z0
E½Yi jxi  denotes the mathematical expecta- median or mode of the relevant posterior distri-
d
tion of Yi conditional on xi; the notation w is bution. Instead of a confidence interval, Bayes-
read as “is distributed as.” The parameters a, ians determine the interval of highest posterior
b, g have the following interpretations: a is the density or HPD interval. This is obtained by
mean response for concentrations between determining the shortest interval for which the
0 and g; b controls the rate of response for con- area under the posterior density is some nominal
centrations greater than g; and g is the NEC. valuedfor example 0.95. A 95% credibility inter-
Although seemingly complex, the Bayesian val for the parameter g is found to be [0, 0.0014]
analysis can be programmed with about 10 consistent with a subjective assessment based on
lines of code using freely available software the data in Fig. 2.5A. Note that the estimated
such as OpenBUGS (Openbugs, 2009), JAGS posterior probability of obtaining a value of g
(Sourceforge, 2015), or Stan (Stan Development at least as large as the previously determined
Team, 2015). NOEC for this data is only 0.003 suggesting an

FIGURE 2.6 Empirical posterior densities for parameters and EC5 for data in Fig. 2.5.
54 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

irreconcilable difference between the Bayesian each concentration. Not only this is unnecessary,
model and the Frequentist estimate. it is potentially wasteful of precious experi-
Another facet of C-R modeling that is attract- mental resources. To be clear, there is a simple
ing increasing attention is the design aspect. The dictum in statistics that states:
high cost of obtaining and analyzing data to
data ¼ model þ error
populate C-R models brings into focus the need
for an experimental design that yields the most The “error” term in the above equation is a
precise estimates of critical parameters that can “repository” for all variation in the data that
be obtained from a fixed sample size. A long- are not accounted for by the modeldwhich is
held view by practicing ecotoxicologists is that why it is often referred to as the “residual.”
a minimum of three replicates is necessary for There are many and varied sources of “error,”
each concentration used in a C-R experiment. but two important ones are stochastic error and
This convention is no doubt driven by the repli- lack-of-fit. Stochastic error is what some software
cation requirements of the one-way ANOVA packages refer to as “pure error”; it represents the
methods used to generate NOECs. Without unpredictable, randomly occurring variations
replication, the mean-square error term in the from “expectation” and, as such, can only be
ANOVA model cannot be estimated and thus described probabilistically. Lack-of-fit, on the
there is no basis for inferencedincluding the other hand, arises from model misspecification
determination of a NOEC. Even though there is and represents the discrepancies that arise
a growing trend to use modeling approaches, it because the chosen model is simply incapable
is not uncommon to see experimental designs of describing the response exactly. The situation
employing a minimum of three replicates at is shown in Fig. 2.7 which is a plot of leaf length

FIGURE 2.7 Relationship between total error, stochastic error, and lack-of-fit. Open circles are data and solid red circles (gray
in print versions) the mean response at each concentration. Vertical axis is measured leaf length (cm); horizontal axis is con-
centration of metsulfuron-methyl (mg/L).
2.6 CONCENTRATIONeRESPONSE MODELING 55
data provided with the R package drc. Thus, The D-optimality criterion was introduced in
provided there is replication, an independent es- Section 2.5 in the context of fractional factorial
timate of the pure error can be obtained and this designs, which are usually analyzed using
can be subtracted from what many packages ANOVA techniques. As was the case in the
report as mean square of residuals to obtain an example in Section 2.5, the optimality criterion
estimate of the lack-of-fit. was a function of the design matrix X only. This
Returning to the issue of design, there is no means that it is possible to compute the effi-
need to replicate unless the significance of the ciency of any given design by constructing Xd
lack-of-fit term is required to be tested and, something which can be done in advance of
even so, there are other “less costly” ways of any data collection. Unfortunately, this is not
assessing the adequacy of alternative model possible when designing a C-R experiment for
forms. For example, the drc package has a func- which the statistical model is nonlinear because
tion to compare the utility of a list of models us- evaluation of the optimality criterion requires
ing either the akaike information or AIC (a knowledge of the model parameters. Thus, we
measure of the utility of a model fitted to data) have a “circular” situation of wanting to design
criterion or the Bayesian information criterion an experiment to efficiently estimate model pa-
(similar to the AIC whose computation does rameters, but identification of the “best” design
not require specification of prior distributions requires us to first gather data to estimate the pa-
in a Bayesian analysis). rameters. Although a “best guess” for the pa-
As noted by the OECD (2012) the choice of the rameters could be used (for example, Chevre
number of dose levels and the dose spacing is crucial and Brazzale, 2008 used what they called “plau-
to achieving the objectives of the study (e.g., hazard sible values”), this is not entirely satisfactory as
identification or dose-response/risk assessment) and there is no way of assessing the quality of the
is important for subsequent statistical analysis. guess and hence of any design constructed based
With the requirement to replicate removed, on it.
experimental resources can now be re-directed More sophisticated solutions to this problem
to increasing the density of information exist. For example, Li and Fu (2013) couple
extracted around critical areas of the C-R Bayesian methods with an adaptive approach
curvedfor example, the NEC. While the design to the design problem. The procedure is complex
of the C-R experiment (ie, the choice of number and requires an understanding of advanced
of doses and dose spacing) can be done subjec- mathematical and statistical concepts. A simpli-
tively or based on heuristic rules, more formal fied outline of a modified version of their
procedures are available which identify the method is as follows:
“best” design by optimizing some criterion
1. Allocate a small fraction of the total sample
such as the D-optimality criterion introduced in
size to a “pilot experiment,” which uses a
Section 2.5.
subjective design (possibly informed by the
results of similar studies). For example, in a
2.6.1 Locally D-Optimal Designs
“conventional” ANOVA-type C-R
for C-R Experiments experiment assume the total sample size n is
Currently, the use of mathematical optimiza- given as n ¼ k$ r where k is the number of
tion methods to design C-R experiments is not concentrations to be used and r is the number
widespread, although this is expected to change of replicates at each concentration. The pilot
as ecotoxicologists move away from ANOVA- experiment could be based on a sample size
based to model-based estimation and inference. n1 ¼ k for example.
56 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

2. Using the results of the pilot experiment, important distinction is made between the popu-
obtain the empirical posterior distribution, lation and the sample. For statisticians, a popula-
pðQjdataÞ for the vector of parameters tion is simply the largest collection of “things”
of interest, Q (how this is done will depend on that (s)he is interested in; this may be biological,
the particular software package used). animate, or inanimate. For example, an elec-
Identify a list of candidate designs, {Di} tronics engineer might be interested in making
i ¼ 1,.,m where Di is an assignment of inference about the failure rate for a population
concentrations to the remaining sample of size of circuit boards, while a plant physiologist
n2 ¼ n  n1 ¼ k(r  1). might be interested in estimating the stomatal
3. For a given design, Di: response for all plants of a certain species to
a. sample from the posterior distribution elevated CO2 levels. In ecotoxicology the SSD al-
pðQjdataÞ to obtain a realization of the lows us to make inference about the toxic effect
predicted responses Y ¼ fY1 ; .; Yn2 g. of a substance on all species in a defined popula-
Repeat this L times; tion when only a small fraction of responses
b. For each of the samples in (a), evaluate the have been observed.
utility function gðYÞ (eg, gð$Þ might be the D- While the SSD has been an important de-
optimality criterion); and velopment in ecotoxicology, its use has also
c. Estimate E½gðYÞjDi  by taking the average been plagued with problems and controversy
of the L values obtained in (a). (Forbes and Forbes, 1993; Forbes and Calow,
4. The “best” design among the set of candidate 2002; Hickey and Craig, 2012; Wheeler et al.,
designs {Di} is that which maximizes the 2002a,b; Zajdlik, 2006, 2015). At least conceptu-
average in (c). ally, the borrowing of ideas from statistical
inference makes sense. If we take a random
sample of species from our defined population
2.7 SPECIES SENSITIVITY we can fit a theoretical probability distribution
DISTRIBUTION MODELING to the collection of toxicity measures obtained
from the sample. The advantage of working
This section presents a brief overview of the with this theoretical construct is that it (presum-
current status of SSD modeling. Discussion of ably) encapsulates our understanding of
the computational and inferential aspects of the toxicity based on a small sample and extends
SSD methodology is not covered here as these it to the entire populationdwhich is why the
have been well documented elsewheredfor use of the SSD in this manner is sometimes
example, Posthuma et al. (2002) and Duboudin referred to as an “extrapolation” technique.
et al. (2004). But there are some severe shortcomings which,
In a sense, the SSD is the inferential tool that despite more than 30 years of use and refine-
allows us to move beyond individual toxicity ment, remain problematic and potentially un-
metrics to saying something about the toxicity dermine the credibility of the whole approach.
of a chemical for an entire “population”; howev- Numerous papers have been written which
er, that population is defined (although gener- have documented the many shortcomings of
ally taken to mean all species comprising an the SSD methodology (and the companion issue
ecosystem). of flawed metrics such as NOECs) (Newman
The SSD is a statistical device and as such et al., 2000; Okkerman et al., 1991; Wang et al.,
there are important statistical considerations 2014). While some of these have obvious if not
that accompany its use and interpretation. As always practical solutions (for example,
with all inferential statistical techniques, an increasing the sample size to improve the
2.7 SPECIES SENSITIVITY DISTRIBUTION MODELING 57
quality of inference), other issues endure sample of test species represents the receiving
without resolution. Most significant among communities.” Given the importance of this
these are (1) unlike other areas of science (for key assumption, it is surprising that so little
example, physics, chemistry, and thermody- has been done to overcome its violation in eco-
namics) where there are theoretical justifica- toxicology. Indeed, various guideline documents
tions for why a particular functional form and regulatory requirements around the world
should apply, there is no such basis for SSDs; actually embed nonrandomness into SSDs by
and (2) despite being a core statistical assump- mandating the use of purposive sampling. For
tion and requirement, the selection of species example, the revised Australian and New
used to obtain the sample statistics referred to Zealand Water Quality Guidelines recommend
above is never random. using toxicity data from at least eight species
With respect to (1), current practice has from at least four taxonomic groups (Batley
enshrined the use of a small number of probabil- et al., 2014; Warne et al., 2014). Random sam-
ity models (notably the log-normal, logistic, and pling will likely remain an unattainable ideal in
log-logistic distributions) as suitable descriptors SSD modeling for at least two reasons: (i) it is
of the SSD. However, there is no guiding theory impossible to identify all species in an
in ecotoxicology to justify one distributional ecosystem; and (ii) testing protocols only exist
form over another. While for many applications for a handful of (non-randomly selected) species.
this would not be a problem because, overall, the Interestingly, although this has been a long-
fits afforded by all choices are reasonable, in eco- standing and widely acknowledged problem, it
toxicology this “degree of freedom” is particu- is only recently that any serious attempt has
larly problematic. Standard goodness-of-fit been made to both quantify and ameliorate the
tests do not help in the elicitation of the “best” effects of nonrandom species selection in SSD
distribution as the invariably small sample size fitting and HCx estimation (Fox, 2015).
results in low-powered goodness-of-fit tests. Although more research is needed into this
This means that any plausible candidate model crucial topic, Fox (2015) has shown that under
is unlikely to be rejected and, for this reason, the very mild assumption that specifies a beta
the vexatious issue of choice of functional form probability density function for the species selec-
is unlikely to ever be resolved. The other reason tion function, the actual SSD is not the assumed
why the choice of functional form is so problem- SSD (unless of course the selection process is
atic in SSD modeling is because our ultimate in- truly random). In particular Fox (2015) notes
terest lies in the most ill-defined portion of the that the actual distribution, when an assumed
fitted curvednamely the extreme left tail log-logistic SSD having parameters (a, b) is
(assuming larger values have a greater adverse used to describe toxicity data (X) that have
outcome). The ultimate objective of fitting the been selected according to a beta distribution
SSD is to estimate either the fraction of species with parameters (a, b), is a modified F distribu-
adversely affected by a prescribed concentration tion having the following probability density
of toxicant or the concentration that is hazardous function (pdf):
to no more than x% of all species. This latter  
 b
quantity is identified as the HCx and is numeri- b b x b1 b x
gX ðx; a; b; a; bÞ ¼ dF ; 2a; 2b
cally equivalent to the xth percentile of the aa a a a
SSD, typically the HC5. (2.10)
On the issue of species selection, Posthuma
et al. (2002) note “one of the serious failings of where the notation dFð$ ; n1 ; n2 Þ denotes a stan-
SSDs is the assumption that a non-random dard F distribution having n1 and n2 degrees of
58 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

freedom. A requirement satisfied by this pdf is seen, a small, biased sample of toxicity data is
that when species selection is truly random (cor- used to estimate the parameters of the theoretical
responding to the special case of the beta distri- SSD. Whether regression or ANOVA-based,
bution with a ¼ b ¼ 1), the actual distribution of these metrics are invariably derived from C-R ex-
the sample data is the assumed log-logistic. The periments for which standard laboratory proto-
impact of nonrandom species selection can be cols have been developed. Part of that
profounddresulting in HCx estimation errors standardization includes the fixing of the dura-
of a factor of 20 or more (Fox, 2015). Fortu- tion over which to run the experimentdtypically
nately, the impact of selection bias on an esti- 24, 48, or 96 h. Thus, the toxicity measures
mated HCx can be ameliorated through the derived from these experiments are, in reality,
use of a bias correction factor or bcf. For the only relevant for one particular period of expo-
assumed log-logistic SSD with a beta selection sure. As noted by Fox and Billoir (2013), there
function with parameters (a, b), the bcf to apply have been few, if any, attempts to fully integrate
to an HCx estimated from the SSD fitted to the the temporal component of SSDs. Using only
(biased) data is: simple assumptions about the manner in which
time was introduced into the SSD model, Fox

1=
b x b and Billoir (2013) were able to quantify the im-
bcf ðxÞ ¼ (2.11) pacts of this extra dimension on important quan-
a xx 100  x
tities such as the HCx, which of course itself
where xx is the xth percentile from the standard F becomes a function of the duration of the C-R
distribution having 2a and 2b degrees of freedom experiment. More recently, Kon Kam King
(Fox, 2015). The only difficulty with Eq. (2.11) is et al. (2015) used a “toxicodynamic” (TD) model
that it requires knowledge of the true value of to add a temporal dimension to the SSD. The TD
the shape parameter b of the assumed log- model is a “lumped” version of the more familiar
logistic SSD. As a workaround Fox (2015) sug- TKTD models used to describe various biolog-
gests replacing b in Eq. (2.11) with its sample ical processes related to the uptake, processing,
estimate b b. By way of example, suppose the and toxic effects within an individual organism
fitted log-logistic SSD had parameters a b ¼ 5:46 (eg, Jager et al., 2011). In this formulation, the
and bb ¼ 1:76. Furthermore, the species selection number of individuals Nijk of species j surviving
function was known to preferentially select the until time tk after having lived until time tk1
more sensitive species, which was adequately when exposed to the ith concentration level is
described by the beta density with a ¼ 0.5 and described by a binomial distribution where the
b ¼ 2.0. Using either published tables of the F probability of survival between tk and tk1 is a
distribution having 1 and 4 degrees of freedom four-parameter function of time and exposure
or, more conveniently, EXCEL’s intrinsic concentration, one of which is the NEC. A
F.INV() function, it is readily determined that Bayesian hierarchical model was used to esti-
for x ¼ 5, x5 ¼ 0.004453 and thus bcf(5) ¼ 8.94. mate parameters and make inference about the
In other words, the HC5 estimated from the time-varying nature of the HC5. Using published
SSD fitted to the sample data needs to be data on the salinity tolerance of 217 macroinver-
increased by almost an order of magnitude to tebrates from the Murray Darling Basin region in
compensate for the biased selection process. Australia, Kon Kam King et al. (2015) demon-
Another area of active research, which aims to strated a strong HC5etime dependency and
improve the effectiveness and applicability of concluded that an HC5 using data from a 72-h
SSD modeling, is the explicit incorporation of C-R experiment was likely to overestimate the
time, which has always been the “missing HC5 obtained from a longer running experiment.
dimension” in the whole approach. As we have From an environmental perspective, this is a
2.8 STATISTICAL SOFTWARE TOOLS FOR ECOTOXICOLOGY 59
disturbing outcome given that the potentially guideline documents and textbooks have not
affected fraction of species having long-term caught up with these advances.
exposure to a salinity not exceeding a More contemporary approaches to the
conventionally derived HC5 is likely to be manipulation, presentation, visualization, and
much higher than the assumed 5%. statistical analyses of data take advantage of
purpose-built software tools. While stand-
2.8 STATISTICAL SOFTWARE alone, computer programs, designed to achieve
TOOLS FOR ECOTOXICOLOGY a single outcome such as the estimation of
toxicity values and/or protective concentra-
This section provides an overview of some tions, have been in widespread use for many
common software tools used in ecotoxicology. yearsdfor example, the ToxCalc program
It is not intended to be exhaustive and focuses (Tidepool Scientific Software, 2016), there
mainly on the analysis of C-R data and SSD has been a seismic shift in academia and
fitting. TKTD models were briefly mentioned in science more generally to the R statistical
the previous section and, while these methods computing environment which commenced
can be used to obtain toxicity estimates, their around 2012. There are two good reasons for
focus is to predict toxic effects over time by this: (1) the R system is open source and there-
modeling the relationship between an individ- fore completely free of charge; and (2) the range
ual’s exposure to a contaminant and internal of intrinsic and user-contributed packages is
concentrations in the body of the organism. unrivalled.
These models therefore need to explicitly Given that the number of user-contributed R
describe processes controlling uptake, elimina- packages is predicted to reach 10,000 in 2016
tion, internal distribution, and metabolism. (Environmetrics Australia, 2014b) and in keep-
Closely related to TKTD models are those devel- ing with a forward-looking assessment of trends
oped around dynamic energy budget (DEB) the- in ecotoxicology, the remainder of this section fo-
ory. The most common of these is DEBtox which cuses on ecotoxicological uses of R.
first appeared in 1996 as an accompaniment to
the book by Kooijman and Bedaux (1996). DEB 2.8.1 R Package webchem
theory “is all about mechanistically linking
webchem is an R package to retrieve chemical
(time-varying) external concentrations of a toxi-
cant to the effects on life-history traits such as information from the web (Sz€ocs, 2015b). It can
survival, growth and reproduction, over time” interrogate and retrieve information from a
(Debtox Information Site, 2011). It does this by number of internet sources including:
constructing rules that govern how organisms • National Cancer Institute’s Chemical
partition energy to control all life stages. Identifier Resolver;
The information and communications tech- • Royal Society of Chemistry’s ChemSpider;
nology “revolution” and the aforementioned • National Center for Biotechnology
rise of “big data” have led to rapid and unparal- Information’s PubChem BioAssay Database;
leled developments in computing hardware and • University of California’s Chemical
software. In ecotoxicology, this has resulted in Translation Service;
the deployment of more sophisticated models • Pesticide Action Network’s Pesticide
whose parameters can be estimated using Database;
computationally intensive techniques such as • Effects information from aquatic and
Markov Chain Monte Carlo (MCMC) and other terrestrial ecotoxicology from the Federal
resampling strategies (Gamerman, 2006; Gilks Environment Agency of Germany’s ETOX
et al., 1998). To some extent, ecotoxicological database;
60 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

• University of Hertfordshire’s pesticide 2.8.3 R Package drc


chemical identity, physicochemical, human
health, and ecotoxicological database; and The package drc provides a comprehensive
• United States National Library of Medicine’s suite of modeling and analyses tools for fitting
TOXNETda group of databases covering complex models to data generated by C-R exper-
chemicals and drugs, diseases and the iments. An excellent overview describing the
environment, environmental health, features and use of drc can be found in Ritz
occupational safety and health, poisoning, and Streibig (2005). As an example, the R code
risk assessment and regulations, and below fits a four-parameter and a five-
toxicology. parameter log-logistic function to C-R data. Us-
ing the anova() function with the fitted models
For example, entering the command get_cid as arguments provides a quick check on the
(“Triclosan”) produces the following list of value of the extra parameter. In this case, the
chemical IDs for triclosan:

#> [1] "5564" "131203" "627458" "15942656" "16220126" "16220128"


#> [7] "16220129" "16220130" "18413505" "22947105" "23656593" "24848164"
#> [13] "25023954" "25023955" "25023956" "25023957" "25023958" "25023959"
#> [19] "25023960" "25023961" "25023962" "25023963" "25023964" "25023965"
#> [25] "25023966" "25023967" "25023968" "25023969" "25023970" "25023971"
#> [31] "25023972" "25023973" "45040608" "45040609" "67606151" "71752714"
#> [37] "92024355" "92043149" "92043150"

while the following examples illustrate how to


convert between IDs:

cts_convert(query ¼ ’3380-34-5’, from ¼ ’CAS’, to ¼ ’PubChem CID’)


#> [1] "5564" "34140"
cts_convert(query ¼ ’3380-34-5’, from ¼ ’CAS’, to ¼ ’ChemSpider’)
#> [1] "31465"
(inchk <- cts_convert(query ¼ ’50-00-0’, from ¼ ’CAS’, to ¼ ’inchikey’))
#> [1] "WSFSSNUMVMOOMR-UHFFFAOYSA-N"

2.8.2 R Package ggplot2


reported p-value of 0.3742 suggests no significant
ggplot2 is a feature-rich, general purpose improvement in overall fit using the 5-parameter
graphics package and, although not specifically model. Plotting the data and fitted models is
developed for ecotoxicologists, it can be used simple and elegant as shown by the brevity of
to great effect to produce highly customizable code required and the resulting output
SSDs (Figs. 2.8 and 2.9). (Fig. 2.10).
2.8 STATISTICAL SOFTWARE TOOLS FOR ECOTOXICOLOGY 61

> head(df) # inspect first 6 rows of dataframe df

concen response
1 0 0.8457565
2 0 0.9741697
3 0 0.9874539
4 0 1.0228782
5 0 1.0405904
6 0 1.0583026

> fit1<-drm(df$response w df$concen,fct¼LL.4()) # fit 4-parameter log-logistic


> fit2<-drm(df$response w df$concen,fct¼LL.5()) # fit 5-parameter log-logistic
>
> anova(fit1,fit2) # compare models
1st model
fct: LL.4()
2nd model
fct: LL.5()

ANOVA table

ModelDf RSS Df F value p value


1st model 59 1.2922
2nd model 58 1.2746 1 0.8019 0.3742

> # plot results


>
> plot(fit1,broken¼TRUE,legend¼TRUE,legendText¼"4-param log-logistic",
þ col¼"red",type¼"confidence",ylab¼"fraction",xlab¼"concen")
>
> plot(fit2,add¼TRUE,legend¼TRUE,legendText¼"5-param log-
logistic",legendPos¼c(0.19,1.2),
þ col¼"blue",type¼"confidence",ylab¼"fraction",xlab¼"concen")
>
> plot(fit2,add¼TRUE,type¼"obs",pch¼16,cex¼0.6)

2.8.3 R Function fitdistr() and Package capable of fitting a variety of univariate proba-
fitdistrplus bility models to both censored and noncensored
data using a variety of algorithms including
These are general-purpose distribution-fitting
maximum likelihood estimation, method of mo-
programs. fitdistr() is part of the MASS package
ments, quantile matching, and maximum good-
and fits univariate probability distributions
ness of fit. fitdistrplus is at the heart of the
using maximum-likelihood estimation. The
online SSD tool MOSAIC (see below).
fitdistrplus package has more features and is
62 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

Parameter estimates for a lognormal distribu-


tion fitted to the data of Fig. 2.8 are readily ob-
tained using either fitdistr() or fitdistrplus:

> fitdistr(data$EC50,"lognormal")
meanlog sdlog
1.9084029 1.8440993
(0.4024155) (0.2845507)

> fit<-fitdist(data$EC50, "lnorm",method¼"mme")


> fit
Fitting of the distribution ’ lnorm ’ by matching moments
Parameters:
estimate
meanlog 3.825651
sdlog 1.716488

FIGURE 2.8 Empirical SSD for endosulfan EC50 data for non-Australian fish. From Hose, G.C., Van den Brink, P.J., 2004.
Confirming the species-sensitivity distribution concept for endosulfan using laboratory, mesocosm, and field data. Environ. Contam. Tox-
icol. 47, 511e520.
2.8 STATISTICAL SOFTWARE TOOLS FOR ECOTOXICOLOGY 63

FIGURE 2.9 Theoretical lognormal SSD [red curve (dark gray in print versions)] for endosulfan data of Fig. 2.8 (solid black
circles) together with lognormal distributions fitted to bootstrap samples [blue curves (gray in print versions)] and 95%
confidence limits (black dashed curves). R source code adapted from http://bit.ly/1OrFC5n.

FIGURE 2.10 Plot of data and fitted models using drc package. Colored bands are 95% confidence intervals for mean
response.
64 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

2.8.4 MOSAIC 2.8.5 BurrliOZ


MOSAIC is an online tool that interfaces with BurrliOZ (CSIRO, 2015) is similar in concept to
R to fit one of a number of predefined probability MOSAIC in that it is an interface to R and is used
models to empirical toxicity data. It also uses to fit either a log-logistic or Burr-type distribution
bootstrapping to provide estimates of the stan- to toxicity data. However, unlike MOSAIC, Burr-
dard error of parameter estimates as well as for liOz is downloaded and run locally on a desktop
derived HCx values. Developed by researchers computer. It was originally released as a stand-
at the Biometry and Evolutionary Biology Labo- alone program developed by statisticians at Aus-
ratory at the University of Lyon, the tool can be tralia’s Commonwealth Scientific and Industrial
accessed online (Biometry and Evolutionary Research Organisation (CSIRO) to support the
Biology Laboratory, 2015). A screenshot of the 2000 revision of the Australian and New Zealand
output for the fish data of Fig. 2.8 is shown in Guidelines for fresh and marine water quality
Fig. 2.11. (ANZECC/ARMCANZ, 2000a). While offering

FIGURE 2.11 Screenshot of MOSAIC output for lognormal SSD fitted to data of Fig. 2.8.
2.9 FUTURE POSSIBILITIES 65
similar capabilities to MOSAIC, BurrliOz is more such as MOSAIC (Biometry and Evolutionary
cumbersome and, although it also uses R to Biology Laboratory, 2015) and BurrliOz
perform the necessary calculations, unlike (CSIRO, 2015). The increasing number of
MOSAIC the R source code cannot be accessed. customized ecotoxicology packages written in
A screenshot from BurrliOz applied to the data R coupled with the development of “wrapper”
of Fig. 2.8 is shown in Fig. 2.12. programs and languages such as R-Studio’s
Shiny software (RStudio, 2016) will improve
user-friendliness thereby encouraging greater
2.9 FUTURE POSSIBILITIES
uptake and use of these tools. This convergence
of technologies will, we predict, result in the
It is true to say that advances in statistical
development of ecotoxicological “apps” for mo-
methodology over the past 40 years have to a
bile computing devices such as tablets, smart
large extent been driven by advances in
phones, and iPads. So, at one level, a new fron-
computing power and software developments.
tier in statistical ecotoxicology will be entirely
Witness, for example the explosive growth of
digital.
R, the pervasiveness of computationally inten-
In terms of theoretical challenges and oppor-
sive techniques such as MCMC, and the ability
tunities, those identified by Eggen et al. (2004)
to produce publication-quality graphics with
remain relevant and are associated with:
the click of a mouse. It therefore seems reason-
able to suggest that opportunities and develop- • Low concentrations of pollutants and chronic
ments in statistical ecotoxicology will exhibit a effects;
similar dependency. Already, stand-alone • Multiple effects by single pollutants;
desktop applications for undertaking many • Complex mixtures of pollutants;
of the analyses presented in this chapter • Multiple stressors; and
are gradually being replaced by online tools • Ecosystem complexity.

FIGURE 2.12 Screenshot of BurrliOz HCx calculator screen for log-logistic SSD fitted to data of Fig. 2.8.
66 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

Statistical tools such as maximum likelihood process, imprecise model specification, and
estimation, probit analysis, and logistic regres- statistical treatment of data;
sion and to a lesser extent generalized linear • Refinement of modeling capabilities to
models and Bayesian methods now underpin undertake external and internal exposure
contemporary methods in ecotoxicological esti- assessments;
mation and inference. The challenge as we see • Elicitation of “expert” opinion in the setting of
it, is to more tightly integrate this important Bayesian priors and protocols for reaching
facet of ecotoxicology into university courses consensus when these are used in a regulatory
and professional training programs. Much of context.
the statistical framework in ecotoxicology has
Because of its importance and widespread
been the result of organic growth rather than
use, the SSD methodology and companion tech-
planned development. Regrettably, biometri-
niques have been a focus of this chapter. Most of
cians are a vanishing species in natural resource
the dot-points above relate to ways of improving
management agencies around the world. An
both the theoretical and applied aspects of the
unintended consequence of this decline is a ten-
approach. It is likely that the SSD will retain its
dency of researchers in ecotoxicology to avoid
place as the preferred (and often mandated)
new and potentially challenging statistical con-
method for establishing protective concentra-
cepts thereby increasing the propensity to main-
tions although one group of experts recently
tain the status quo in terms of the statistical
identified a need to “move away from using
analysis of ecotoxicological data. It is, we sus-
SSDs as a purely statistical construct applied to
pect, also partly responsible for on-going de-
poorly understood species sensitivity data to
bates that have outlived their usefulness about
one in which SSDs provide the framework for
indeterminate questions of the sort is the log-
a more process-based approach” (ECETOC,
normal distribution better than the log-logistic dis-
2014).
tribution as an SSD? or what value of x should be
In its defense the SSD has provided a consis-
used to reconcile ECx and NOECs? More produc-
tent and reproducible way of establishing water
tive lines of investigation will be associated
quality criteria and while not totally free of sub-
with:
jective choices concerning choice of distribution,
• Addressing and compensating for small, parameterization, and estimation techniques, it
biased samples in SSD and HCx has removed the arbitrariness of the assessment
estimation; factor method it replaced. Problems that have
• Development of statistical methods to been identified with this “statistical construct”
validate predictions from TKTD models and are not the result of flawed statistical science,
SSD-based approaches; but invariably arise from violations of critical
• How to design a C-R experiment that assumptions and/or inappropriate application.
maximizes information content for minimum How the research community deals with these
cost; issues remains to be seen.
• How, or establish if it is possible, to set an HCx
for a mixture of chemicals;
• How to seamlessly integrate the temporal
References
dimension into SSD modeling and HCx Adelmann, S., 1961. Irregular fractions of the 2n factorial
estimation rather than marginalizing it; experiments. Technometrics 3, 479e496.
ANZECC/ARMCANZ, 2000a. Australian and New Zealand
• Strategies for error propagation to incorporate
Guidelines for Fresh and Marine Water Quality. Austra-
uncertainties arising from data collection lian and New Zealand Environment and Conservation
REFERENCES 67
Council/Agricultural and Resource Management Council CRAN, 2015. Design of Experiments (DoE) and Analysis of
of Australia and New Zealand, Canberra, ACT, Australia. Experimental Data. Available at: https://cran.r-project.
ANZECC/ARMCANZ, 2000b. Australian Guidelines for org/web/views/ExperimentalDesign.html.
Water Quality Monitoring and Reporting. Australian Crane, M., Newman, M.C., 2000. What level of effect is a no
and New Zealand Environment and Conservation Coun- observed effect? Environ. Toxicol. Chem. 19, 516e519.
cil, Agriculture and Resource Management Council of CSIRO, 2015. BurrliOz 2.0. Available at: https://research.
Australia and New Zealand (National Water Quality csiro.au/software/burrlioz/.
Management Strategy number 7). Dasu, T., Johnson, T., 2003. Exploratory Data Mining and Data
Barnett, V., O’Hagan, A., 1997. Setting Environmental Stan- Cleaning. John Wiley & Sons. http://ca.wiley.com/
dards: The Statistical Approach to Handling Uncertainty WileyCDA/WileyTitle/productCd-0471268518,subjectCd-
and Variation. Chapman and Hall, London, UK. CSB0.html.
Batley, G.E., Braga, O., Van Dam, R., Warne, M.S.J., Debtox Information Site, 2011. Debtox Information: Making
Chapman, J.C., Fox, D.R., Hickey, C., Stauber, J.L., 2014. Sense of Ecotoxicity Test Results. Available at: http://
Technical Rationale for Changes to the Method for www.debtox.info/about_debtox.html.
Deriving Australian and New Zealand Water Quality Dey, A., 1985. Orthogonal Fractional Factorial Designs. John
Guideline Values for Toxicants. Council of Australian Wiley and Sons, Haslted Press, New York.
Government’s Standing Council on Environment and Duboudin, C., Ciffroy, P., Magaud, H., 2004. Effects of data
Water, Sydney, Australia. manipulation and statistical methods on species sensi-
Berkson, J., 1944. Application of the logistic function to bio- tivity distributions. Environ. Toxicol. Chem. 23, 489e499.
assay. J. Am. Stat. Assoc. 39, 357e365. ECETOC, 2014. Estimating Toxicity Thresholds for Aquatic
Bickel, P.J., Doksum, K.A., 2016. Mathematical Statistics: Ecological Communities From Sensitivity Distributions
Basic Ideas and Selected Topics. CRC Press, Boca Raton, 11e13 February, 2014. Amsterdam Workshop Report
FL, USA. No. 28. European Centre for Ecotoxicology and Toxi-
Billoir, E., Delignette-Muller, M.L., Pery, A.R.R., Charles, S., cology of Chemicals.
2008. A Bayesian approach to analyzing ecotoxicological Eggen, R.I.L., Behra, R., Burkhardt-Holm, P., Esche, B.I.,
data. Environ. Sci. Technol. 42, 8978e8984. Schweigert, N., 2004. Challenges in ecotoxicology. Envi-
Biometry and Evolutionary Biology Laboratory, 2015. ron. Sci Technol. 38, 58Ae64A.
MOSAIC: Modelling and Statistical Tools for Environment Canada, 2005. Guidance Document on Statisti-
Ecotoxicology. University of Lyon, France. Available at: cal Methods for Environmental Toxicity Tests. EPS
http://pbil.univ-lyon1.fr/software/mosaic/ssd/. 1/RM/46, Ottawa, ON, Canada.
Bliss, C.I., 1935. The calculation of the dosage-mortality Environmetrics Australia, 2014a. Big Data Is Watching You.
curve. Ann. Appl. Biol. 22, 134e167. Available at: http://www.environmetrics.net.au/index.
Bose, R.C., Bush, K.A., 1952. Orthogonal arrays of strength php?news&nid¼81.
two and three. Ann. Math. Stat. 23, 508e524. Environmetrics Australia, 2014b. The Explosive Growth of R.
Cash, J.T., 1908. The relationship of action to dose especially Available at: http://environmetrics.net.au/index.php?
with reference to repeated administration of indaconitine. news&nid¼79.
Br. Med. J. 1908, 1213e1218. European Commission, 2011. Technical guidance for
CCME, 2007. A Protocol for the Derivation of Water quality deriving environmental quality standards. In: Guidance
Guidelines for the Protection of Aquatic Life. Canadian Document No. 27, Common Implementation Strategy
Council of Ministers of the Environment, Ottawa, ON, for the Water Framework Directive, Brussels, Belgium,
Canada, 37 pp. 204 pp.
Chapman, P.M., Cardwell, R.S., Chapman, P.F., 1996. Evans, D.A., Newman, M.C., Lavine, M., Jaworska, J.S.,
A warning: NOECs are inappropriate for regulatory Toll, J., Brooks, B.W., Brock, T.C.M., 2010. The Bayesian
use. Environ. Toxicol. Chem. 15, 77e79. vantage for dealing with uncertainty. In: Warren-
Chevre, N., Brazzale, A.R., 2008. Cost-effective experimental Hicks, J., Hart, A. (Eds.), Application of Uncertainty Anal-
design to support modelling of concentration-response ysis to Ecological Risks of Pesticides. CRC Press, Boca
functions. Chemosphere 72, 803e810. Raton, FL, USA.
Cliffroy, P., Keller, M., Pasanisi, A., 2013. Estimating hazard- Forbes, T.L., Forbes, V.E., 1993. A critique of the use of
ous concentrations by an informative Bayesian approach. distribution-based extrapolation models in
Environ. Toxicol. Chem. 32, 602e611. ecotoxicology. Funct. Ecol. 7, 249e254.
Cox, D.R., Snell, E.J., 2000. Statistics: Principles and Forbes, V.E., Calow, P., 2002. Species sensitivity distribu-
Examples. Chapman and Hall/CRC, Boca Raton, FL, tions: a critical appraisal. Hum. Ecol. Risk Assess. 8,
USA. 473e492.
68 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

Fox, D.R., 2006. Statistical issues in ecological risk Hickey, G.L., Craig, P.S., 2012. Competing statistical methods
assessment. Hum. Ecol. Risk Assess. 12, 120e129. for the fitting of normal species sensitivity distributions:
Fox, D.R., 2008. NECs, NOECs, and the ECx. Australas. J. recommendations for practitioners. Risk Anal. 32,
Ecotoxicol. 14, 7e9. 1232e1243.
Fox, D.R., 2010. A Bayesian approach for determining the no Hinkelmann, K., Kempthorne, O., 2008. Design and Analysis
effect concentration and hazardous concentration in of Experiments. In: Introduction to Experimental Design,
ecotoxicology. Ecotoxicol. Environ. Saf. 73, 123e131. vol. 1. Wiley, NJ, USA.
Fox, D.R., 2015. Selection bias correction for species Hoekstra, J.A., Van Ewijk, P.H., 1993. The bounded effect
sensitivity distribution modelling and hazardous con- concentration as an alternative to the NOEC. Sci. Tot. En-
centration estimation. Environ. Toxicol. Chem. 34, viron. 134, 705e711.
2555e2563. Hose, G.C., Van den Brink, P.J., 2004. Confirming the species-
Fox, D.R., Landis, W.G., April 18, 2016a. Don’t be fooled - A sensitivity distribution concept for endosulfan using
NOEC is no substitute for a poor concentration-response laboratory, mesocosm, and field data. Environ. Contam.
experiment. Environ. Toxicol. Chem. http://dx.doi.org/ Toxicol. 47, 511e520.
10.1002/etc.3459. Jager, T., 2012. Bad habits die hard: the NOECs persistence re-
Fox, D.R., Landis, W.G., 2016b. Comment on ET&C perspec- flects poorly on ecotoxicology. Environ. Toxicol. Chem.
tives November 2015 - A Holistic View. Environ. Toxicol. 31, 228e229.
Chem. 35, 1337e1339. Jager, T., Albert, C., Preuss, T.G., Ashauer, R., 2011. General
Fox, D.R., Billoir, E., 2013. Time dependent species sensitivity unified threshold model of survival e a toxicokinetic-
distributions. Environ. Toxicol. Chem. 32, 378e383. toxicodynamic framework for ecotoxicology. Environ.
Fox, D.R., Landis, W.G., April 18, 2016a. Don’t be fooled e Sci. Technol. 45, 2529e2540.
A NOEC is no substitute for a poor concentration- Jaworska, J., Gabbert, S., Aldenberg, T., 2010. Towards opti-
response experiment. Environ. Toxicol. Chem. http:// mization of chemical testing under REACH: a Bayesian
dx.doi.org/10.1002/etc.3459. network approach to integrated testing strategies. Regul.
Fox, D.R., Landis, W.G., 2016b. Comment on ET&C per- Toxicol. Pharmacol. 57, 157e167.
spectives November 2015 e A Holistic View. Environ. Kempthorne, O., 1947. A simple approach to confounding
Toxicol. Chem. 35, 1337e1339. and fractional replication in factorial experiments. Bio-
Fox, D.R., Billoir, E., Charles, S., Delignette-Muller, M.L., metrika 34, 255e272.
Lopes, C., 2012. What to do with NOECs/NOELs e Kon Kam King, G., Delignette-Muller, M.L., Kefford, B.J.,
prohibition or innovation? Integr. Environ. Assess. Piscart, C., Charles, S., 2015. Constructing time-resolved
Manag. 8, 764e766. species sensitivity distributions using hierarchical toxico-
Gad, S.C., 2006. Statistics and Experimental Design for Toxi- dynamic model. Environ. Sci. Technol. 49, 12465e12473.
cologists and Pharmacologists, 4th ed. Taylor and Francis, Kooijman, S., Bedaux, J.J.M., 1996. The Analysis of Aquatic
Boca Raton, FL, USA. Toxicity Data. VU University Press, Amsterdam,
Gamerman, D., 2006. Markov Chain Monte Carlo: Stochastic Netherlands.
Simulation for Bayesian Inference. Chapman and Hall/ Kutner, M.H., Nachtsheim, C.J., Neter, J., Li, W., 2016.
CRC, Boca Raton, FL, USA. Applied Linear Statistical Models, 5th ed. Irwin/
Gilks, W.R., Richardson, S., Spiegelhalter, D.J., 1998. Markov McGraw-Hill, New York, USA.
Chain Monte Carlo in Practice. Chapman and Hall/CRC, Landis, W.G., Chapman, P.M., 2011. Well past time to stop
Boca Raton, FL, USA. using NOELs and LOELs. Integr. Environ. Assess.
Graybill, F.A., 1976. Theory and Application of the Linear Manag. 7, vieviii.
Model. Duxbury Press, North Scituate. Lawson, J., 2015. Design and Analysis of Experiments with R.
Green, J.W., 2016. Issues with using only regression models CRC Press, Boca Raton, FL, USA.
for ecotoxicology studies. Integr. Environ. Assess. Manag. Li, J., Fu, H., 2013. Bayesian adaptive D-optimal design with
12, 198e199. delayed responses. J. Biopharm. Stat. 23, 559e568.
Green, J.W., Springer, T.A., Staveley, J.P., 2013. The drive to Link, W.A., Albers, P.H., 2007. Bayesian multimodel infer-
ban the NOEC/LOEC in favor of ECx is misguided and ence for dose-response studies. Environ. Toxicol. Chem.
misinformed. Integr. Environ. Assess. Manag. 9, 12e16. 26, 1867e1872.
Grist, E.P.M., O’Hagan, A., Crane, M., Sorokin, N., Sims, I., Malthus, T.R., 1798. An Essay on the Principle of Population.
Whitehouse, P., 2006. Comparison of frequentist and J. Johnson, London, UK.
Bayesian freshwater species sensitivity distributions for McBride, M.F., Garnett, S.T., Szabo, J.K., Burbidge, A.H.,
chlorpyrifos using time-to-event analysis and expert Butchart, S.H.M., Christidis, L., Dutson, G., Ford, H.A.,
elicitation. Environ. Sci. Technol. 40, 295e301. Loyn, R.H., Watson, D.M., Burgman, M.A., 2012.
REFERENCES 69
Structured elicitation of expert judgments for threatened Posthuma, L., Suter, G.W., Traas, T.P., 2002. Species Sensi-
species assessment: a case study on a continental scale us- tivity Distributions in Ecotoxicology. Lewis Publishers,
ing email. Methods Ecol. Evol. 3, 906e920. Boca Raton, FL, USA.
Moore, D.R.J., Warren-Hicks, W.J., Qian, S., Fairbrother, A., Quora, 2014. What Is Data Munging? Available at: https://
Aldenberg, T., Barry, T., Luttik, R., Ratte, H.-T., 2010. Un- www.quora.com/What-is-data-munging.
certainty analysis using classical and Bayesian hierarchi- R Development Core Team, 2004. R: A Language and Envi-
cal models. In: Warren-Hicks, W.J., Hart, A. (Eds.), ronment for Statistical Computing. R Foundation for Sta-
Application of Uncertainty Analysis to Ecological Risks tistical Computing, Vienna, Austria. ISBN: 3-900051-00-3,
of Pesticides. CRC Press, Boca Raton FL, USA. URL http://www.R-project.org.
Newman, M.C., 2012. Quantitative Ecotoxicology. CRC Rao, C.R., 1950. The theory of fractional replication in facto-
Press, Boca Raton, FL, USA. rial experiments. Sankhya 10, 81e86.
Newman, M.C., Clements, W.H., 2008. Ecotoxicology: A Ritz, C., Streibig, J.C., 2005. Bioassay analysis using R. J. Stat.
Comprehensive Treatment. CRC Press, Boca Raton, Softw. 12, 1e22.
FL, USA. RStudio, 2016. Shiny Software. Available at: https://www.
Newman, M.C., Ownby, D.R., Mezin, L.C.A., Powell, D.C., rstudio.com/products/shiny/.
Christensen, T.R.L., Leerberg, S.B., Anderson, B.-A., Shieh, J.N., Chao, M.R., Chen, C.Y., 2001. Statistical compar-
2000. Applying species sensitivity distributions in ecolog- isons of the no-observed-effect concentration and the
ical risk assessment: assumptions of distribution type and effective concentration at 10% inhibition (EC10) in algal
sufficient numbers of species. Environ. Toxicol. Chem. 19, toxicity tests. Water Sci. Technol. 43, 141e146.
508e515. Sourceforge, 2015. JAGS: Just Another Gibbs Sampler. Avail-
New York Times, 2014. For Big-data Scientists, ‘Janitor able at: http://mcmc-jags.sourceforge.net/.
Work’ is Key Hurdle to Insights. Available at: http:// Sparks, T., 2000. Statistics in Ecotoxicology. Wiley and Sons,
www.nytimes.com/2014/08/18/technology/for-big-data- Chichester, UK.
scientists-hurdle-to-insights-is-janitor-work.html?_r¼2. Stan Development Team, 2015. Stan Modeling Language
OECD (Organization for European Cooperation and Devel- Users Guide and Reference Manual, Version 2.8.0. Avail-
opment), 1998. Report of the OECD workshop on statisti- able at: http://mc-stan.org/documentation/.
cal analysis of aquatic toxicity data. In: OECD Sz€
ocs, E., 2015a. Species Sensitivity Distributions (SSD) with
Environmental Health and Safety Publications Series on R. Available at: http://edild.github.io/ssd/.
Testing and Assessment No. 10, Paris, France. Sz€
ocs, E., 2015b. Introducing the Webchem Package. Avail-
OECD, 2012. Guidance document 116 on the conduct and able at: http://edild.github.io/webchem/.
design of chronic toxicity and carcinogenicity studies, Tidepool Scientific Software, 2016. ToxCalc Software. Avail-
supporting test guidelines 451, 452, and 453. In: Series able at: https://tidepool-scientific.com/ToxCalc/Tox-
on Testing and Assessment No. 116, Paris, France, Calc.html.
2nd ed. Trevan, J.W., 1927. The error of determination of toxicity.
OECD, 2014. Current approaches in the statistical analysis Proc. R. Soc. B 101 (712), 483e514.
of ecotoxicity data: a guidance to application (annexes USEPA (U.S. Environmental Protection Agency), 2002.
to this publication exist as a separate document). In: Guidance on Choosing a Sampling Design for Environ-
OECD Series on Testing and Assessment, No. 54, Paris, mental Data Collection for Use in Developing a Quality
France. Assurance Project Plan. EPA QA/G-5S, Washington,
Okkerman, P.C., Plassche, E.J.V.D., Slooff, W., Van DC, USA.
Leeuwen, C.J., Canton, J.H., 1991. Ecotoxicological effects USEPA, 2006. Data Quality Assessment: Statistical Methods
assessment: a comparison of several extrapolation for Practitioners. EPA QA/G-9S, Washington, DC, USA.
procedures. Ecotoxicol. Environ. Saf. 21, 182e193. USEPA, 2015. ECOTOX User Guide: ECOTOXicology Data-
Openbugs, 2009. Bayesian Inference Using Gibbs Sampling. base System. Version 4.0. Available at: http://www.
Available at: http://www.openbugs.net/w/FrontPage. epa.gov/ecotox/.
Payet, J., 2004. Assessing Toxic Impacts on Aquatic Ecosys- van Dam, R.A., Harford, A.J., Warne, M.S., 2012. Time to get

tems in Life Cycle Assessment (LCA) (Ph.D. thesis). Ecole off the fence: the need for definitive international guid-
Polytechnique Federale de Lausanne, Lausanne, ance on statistical analysis of ecotoxicity data. Integr. En-
Switzerland. viron. Assess. Manag. 8, 242e245.
Pearl, R., Reed, L.J., 1920. On the rate of growth of the pop- van der Hoeven, N., Noppert, F., Leopold, A., 1997. How to
ulation of the United States since 1790 and its mathemat- measure no effect. Part I: towards a new measure of
ical representation. Proc. Natl. Acad. Sci. USA 6, chronic toxicity in ecotoxicology. Introduction and work-
275e288. shop results. Environmetrics 8, 241e248.
70 2. CONTEMPORARY METHODS FOR STATISTICAL DESIGN AND ANALYSIS

Verhulst, P.F., 1838. Notice sur la loi que la population pour- Wheeler, J.R., Leung, K.M.Y., Morritt, D., Whitehouse, P.,
suit dans son accroissement. Corresp. Math. Phys. 10, Sorokin, N., Toy, R., Holt, M., Crane, M., 2002b. Fresh-
113e121. water to saltwater toxicity extrapolation using species
Wang, Y., Zhang, L., Meng, F., Zhou, Y., Jin, X., Giesy, J.P., sensitivity distributions. Environ. Toxicol. Chem. 21,
Liu, F., 2014. Improvement on species sensitivity distribu- 2459e2467.
tion methods for deriving site-specific water quality Wickham, H., 2009. ggplot2: Elegant Graphics for Data
criteria. Environ. Sci. Pollut. Res. 22, 5271e5282. Analysis. Springer, New York, NY, USA.
Warne, MStJ., van Dam, R., 2008. NOEC and LOEC data Wickham, H., 2014. Tidy data. J. Stat. Softw. 59, 1e23.
should no longer be generated or used. Australas. J. Eco- Wikipedia, 2015. Big Data. Available at: https://en.
tox. 14, 1e5. wikipedia.org/wiki/Big_data.
Warne, MStJ., Batley, G.E., Braga, O., Chapman, J.C., Wikipedia, 2016. Determinant. Available at: https://en.
Fox, D.R., Hickey, C.W., Stauber, J.L., van Dam, R.A., wikipedia.org/wiki/Determinant.
2013. Revisions to the derivation of the Australian and Zajdlik, B.A., 2006. Potential statistical models for describing
New Zealand guidelines for toxicants in fresh and marine species sensitivity distributions. In: Prepared for Cana-
waters. Environ. Sci. Pollut. Res. 21, 51e60. dian Council of Ministers for the Environment CCME
Warne, MStJ., Batley, G.E., Van Dam, R.A., Chapman, J., Project # 382e2006. Available at: http://www.ccme.ca/
Fox, D., Hickey, C., Stauber, J., 2014. Revised method assets/pdf/pn_1415_e.pdf.
for deriving Australian and New Zealand water quality Zajdlik, B.A., 2015. A statistical evaluation of the safety factor
guideline values for toxicants. In: Prepared for the Coun- and species sensitivity distribution approach to deriving
cil of Australian Government’s Standing Council on Envi- environmental quality guidelines. Integr. Environ.
ronment and Water, Sydney, Australia. Assess. Manag.. http://dx.doi.org/10.1002/ieam.1694.
Webb, J.M., Smucker, B.J., Bailer, A.J., 2014. Selecting the best Zhang, J., Bailer, A.J., Oris, J.T., 2012. Bayesian approach to
design for nonstandard toxicology experiments. Environ. estimating reproductive inhibition potency in aquatic
Toxicol. Chem. 33, 2399e2406. toxicity testing. Environ. Toxicol. Chem. 31, 916e927.
Wheeler, J.P., Grist, E.P.M., Leung, K.M.Y., Morritt, D.,
Crane, M., 2002a. Species sensitivity distributions: data
and model choice. Mar. Pollut. Bull. 45, 192e202.

You might also like