Professional Documents
Culture Documents
Microbiology - Epidemiology PDF
Microbiology - Epidemiology PDF
- An Introduction
Dirk U. Pfeiffer
Epidemiology Division
Department of Veterinary Clinical Sciences,
The Royal Veterinary College,
University of London
September 2002
Postal: Royal Veterinary College, Hawkshead Lane, North Mymms, Hertfordshire AL9 7TA, United Kingdom;
E-mail: pfeiffer@rvc.ac.uk; Fax: +44 (1707) 666 346; Tel: +44 (1707) 666 342
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 2
TABLE OF CONTENTS
BASIC CONCEPTS OF VETERINARY EPIDEMIOLOGY 3
Learning Objectives 3
Challenges to Today’s Veterinary Medicine 3
Veterinary Epidemiology 3
Evidence-based Veterinary Medicine (EBVM) 4
Evidence-based Veterinary Medicine in Practice 4
Basic Epidemiological Concepts 5
Causation 9
DESCRIPTIVE EPIDEMIOLOGY 12
Learning Objectives 12
Measurement of Disease Frequency and Production 12
Survival 15
Standardisation of Risk 16
ANALYTICAL EPIDEMIOLOGY 18
Learning Objectives 18
Introduction 18
Epidemiological Studies 19
Concept of Risk 22
Identification of Risk Factors 22
From Association to Inference in Epidemiological Studies 25
SAMPLING OF ANIMAL POPULATIONS 28
Learning Objectives 28
Introduction 28
Sample Size Considerations 34
INTERPRETATION OF DIAGNOSTIC TESTS 37
Learning Objectives 37
Uncertainty and the Diagnostic Process 37
Diagnostic Tests 38
Evaluation and Comparison of Diagnostic Tests 38
Test Performance and Interpretation at the Individual Level 39
Methods for choosing Normal/Abnormal Criteria 43
Likelihood Ratio 44
Combining Tests 46
Decision Analysis 50
EPIDEMIOLOGICAL ANIMAL DISEASE INFORMATION MANAGEMENT 52
Learning Objectives 52
Outbreak Investigation 52
Assessment of Productivity and Health Status of Livestock Populations 57
Theoretical Epidemiology 58
Information Technology and Veterinary Applications 58
RECOMMENDED READING 60
INDEX 61
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 3
clinician
Causal
Economic
hypothesis
evaluation
testing
best and
Figure 1: Components of veterinary economically
epidemiology sustainable
clinical outcome
and optimum
Evidence-based Veterinary Medicine animal welfare
(EBVM)
animal
In the context of individual animal treatment, owner
animal
the veterinary practitioner has to integrate best
research evidence with clinical expertise and Figure 2: Relationships influencing EBVM
patient or owner values. The best research
evidence has to be derived from varied Evidence-based Veterinary Medicine in
sources such as clinically relevant research Practice
which may be generated by the basic sciences An approach incorporating EBVM into
of veterinary medicine or patient-centred clinical practice consists of the following
clinical research. The latter will be of often steps:
underestimated particular relevance, in that it 1. convert need for information into
is research evidence for example describing answerable questions
the accuracy and precision of diagnostic tests, 2. track down best evidence
the power of prognostic markers or efficacy 3. critically appraise evidence for
and safety of therapeutic, rehabilitative and validity, impact and applicability
preventive regimens. The veterinary 4. integrate critical appraisal with clinical
practitioner is constantly exposed to emerging expertise and patient’s unique biology,
new evidence which may invalidate values and circumstances
previously accepted diagnostic tests and 5. evaluate one’s effectiveness and
treatments and replace them with new more efficiency in steps 1-4 and seek ways
powerful, more accurate, more efficacious and for improvement
safer ones. Veterinary epidemiology provides An EBVM practicing clinician will
the tools which are used to generate research implement these steps depending on the
evidence. frequency of a condition. In any case, they
Clinical expertise is the ability to combine will integrate evidence with patient factors
clinical skills and past experience to rapidly (step 4). But there differences with respect to
identify a patient’s unique health state, the other steps. With conditions encountered
individual risks, benefits of potential every day, one should work in an appraising
interventions, animal welfare needs as well as mode involving constant searching (step 2)
personal values and expectations of the and critical appraisal (step 3) of evidence. In
owner. This is the main focus of the the case of conditions encountered less
undergraduate training. Best research frequently, the EBVM practitioner will
evidence and clinical expertise have to be function in a searching mode which represents
applied in the context of the unique consultation of critical appraisals done by
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 5
others (step 2). And with conditions analogy of an iceberg (see Figure 3). It
encountered very infrequently, they work in a assumes that typically a substantial number of
replicating mode where they blindly seek, animals which were exposed to infection
accept and apply recommendations received remain uninfected and these represent the
from authorities in specific branch of base of the iceberg. These animals could be
veterinary medicine. Unfortunately, this is susceptible to infection in the future or
also the clinical practice mode which is develop immunity as a consequence of past
typically applied by students, recent graduates exposure. Another group of animals may
and residents. It also means that one does not become infected, but has not developed
consider whether the source of advice clinical disease. This group of animals may
operates in an appraising or opinion-based always remain in this category, or could at
mode, and it is therefore difficult to assess some stage develop clinical disease depending
whether the advice is effective, useless or on the influence of different factors including
even harmful. for example environmental stress. The tip of
the iceberg includes animals with different
Basic Epidemiological Concepts manifestations of clinical disease. The ability
The basis for most epidemiological of animals within these different groups to
investigations is the assumption that disease transmit infection becomes a very important
does not occur in a random fashion, because factor in the epidemiology of an infectious
one of their main objectives is to identify disease.
causal relationships between potential risk
factors and outcomes such as disease or
productivity losses. Both types of losses are CLINICAL DEATH
DISEASE
assumed to be influenced by multiple, SEVERE
potentially interacting factors. DISEASE
New cases
6000
cause of the disease and its epidemiological 5000
4000
characteristics. 3000
2000
1000
0
30 1946 1948 1950 1952 1954 1956 1958 1960 1962 1964
25
15
wildlife and dogs in USA
10 Figure 7 shows examples of the four standard
5 types of curves of disease occurrence. In the
0
case of a propagating epidemic, disease could
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 have been introduced through a single source,
Days and subsequently have been transmitted from
infected animals to other susceptible ones
Figure 4: Epidemic curve
within the same population. With sporadic
Clustering of disease occurrence in time can disease occurrence only a small number of
be described as short-term variation as in the cases are observed during a short period of
case of classical epidemics, periodic or time which would infer that the disease
seasonal variation such as in the case of process is not infectious under the prevailing
human leptospirosis in the U.S.A. (see Figure conditions. In the case of a point epidemic a
5) and long-term variation such as with large number of cases have been observed
reported wildlife and dog rabies in the U.S.A. during a relatively short period of time, but
(Figure 6). the disease disappears after that time.
Endemic disease occurrence refers to the
appearance of cases at all times.
250
200
New cases
150
100
50
0
1 2 3 4 5 6 7 8 9 10 11 12
Month of year
New cases
10
0
1 2 3 4 5 6
Time (in weeks)
Propagating
New cases
10
Figure 8: Occurrence of tuberculosis in wild
5
possums during a longitudinal field study
0 (draped over a digital terrain model of the
1 2 3 4 5 6
study site)
Time (in weeks)
The concept of causation links potential risk
Sporadic
factors with the occurrence of disease or
impaired productivity. Knowledge of these
New cases
10
factors potentially allows control or
5 eradication of disease or increase in
0 productivity. Investigations into cause-effect
1 2 3 4 5 6 relationships become useful in populations
Time (in weeks)
where not every individual is affected by the
Endemic problem under study. The objective is then to
measure factors describing variation within
husbandry systems subject to economic,
New cases
10
5
social, physical and biologic parameters.
These factors are also called determinants of
0
1 2 3 4 5 6
health and disease. They can be factors from
Time (in weeks) one or more of these parameter groups, and as
risk factors, they may alter the nature or
Epidemic
frequency of disease or impaired productivity.
Figure 7: Standard types of temporal patterns Determinants of disease include any factor or
of disease occurrence variable, which can affect the frequency of
Disease occurrence can also be characterised disease occurrence in a population. These can
through its spatial pattern which is typically be of an intrinsic nature such as physical or
the consequence of environmental factors physiological characteristics of the host or
differing between locations. Spatial patterns disease agent, or extrinsic such as
can result from variation between regions and environmental influences or interventions by
countries, variation within countries or simply man. Intrinsic factors include disease agents
local patterns. With the advent of which can be living (viruses, bacteria etc.) or
computerised mapping software these types of non-living (heat and cold, water etc.).
analyses have become much more accessible Intrinsic determinants of living disease agents
to epidemiologists. Figure 8 shows the include infection which refers to the invasion
locations used by wild possums (Trichosurus of a living organism by another living
vulpecula Kerr) infected with Mycobacterium organism, infectivity which is the ability of an
bovis draped over a three-dimensional agent to establish itself in a host (ID50 =
representation of the study area. It indicates numbers of agents required to infect 50% of
that locations used by clinically diseased exposed susceptible animals under controlled
possums are clustered in space. conditions) and pathogenicity (or virulence)
which is the ability of an agent to produce
disease in a range of hosts under a range of
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 8
that one sex may be of higher value to farmers identifying cause-effect relationships in
resulting in more care and thereby reduced modern epidemiology. It includes the
disease incidence. Extrinsic determinants of following criteria:
disease affect the interaction between host and
agent. They include factors such as climate, • The proportion of individuals with the
soils and man. disease should be higher in those exposed
to the putative cause than in those not
Causation exposed.
Most scientific investigations are aimed at • The exposure to the putative cause should
identifying cause-effect relationships. be more common in cases than in those
Webster's dictionary defines a cause as without the disease.
"something that brings about an effect or a • The number of new cases should be
result". A cause of a disease is an event, higher in those exposed to the putative
condition, or characteristic which plays an cause than in those not exposed, as shown
essential role in producing an occurrence of in prospective studies.
the disease. Knowledge about cause-and- • Temporally, the disease should follow
effect relationships underlies every exposure to the putative cause.
therapeutic manoeuvre in clinical medicine. • There should be a measurable biological
The situation is complicated if multiple spectrum of host responses.
causes are involved. The Henle - Koch • The host response should be repeatable
postulates developed in 1840 (Henle) and following exposure to the putative cause.
1884 (Koch) were the first set of criteria used • The disease should be reproducible
to provide a generally accepted framework for experimentally.
identifying causes of disease. They demanded • Preventing or modifying the host response
the following criteria to be met before an should decrease or eliminate the
agent could be considered the cause of a expression of disease.
disease: • Elimination of the putative cause should
result in a lower incidence of the disease.
• It has to be present in every case of the
• The relationship should be biologically
disease.
and epidemiologically plausible
• It has to be isolated and grown in pure The web of causation is often used to describe
culture. modern disease problems where presence or
• It has to cause specific disease, when absence of disease is not just a matter of the
inoculated into a susceptible animal and agent being present or absent and the disease
can then be recovered from the animal and occurrence is determined by a complex web
identified. of interacting factors involving agent, host
Koch’s postulates brought a degree of order and environment. Figure 9 presents the causes
and discipline to the study of infectious of tuberculosis in humans as an example of a
diseases, but had the following basic web of causation.
assumptions which were often impossible to
fulfil. They require that a particular disease
Exposure to
has to have only one cause and a particular Malnutrition Mycobacterium
DISEASE
Figure 12: Web of causation for rhinitis in pigs
The presence of Pasteurella is a necessary not the true cause. This illustrates that it
cause for pasteurellosis, but it is not a typically is quite difficult to prove a cause-
necessary cause for pneumonia. effect relationship.
Bias in selection
strong
and measurement
incidence lame
Individual animals
• understand the meaning of survival lame
lame
Withdrawal
Population
September
November
December
Diseased
February
October
January
Animal
August
March
June
April
May
July
A Disease 4 yes 1 no
B 12 no 1 no
C Withdrawn 7 no 1 yes
D Disease 1 yes 1 no
E 12 no 1 no
F Disease 5 yes 1 no
G Disease 10 yes 1 no
H 12 no 1 no
I 12 no 1 no
J Withdrawn 5 no 1 yes
Total 80 4 10 2
Point Prevalence in December 0.50 = 4/8
Point Prevalence in June 0.33 = 3/9
Cumulative Incidence per Year 0.50 = 4/8
Incidence Density per Year
approximate 0.57 = 4 /( (10 - 0.5 * 4 - 0.5 * 2))
exact 0.60 = 4 / ((4 + 12 + 7 + 1 + 12 + 5 + 10 + 12 + 12 + 5)/12)
Figure 17: Calculation of measures of disease cumulative incidence calculation they were
occurrence excluded, and in the case of incidence density
the approximate and the exact calculation
resulted in very similar estimates. Any
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 15
interpretation of the incidence figures should the survivor and hazard function. The
take the risk period into account, which in this survivor function (=cumulative survival
case is one year. probability) is a functional representation of
the proportion of individuals not dying or
Miscellaneous measures of disease becoming diseased beyond a given time at
occurrence risk. It can be interpreted as the probability of
Other measures of disease frequency include
remaining alive for a specific length of time.
the attack rate which is defined as the number
The survival function is often summarised
of new cases divided by the initial population
using the median survival time which is the
at risk. As it is based on the same calculation
time at which 50% of individuals at risk have
as cumulative incidence it really is a subtype
failed (died or became diseased). The hazard
of cumulative incidence. It is confusing
function (=instantaneous failure rate, force of
though that despite its name it is in fact a
mortality, conditional mortality rate, age-
probability and not a rate. The attack rate is
specific failure rate) is calculated by dividing
used when the period at risk is short.
the conditional probability of an individual
Mortality rates are applied using a number of
dying or becoming diseased during a specific
different interpretations, and often do not
time interval provided it has not died or
represent a true rate. The crude mortality rate
become diseased prior to that time divided by
has death as the outcome of interest and is
the specified time interval. This parameter
calculated analogous to incidence density.
does represent a rate expressing the potential
The cause-specific mortality rate is estimated
of failing at time t per unit time given survival
for specific causes of death and also
up until time t. In the context of survival data,
calculated analogous to incidence density.
censoring is an extremely important concept.
Case fatality rate represents the proportion of
In the case of right censoring, individuals are
animals with a specific disease that die from
lost to follow-up or are not dead/diseased at
it. It is a risk measure, not a rate, and is used
the end of the follow-up period. This
to describe the impact of epidemics or the
particular type of censoring can be easily
severity of acute disease.
accounted for in survival analysis by
excluding them from the denominators of the
Survival calculations following their departure from
Time to the occurrence of an event such as
the population at risk. With left censoring,
death or onset of clinical disease can be
beginning of the time at risk is not known,
estimated for many epidemiological data sets
and the commonly used analysis techniques
describing repeated observations on the same
cannot take account of this type of censoring.
sample of animals. This type of data can be
summarised using for example mean survival An example calculation for survival data is
time. This particular method has the presented in Figure 18. Animal A survived for
disadvantage that the estimate will depend on 4 months, animal B survived the whole period
the length of the time period over which data of 12 months and animal C was removed
was collected. If the interval is too short, from the population after 7 months. The
survival time is likely to be estimated number of survivors is based on the actual
incorrectly, as only individuals who number of animals still alive after a given
experienced the event of interest can be time period. The value for the cohort is used
included in the calculation. As an alternative, as the denominator for calculation of
it is possible to calculate an incidence density cumulative survival. The number is adjusted
by using person-years of observation in the for censored observations. The failures
denominator and number of events in the represents the number of deaths during a
enumerator. But this calculation assumes that particular time interval. And the hazard rate is
the rate at which the event occurs is constant the probability of death per unit time (one
throughout the period of study. The most month in this case). The resulting graphs for
appropriate technique for this data is based on
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 16
the survivor and hazard functions are shown available to perform these calculations are
in Figure 19. direct and indirect standardisation.
Direct standardisation
Month of year Direct standardisation
Animal 1 2 3 5 64 7 8 9 10 11 12 involves weighting a set
A Death of observed category-
B specific risk estimates
C CENSORED according to a standard
D Death distribution. First,
E stratum-specific risk rates
F Death are calculated. Then a
G Death
standard population
H
distribution is estimated
I
J
and the proportion of the
CENSORED
4
standard population in
Survivors 10 9 9 9 8 6 6 5 5 5 4
Cohort 10 10 10 10 10 9 9 8 8 8 8 8
each stratum calculated.
Survival 1.00 0.90 0.90 0.90 0.80 0.67 0.67 0.63 0.63 0.63 0.50 0.50 The direct adjusted risk
Failures 0 1 0 0 1 1 0 0 0 0 1 0 rate estimate is obtained
Hazard 0.00 0.11 0.00 0.00 0.13 0.17 0.00 0.00 0.00 0.00 0.25 0.00 as the sum of the products
across the strata between
the proportion of the standard population in
Figure 18: Example calculation for survival stratum i and the observed risk rate estimate
data in stratum i in the study population. As an
example, the mortality in humans in 1963 was
compared between Sweden and Panama. In
1.20
this particular year, Sweden had a population
1.00
Survivor function size of 7.496.000 and 73.555 deaths resulting
in a mortality rate of 0.0098 per year. Panama
Survival probability
0.80
distribution of the study population. This Incidence Rate 0.0005 0.005 0.005
Old
technique is used if stratum-specific risk Cases 4 2 40 20
estimates are not available. But in order to be Person-Years
Incidence Rate 0.002
1000
0.004
10000
0.004
able to use the method, stratum-specific risk Crude Rate 0.0008 0.0049 0.0041
Total Cases 0 54 7 45 20.5
estimates for the standard population and the Total Population 0 11000 11000
frequency of the adjusting factor in the study Calculation 0.0008*7.71 54 / 7 0.0008*2.2 45 / 20.5
population have to be available. As a first Standardized Mortality ratio
Indirect adjusted rate 0.0062
7.71
0.0018
2.20
comparison of different groups and response are then compared between both
investigation of cause-effect relationships. subgroups. Differences in the summary values
Both designs can be based on a census where suggest the presence of an effect of the
all members of the population are included treatment on the response variable. Non-
thus allowing exact measurement of variables observational or experimental studies can be
of interest, or alternatively on a sample where conducted as laboratory experiments or as
a subset of the population is included, thereby field studies such as clinical trials. The latter
providing only estimates of the variables of are usually used to evaluate therapeutic or
interest. preventive effects of particular interventions,
but are also useful to investigate etiologic
Epidemiological Studies relationships. The non-observational study
Epidemiological studies are broadly provides the researcher with effective control
categorised into non-observational (or over the study situation. If the sample size is
experimental studies) and observational large enough a well-designed experiment will
studies. The first group includes clinical trials limit the effect of unwanted factors even if
or intervention studies and the basic principle they are not measurable. Control of factors
is that the design of the study involves other than the treatment which are likely to
deliberately changing population parameters have an effect on disease can be achieved by
and assessing the effect. With this type of using them to define homogeneous subgroups
study an attempt is made to simplify with respect to the status of these variables -
observation by creating suitable conditions blocking or matching- within which treatment
for the study. The second group assumes that is then allocated randomly. The possibility to
the study does not interfere with the have excessive control over the study
population characteristics. Here, the situation can become a weakness of the non-
investigator is only allowed to select suitable observational approach as it may not be
conditions for the study. Observational representative of the real situation in the
studies can be further categorised into biological system anymore. Clinical trials are
prospective cohort studies and retrospective considered the method of choice for
studies as well as cross-sectional studies. investigation of causal hypotheses about the
Retrospective studies include case-control and effectiveness of preventive measures, and
retrospective cohort studies. Another type of compared with the other types of field studies
observational study is the longitudinal study they can provide the strongest evidence about
which is a mix between prospective cohort causality. There is less opportunity for
and repeated cross-sectional studies. Within systematic error compared with the
the group of observational studies mixtures of observational studies. Amongst their
study designs are common, such as for disadvantages are the following
example the case-cohort study. The case characteristics. They require large groups, are
series is a separate group of studies and costly, bias may be introduced through
frequently used in a clinical context. selection error and the required duration can
be long if disease incidence is low.
Non-observational studies
This type of study typically involves dividing
a group of animals into a subgroup which is
being treated and another subgroup which is
being left untreated and acts as a control (see
Figure 22). The decision to treat an animal or
leave it untreated is typically based on random
allocation - randomisation. After a period of
time the status with respect to a response
variable (e.g. disease status) is assessed for
each animal. Summary measures of the
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 20
Exposed with
outcome
Cases
Unexposed Sample
Exposed without
Controls outcome
Unexposed
Time
Onset
Time
Onset of study
of study
No direction of inquiry
Direction of inquiry
Figure 24: Schematic diagram for case-control Figure 25: Schematic diagram of cross-
study sectional study
relationship between a factor and a disease. predictive purposes where knowledge of the
Epidemiological field studies can provide disease risk in individuals with the risk factor
strong support for causal hypotheses. present is used to manage disease. For
Combined epidemiological and other diagnostic purposes, the presence of a known
evidence can lead to the conclusion that a risk factor in an individual increases the
causal hypothesis becomes highly probable. likelihood that the disease is present. If it is a
strong risk factor, absence can be used to rule
Criteria Cross- Case-control Prospective
sectional study cohort study out specific diseases. If the risk factor is also
study the cause of the disease, its removal can be
Sampling random separate separate
sample of samples of samples of used to prevent disease. When assessing the
study diseased and exposed and cause-effect relationship, one should always
population non-diseased non-exposed
units units be aware of potential confounding factors.
Time one point usually follow-up over
retrospective specified
period Identification of Risk Factors
Causality association preliminary causality Epidemiological studies are conducted to
between causal through
disease and hypothesis evidence of identify risk factors through the comparison
risk factor temporality of incidence or prevalence between groups
Risk prevalence none incidence
density, exposed and not exposed to a risk factor.
cumulative Probabilities of disease occurrence can be
incidence
Comparison relative risk, odds ratio relative risk, compared using measures of strength of
of risks odds ratio odds ratio association or measures of potential impact.
Figure 26: Comparison of observational field The first group involves calculation of ratios
studies such as relative risk and odds ratio which
measure the magnitude of a statistically
Concept of Risk significant association between risk factor and
Any investigation into the cause-effect disease. They are used to identify risk factors,
relationships between potential risk factors but do not provide information on absolute
and an outcome parameter such as disease or risk. In contrast, measures of potential impact
death involves calculation of risks. A generic include differences such as the attributable
definition of risk says that it is the probability risk or fractions such as the attributable
of an untoward event. Risk factors include any fraction. These allow quantifying the
factors associated with an increased risk of consequences from exposure to a risk factor,
becoming diseased or to die. Exposure to a and are used to predict, quantify the effect of
risk factor means that an individual has, prevention and to plan control programs.
before becoming ill, been in contact with the
risk factor. Risk assessment is performed Relative risk
intuitively by everyone on a daily basis. Most The relative risk (= RR; risk ratio, cumulative
of the time it is done based on personal incidence ratio or prevalence ratio) is used if
experience, but this approach is insufficient to the following question is asked: How many
establish a relationship between exposure and times more (or less) likely are exposed
disease particularly with an infectious process individuals to get the disease relative to non-
involving long latency periods, with exposure exposed individuals? It is calculated as the
to the risk factor being common, with diseases ratio of cumulative incidence or prevalence
of low incidence or of high prevalence, or in between exposed and non-exposed
the presence of multiple exposures. Under any individuals. Cumulative incidence ratio and
such circumstances it is preferable to base a prevalence ratio are similar if disease
comparison on quantitative estimates of risk duration is unrelated to the risk factor. The
such as cumulative incidence. RR is interpreted as follows: The disease is
RR times more likely to occur among those
The relationship between measures of disease exposed to the suspected risk factor than
frequency and risk factors can be used for among those with no such exposure. If RR is
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 23
close to 1, the exposure is probably not estimated on the basis of data from cohort
associated with the risk of disease. If RR is studies.
greater or smaller than 1, the exposure is
likely to be associated with the risk of disease, Attributable risk
The question "What is the additional risk of
and the greater the departure from 1 the
disease following exposure, over and above
stronger the association. RR cannot be
that experienced by individuals who are not
estimated in case-control studies, as these
exposed ?" can be answered through
studies do not allow calculation of risks.
calculation of the attributable risk (=AR, risk
Odds ratio difference, excess risk, cumulative incidence
The odds ratio (= OR; relative odds, cross- difference or prevalence difference). AR is
product ratio and approximate relative risk) is estimated through subtracting cumulative
calculated as the ratio between the odds of incidence or prevalence of disease in non-
disease in exposed individuals and the odds of exposed from the corresponding values in
disease in non-exposed individuals. It is exposed individuals. It makes the assumption
interpreted as the odds of having the disease that the risk of disease in the un-exposed
among those exposed to the suspected risk group represents the background risk of
factor being OR times the odds of disease disease. The AR is interpreted as the risk of
among those with no such exposure. If OR is developing the disease being increased by AR
close to 1, the exposure is unlikely to be for those individuals exposed to the risk
associated with the risk of disease. For an OR factor. Different estimates are obtained for
greater or smaller than 1, the likelihood that AR in the exposed group and AR in the
the exposure is associated with risk of disease population (PAR). PAR can be estimated by
increases, and the greater the departure from 1 multiplying AR with the prevalence of the
the stronger the potential cause-effect risk factor in the population. The information
relationship. In contrast to RR, OR can be contained in AR combines the relative risk
used irrespective of the study design, and the risk factor prevalence. The larger the
including case-control studies. OR is also AR, the greater the effect of the risk factor on
insensitive to whether death or survival is the exposed group. The parameter cannot be
being analysed. OR can be used to estimate estimated for most case-control studies.
RR if the disease is rare (less than 10%). Odds
and risks have the same enumerator (=the Number needed to treat
Evidence-based medicine has resulted in the
number of diseased), but differ in the
definition of a new method for expressing
denominator, which in the case of odds
attributable risks which is considered to be
includes only events which are not in the
easier to understand in a clinical context. It is
numerator and in the case of risks includes all
called the number needed to treat (NNT), and
events.
is interpreted as the number of animal patients
Rate ratio needed to treat with a therapy during the
If the researcher asks the question "How much duration of a trial in order to prevent one bad
more likely it is to get cases of disease in the outcome. It is calculated as the inverse of the
exposed compared with the non-exposed attributable risk (i.e. 1 / AR). When applying
population?", the rate ratio (incidence rate this parameter, it is important to take the
ratio) is the parameter of choice. It is follow-up period into account which was used
calculated as the ratio of incidence density to generate it. The number needed to harm can
estimates in exposed and unexposed also be calculated.
individuals. Similar to RR and OR, if its value
is close to 1, it is unlikely that the exposure is Attributable fraction
The attributable fraction (= AF; etiologic
associated with the disease frequency. The
fraction, attributable risk) is used to answer
further the value from unity, the more likely it
the question: "What proportion of disease in
is that the exposure is related to disease
the exposed individuals is due to the
frequency. This quantity can only be
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 24
exposure?". AF is calculated as the proportion with two farrowing sheds and a total of 200
that the attributable risk represents within sows with equal numbers going through each
total disease risk in exposed individuals. shed. The design of one of the two sheds
Different estimates are calculated for AF in allows for easy disinfection and cleaning
the exposed group and AF in the population (=good hygiene), whereas the other shed is
(PAF). PAF can be estimated through very difficult to clean (=poor hygiene). The
dividing PAR by the disease prevalence in the relevance of the different epidemiological
population. It is interpreted as the probability measures can be illustrated by estimating the
that randomly selected individuals from a effect of shed hygiene as a potential risk
group/population develop the disease as a factor affecting the cumulative incidence of
result of the risk factor. If the proportion piglet mortality (measured as occurring or not
exposed declines in the general population, occurring on a litter basis) and mastitis-
PAF also decreases, even if RR remains the metritis-agalactiae complex (MMA) in sows.
same. A high PAF implies that the risk factor Summary data for 200 sows and their litters
is important for the general animal over a period of 6 months provides the
population. AF cannot be estimated for most information listed in Table 2. The figure
case-control studies. presented in the table indicate that the risk
factor hygiene status of the shed has the same
RR =
a c strength of association (RR= 5) with both
Disease No Total a + b c + d
disease cumulative incidence of piglet deaths and
Exposed a b a+b
OR =
a c
=
a × d MMA. Attributable risk is considerably
Non-
exposed
c d c+d b d c × b
higher for litter deaths, because they are more
Total a+c b+d N AR =
a
−
c common. Hence, the probability of having
a + b c + d
AR
piglet deaths in a litter given the presence of
=
AF
a (a + b ) NNT =
1
AR
the risk factor is much higher than of having a
sow with MMA. Control of the risk factor
Figure 27: Calculation of different measures (improving the hygiene standard of the
for comparing risk factors farrowing shed) is clearly justified on the
Vaccine efficacy basis of the economic benefits resulting from
Vaccine efficacy (= VE, prevented fraction) decreasing piglet mortality, but not
stands for the proportion of disease prevented necessarily, if it were only to control the
by the vaccine in vaccinated animals. VE is incidence of MMA alone. The proportion of
estimated through subtracting cumulative cases (litters with piglet deaths or sows with
incidence in vaccinated animals from MMA) due to the presence of the risk factor
cumulative incidence in unvaccinated (the attributable fraction) is in both cases the
animals, and dividing the resulting value by same.
the cumulative incidence in unvaccinated
animals. Table 2: Example calculation for a risk factor
comparison
Calculation of measures for comparing risk
Cumulative Incidence
factors
Hygiene Number of Litters with MMA
The recommended method for calculating the Status Sows Deaths
different quantities is to first set up a 2-by-2 Poor 100 0.25 0.05
table as shown in Figure 27. Many computer Good 100 0.05 0.01
programs will automatically perform the Epidemiological Measures
required calculations and also include Relative Risk 5 5
estimates of confidence intervals. Attributable 0.20 0.04
Risk
Example calculation for comparison of risk Attributable 0.8 0.8
factors Fraction
Data on piglet mortality and MMA
occurrence has been collected on a piggery
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 25
a: no interaction
Relative Risk
Cumulative Relative Risk
Factor A Factor B (compared with both
Incidence (within stratum)
factors being absent)
present 0.16 2 8
present
absent 0.08 4
present 0.04 2 2
absent
absent 0.02 reference
equal RR's
b: with interaction
Incidence of Relative Risk Relative Risk
Prosta-
GnRH Pre-breeding (within (compared with both
glandin
Anestrus stratum) factors being absent)
present 0.13 0.37 1.3
present
absent 0.29 2.85
present 0.13 1.35 1.31
absent
absent 0.1 reference
different RR's
High levels
of dry cat
Castration
food
R:33.60
RR:5.00 RR:4.3
R:168.00
R:6.00
R:7.20
R:3.43
Low levels of
outdoor
activity
TARGET POPULATION
COMMON
SENSE
JUDGMENT
STUDY
POPULATION
variance variance and bias
Figure 33: Variance and bias
RANDOM The aim of probability sampling is to obtain
SAMPLING estimates of a variable of interest which are as
close as possible to the true (albeit unknown)
value for the target population. The result
SAMPLE
should be unbiased and the effect of sampling
variation should be minimal. The probability
Figure 32: The sampling process sample will be subject to sampling and
The aim of the sampling process is to draw a systematic error. Sampling error can be
sample which is a true representation of the quantified and expressed through estimates of
population and which leads to estimates of variance or confidence limits. Systematic
population characteristics having an error or bias can manifest itself as non-
acceptable precision or accuracy. Samples observational error (selection bias) such as
can be selected as probability or non- through non-inclusion or non-response, and
probability samples. With non-probability through observational errors including
sampling, the investigator chooses the sample response error and measurement error.
as a convenience sample , where the most Confidence intervals are now commonly used
easily obtainable observations are taken or as to measure sampling variability (not bias). It
a purposive or judgmental sample where expresses how far away a sample estimate can
deliberate subjective choice is used in be from the true value. The correct
deciding what the investigator regards to be a interpretation of a 95% confidence interval is
‘representative sample’. The main that given repeated sampling and calculation
disadvantage of the non-probability sampling of 95% confidence intervals for each sample
approach is that ‘representativeness’ cannot estimate, 95% of them will include the true
be quantified. Probability sampling requires estimate. Variance or the width of confidence
random selection of the sample. Sampling intervals can be influenced through sample
units will be accessed through simple random size, the selection procedure or mathematical
sampling where each animal in the study methods. Doubling the sample size will half
population has the same probability of being the variance and quadrupling the sample size
selected, independently from any other halves the confidence interval. Stratified
animal, or through systematic sampling where sampling after selection of specific groups or
the first unit is selected randomly followed by geographical regions can be used to reduce
selection at equal intervals. variation. Figure 34 demonstrates the effect of
sampling variation based on a study
population of 10 animals where 50% of
animals are of female sex. The estimates
generated by 5 samples of 4 animals each vary
between 25% and 75% female animals.
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 30
Population
F
M 1
2
3
F
M 4
5
F 6
7
M Sample 1 Sample 2 Sample 3 Sample 4 Sample 5
F F M F F F 8
9
M M M M M F 10
F M F F M F
Take a simple random sample
M M M M F M
of 5 animals !
50% 25% 25% 50% 50% 75%
Figure 34: Example of sampling variation Sampling Frame based on Cow Numbers
7 6
8
9
10
5 / 10 = 2
outcome variable, but is not the target of the individual units. Typically, the individual still
analysis. For example, in the case of milk remains the unit of interest such as for
production in a population of dairy cows of example its disease status, but the sampling
the Jersey and Holstein breeds, sampling unit becomes a grouping of individual animals
variation of estimates will be substantial, such as the herd or mob they belong to. All
largely due to genetic differences affecting elements within each randomly selected group
milk volume between the two breeds. are then included in the sample. Therefore,
Stratification on breed will allow reducing the this technique does only require a sampling
overall variation of the milk production frame for the groups, but not for the members
estimate. Allocation of individuals to the within the groups. The groups or clusters can
different strata can be in equal numbers (same represent natural groupings such as litters or
n per stratum) or proportional (same n/N per herds, or they can be based on artificial
stratum). The latter is used if the researchers groupings such as geographic areas or
would like to ensure that the sample has a administrative units. The random selection of
distribution of observations across strata the clusters as the sampling units can be
which is representative of the target performed using simple random, systematic or
population. See Figure 37 for an example of stratified random sampling. See Figure 38 for
applying stratified sampling. The technique an example application of cluster sampling.
will also allow easy access to information With data collected on the basis of cluster
about the sub-populations represented by the sampling, the variance is largely influenced by
strata. For stratified sampling to be effective the number of clusters, not the number of
at reducing variation, the elements within the animals in the sample. The technique assumes
strata should be homogeneous and variance that the elements within the different clusters
between the strata should be large. As a are heterogeneous (unlike stratified
disadvantage, one has to know the status of sampling). It is often applied as part of multi-
the sampling units with respect to the centre trials. Cluster sampling can lead to an
stratification factor and more complex increased sampling variance following the
methods are required to obtain variance saying that “birds of a feather flock together”.
estimates. In this situation, a larger sample size would be
required to reduce variance to acceptable
levels. Large variation between and small
variation within clusters will result in biased
parameter estimates.
Stratify by Breed
1
3
2
1
3 2 4
7 5
4 6
5
12 8
7 6 9
8 10
9 13 11
10 14
Breed A Breed B
Take Take
random sample random sample
1 2
3 Sam ple size of 30 piglets
Random selection of 5 sow s
8 10
9 1 4
8 6 10
Figure 38: Example of cluster sampling Multistage sampling is often used as part of
epidemiological studies. In combination with
Multistage sampling stratified sampling the use of multistage
Multistage sampling involves the use of sampling is recommended by the Office
random sampling at different hierarchical International des Epizooties as part of the
levels of aggregated units of interest. It is official pathway to declaration of freedom
frequently applied as two-stage sampling, from infection with the rinderpest virus (see
where herds are selected randomly as primary Figure 40).
sampling units and within each of the selected
herds animals are selected randomly as
secondary sampling units. See Figure 39 for
an example of the selection process. The
optimal ratio between the number of primary
and secondary sampling units can be
determined on the basis of cost and/or
variability between and within primary
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 33
provisional freedom
freedom from
freedom from
clinical
from Rinderpest
Rinderpest
Rinderpest virus
clinical and
serological random
Vaccination
sample surveys
(during least 2 years)
intention
to eradicate
Rinderpest
at least
at least 3 years 2 years
2 years
-2 -1 0 1 2 3 4 5
no clinical disease
effective system for animal disease surveillance
targetted random sample surveys in risk areasl
Figure 42: Central limit theorem and While it is useful to understand the principles
confidence limits behind these calculations, the required sample
sizes can be obtained much more quickly
P P (1-P)
from tables or specialised epidemiological
0.5 0.25 computer software such as EpiInfo or
0.4 0.24 EpiScope. Figure 44c presents a table of the
0.3 0.21 sample sizes given a 95% confidence level for
0.2 0.16 different prevalence and absolute precision
0.1 0.09 levels. For example, to estimate the
prevalence of disease in a large population to
Figure 43: Prevalence and sample size
within +/- 5% at the 95%confidence level
Estimation of level of disease occurrence with an expected prevalence of 20%, it is
If the objective is to estimate the disease necessary to examine a random sample of 246
frequency, the following information is animals.
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 35
correctly identify diseased animals and The calculation of the Kappa statistic involves
therefore gives an indication of how many estimation of the observed proportion of
false negative results can be expected. agreement and the expected proportion
Specificity on the other hand is the proportion assuming chance agreement as follows:
of animals without the disease which test Observed proportion agreement: OP = (a + d) / n
negative. It represents the ability of the
Expected proportion agreement:
diagnostic tests to correctly identify non- EP = [{(a+b)/n} x {(a+c)/n}] + [{(c+d)/n} x {(b+d)/n}]
diseased animals and gives an indication of
kappa = (OP - EP) / (1 - EP)
how many false positive results can be
expected. The two measures are inversely The following example calculation for the
related and in the case of test results measured kappa statistic is based on the comparison of
on a continuous scale they can be varied by longitudinal and cross-sectional examination
changing the cut-off value. In doing so, an of pig’s heads for diagnosis of atrophic
increase in sensitivity will often result in a rhinitis (see Figure 52). The result indicates
decrease in specificity, and vice versa. The that there is moderate agreement between
optimum cut-off level depends on the both types of examination. The longitudinal
diagnostic strategy. If the primary objective is examination does seem to miss a substantial
to find diseased animals meaning false number of pigs (n=16) with atrophic rhinitis.
negatives are to be minimised and a limited
Longitudinal Cross-sectional Examination
number of false positives is acceptable, a test Examination Atrophy present Atrophy absent
with a high sensitivity and good specificity is Atrophy present 8 (a) 1 (b)
Atrophy absent 16 (c ) 223 (d)
required. If the objective is to make sure that OP = (8 + 223) / 248 = 0.932
every test positive is “truly” diseased EP = [{(8+1)/248} x {(8+16)/248}] + [{(16+223)/248} x
(meaning no false positives, but limited [(1+223)/248}] = 0.874
kappa = (0.932 - 0.874) / (1 - 0.874) = 0.460
amount of false negatives acceptable), the
diagnostic test should have a high specificity Figure 52: Example calculation for kappa
and good sensitivity. statistic
a negative test 98%. If disease prevalence is test result. It can be more than, less than or
only 3%, and test characteristics remain the equal to the true prevalence. Estimates of the
same, the predictive value of a positive test true prevalence can be obtained taking
will be 23% and for a negative test 99.8%. account of test sensitivity and specificity
using the formula presented in Equation 1.
1.00
apparent prevalence + ( specificit y - 1)
0.90 Test with 70% true prevalence =
0.80 sensitivity and 95% specificit y + ( sensitivit y − 1 )
Predictive Value
0.70 specificity
0.60
0.50
0.40 Positive test
Equation 1: Formula for estimation of true
0.30
Negative test prevalence
0.20
0.10
0.00
Calculations for evaluation of tests and test
0 0.05 0.15 0.25 0.35 0.45 0.55 results
Prevalence The different parameters which can be
calculated for diagnostic tests and their results
1.00
are summarised in Figure 54. Rather than
0.90 memorising the calculations it is easier to
Test with 90%
0.80
sensitivity and 80% work through them on the basis of the
Predictive Value
0.70
0.60
specificity relationship between test results and true
0.50 disease status using a 2-by-2 table layout.
0.40
0.30
Positive test
Even if no information about the actual
0.20
Negative test
numbers of animals in each cell is available,
0.10
0.00
the table can be populated with proportions to
0 0.05 0.15 0.25 0.35 0.45 0.55 calculate positive and negative predictive
Prevalence values, as long as prevalence, sensitivity and
specificity are known.
Figure 53: Relationship between prevalence
and positive predictive value for tests with DISEASE NO TOTAL
different sensitivity/specificity DISEASE
Remember the following general rules about TEST a b a+ b
diagnostic tests: POSITIVE
• sensitivity and specificity are independent TEST c d c+ d
of prevalence (note: this is not necessarily NEGATIVE
correct) TOTAL a+ c b+ d N
• if prevalence increases, positive predictive
value increases and negative predictive SENSITIVITY =
a
value decreases a+c
• if prevalence decreases, positive SPECIFICITY =
d
predictive value decreases and negative b+d
predictive value increases a
POSITIVE PREDICTIVE VALUE =
• the more sensitive a test, the better is the a+b
d
negative predictive value NEGATIVE PREDICTIVE VALUE =
c+d
• the more specific a test, the better is the a+b
positive predictive value. APPARENT PREVALENCE =
N
Prevalence estimation with diagnostic Tests a+c
TRUE PREVALENCE =
Tests produce false negatives and false N
positives, therefore any diagnostic test can Figure 54: Formulas for comparison and
only produce an estimate of the apparent interpretation of diagnostic tests
prevalence. The apparent prevalence is the
proportion of all animals that give a positive
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 41
the pre-test probability is revised to 60%. Table 4: Characteristics of serological test kits
Because of the pre-test probability of 60% the for D.immitis infection
dog is likely to have the disease, the Product Sensitivity Specificity
diagnostic test is used to “rule-in” the disease. (in brackets (in brackets
The serological test result is positive, which number of number of non-
diseased dogs in diseased dogs in
increases the probability of heartworm disease
trial) trial)
to above 95%. Dirokit Latex 85.3% (34) + 95.7% (23) +
(2 trials) 82.9% (76) 100% (30)
Case 3 again is a 5 year old untreated dog
Diromail 92.2% (103) 97.4% (78)
presented at a Brisbane clinic with 50% (ELISA)
prevalence in the local dog population (=first Dirocheck 92% (82) + 73% 99% (91) + 94%
estimate of pre-test probability). Because (ELISA)
history and clinical signs are consistent with Filarocheck 97.3% (149) 98.2% (165)
heartworm disease, the clinician revises the (ELISA)
CITE (ELISA) 85% (266) 100% (176)
pre-test probability to 80%. Examination of
the blood smear does not show any evidence
of microfilariae. The pre-test probability is a)
therefore revised to 60%. The serological test PPV =
sensitivit y * prevalence
is applied with the objective in mind to “rule- (sensitivit y * prevalence ) + (1 − specificit y ) * (1 − prevalence )
in” disease. The result from the serology is
b)
negative, which means that the probability of
specificity * (1 − prevalence )
the dog not having heartworm disease NPV =
specificity * (1 − prevalence ) + (1 − sensitivity ) * prevalence
becomes about 80%. To confirm that the dog
does not have the disease, a “rule-out” test Equations 2: Formulas for calculating positive
with high sensitivity should be used to and negative predictive values
confirm the result. The appropriate formulas are shown as
As case 4, a 5 year old untreated dog is Equations 2. The resulting predictive values
presented at a Brisbane clinic because the are presented in Figure 55. Particularly from
owner would like to take the dog to New the perspective of the country allowing the
Zealand, and have the clinician provide a importation of this dog, in this situation the
certificate that the dog is free from D.immitis predictive values of a negative test result are
infection. As above, the local infection important. The results suggest that the
prevalence is about 50% (= first estimate of Dirocheck kit assuming that the sensitivity /
pre-test probability). No clinical work-up and specificity values from the second trial are
no blood smear is used as part of the correct, performs rather poorly with a chance
diagnostic process. A serological test is used of 78% that a negative test result truly is
with the objective to “rule-out” Dirofilaria negative. The Filarocheck kit would appear to
immitis infection. The clinician has the choice provide the best performance, and the
between different test kits. The operating sensitivity/specificity data is based on
characteristics of the serological tests are reasonable sample sizes of dog populations.
presented in Table 4. A comparison of the The possibility remains though that the dog
tests can be based on estimating the predictive population used in the evaluation of this
values for positive and negative test results. particular test while large in size may in fact
not be representative of the typical
characteristics of the dog population in
Brisbane.
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 43
PRIOR
0.5
possible to use the lower 2.5% and higher
PROBABILITY
TEST SENSITIVITY SPECIFICITY PPV NPV 2.5% instead. The percentile method is as
DIROKIT 0.85 0.96 0.95 0.87
0.83 1.00 1.00 0.85 simple as the Gaussian, but has the additional
DIROMAIL 0.92 0.97 0.97 0.93 advantage that it is also applicable to non-
DIROCHECK 0.92 0.99 0.99 0.93
0.73 0.94 0.92 0.78 normal distributions. Its disadvantages are
FILAROCHECK 0.97 0.98 0.98 0.97
CITE 0.85 1.00 1.00 0.87
otherwise the same as for the Gaussian
method.
Figure 55: Calculations for D.immitis In the case of the therapeutic method, the cut-
serological diagnosis example (NPV = negative off value is decided on the basis of the level,
predictive value) at which treatment is recommended. New
results from research will allow adjustment of
Methods for choosing
the value. Its advantage is that only animals
Normal/Abnormal Criteria
which are to be treated will be classified as
The criteria for deriving cut-off values from
diseased. A major disadvantage is its
diagnostic devices measuring quantities on a
dependence on knowledge about therapeutic
continuous scale such as optical densities of
methods which has to be up-to-date.
ELISA tests can be based on a range of
different methods. The most popular The risk factor method uses the presence of
technique is called the Gaussian distribution known causally or statistically related risk
method, but in addition there are the factors to determine disease status. It has an
percentile, therapeutic, risk factor, diagnostic advantage if risk factors are easy to measure
or predictive value and the culturally desirable and it facilitates prevention. But as a
methods. disadvantage, risk of disease may increase
steadily (dose-response) which means that it
The Gaussian distribution method is used to
becomes difficult to determine a cut-off.
derive a cut-off value on the basis of test
Therefore, the positive predictive value may
results from a disease-free population. A
be low.
histogram is drawn to confirm the Gaussian
shape of the data, and the mean and standard Finally, there is the diagnostic or predictive
deviation of the test results are computed. The value method which is considered the most
upper and lower limits of the test results are clinically sound approach. With this
defined using 2 standard deviations. That technique, the cut-off is selected so that it
means 95% of values in the disease-free produces a desired sensitivity and specificity.
population will have values within this This can be done on the basis of the
interval. The other values would be classified information contained in a receiver operating
as abnormal. The advantage of the technique characteristic (ROC) curve. The choice of a
is that it is simple. But there are many suitable cut-off can be influenced by whether
disadvantages. Firstly, the distribution of false-positives or false-negatives are
values is likely to be skewed or bimodal. In considered less desirable by the clinician. The
addition, it is assumed that prevalence is fixed advantages of the predictive value method
whereas in reality it will often vary between include that it can be applied to any value
populations. There is also no biological basis distribution. It uses realistic, clinical data for
for defining disease on the basis of such a cut- the development of the cut-off values. At the
off. True normal ranges will differ between same time, it uses information about the
populations, and the approach does not diseased as well as the non-diseased
recognise that changes over time in normal population, and most importantly it can be
values can be pathologic. adjusted to suit particular diagnostic
objectives. The disadvantage of the method is
With the percentile method, test values are
that it requires monitoring of prevalence,
obtained for a large number of disease-free
positive and negative predictive values, and it
animals, and the lower 95% are classified as
could be seen as a disadvantage that the
normal, the upper 5% as abnormal. It is
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 44
clinician has to choose between a range of the graph. A diagonal ROC curve (from lower
cut-offs. left to upper right corner) indicates a
diagnostic test which does not produce any
Specificity
useful differentiation between disease and
1
1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0
non-diseased. The area under the ROC curve
0.9
>= 80
>= 40
can be used to quantify overall test accuracy.
0.8
0.6
0.02 0.42 >=280
0.5
>= 280
0.10 0.92 >= 80 plotting their ROC curves. The test whose
0.4 0.31 0.97 >= 40
0.3 1 1 ROC curve comes closest to the upper left
0.2
corner is the best test. The ROC Curve can be
used to adjust cut-off values according to
0.1
0
0 0.1 0.2 0.3 0.4 0.5 0.6
quotient between (1-sensitivity) and combination with the initial expectation of the
specificity (see Figure 57). ). It is used less probability that an animal has a certain
frequently than the one for a positive test. condition (= pre-test probability), a revised
This result is then interpreted as how likely it estimate of the overall probability of the
is to find a negative test result in diseased condition (= post-test probability) can be
compared with non-diseased animals. calculated. A complete set of multilevel
likelihood ratios does, in fact, contain the
Likelihood ratios (LR) can be calculated using
same information as a ROC curve.
single cut-off values, so that one obtains only
one pair of likelihood ratios, one for a positive In order to perform the revision of the pre-test
(LR+) and another for a negative test result probability, it has to be converted into a pre-
(LR-). More powerful information can be test odds as described in Figure 57a. The
extracted from the diagnostic test by using result is then multiplied with the likelihood
multilevel likelihood ratios. In this case every ratio for a particular test value to produce an
test value, or more often several ranges of test estimate of the post-test odds. This in turn
values, will have a specific LR+ and LR-. The then has to be converted back into a
advantage of the multilevel LR method is that probability as shown in Figure 58b.
it allows the clinician to take account of the
degree of abnormality, rather than just use
sensitivity
crude categories such as presence or absence LR (+) =
of disease. It will be possible to place more 1 - specificity
emphasis on extremely high or low test values 1 − sensitivity
LR( − ) =
than on borderline ones, when estimating the specificity
probability of disease presence (i.e. post-test
Figure 57: Formulas for likelihood ratio
probability). So for example a LR values
calculations
around 1 will not allow much differentiation
between diseased and non-diseased animals.
These values are often generated for test probability of event
odds =
values around the cut-off point which would a) 1 - probability of event
have been used with the traditional
positive/negative interpretation of a diagnostic odds
test. On the other hand, an LR of 10 will probabilit y =
b)
1 + odds
provide much stronger indication of disease
presence, and would typically be obtained for Figure 58: Calculation of pre-test odds and
test values further away from the post-test probability
aforementioned cut-off value. The multilevel
LR values for each range of test values are
calculated as the ratio between the proportion As an example, somatic cell counts (SCC) are
of diseased animals and the proportion of used as a screening test for sub-clinical
non-diseased animals within that particular mastitis. A particular dairy client is known to
range of test values. have a mastitis prevalence of about 5% in the
cow herd. One cow shows an SCC of 320.
In essence, with both types of likelihood ratio, The probability that the cow does indeed have
the interpreted output of the diagnostic device mastitis can be estimated based on a fixed
for a particular sample will become a SCC threshold or using the likelihood ratio
likelihood ratio value rather than a test- for a positive test. Using the first method,
positive/negative assessment. assuming a threshold of SCC=200 and a
Likelihood ratios (LR) do not depend on sensitivity of 80% and a specificity of 80% a
prevalence, and they provide a quantitative positive predictive value of 17% can be
measure of the diagnostic information calculated (see Figure 59a). Using the
contained in a particular test result. If used in likelihood ratios, the post-test probability
becomes 43.5% which gives the test result a
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 46
Combining Tests
Different diagnostic methods are frequently
used in combination to allow improved
diagnosis through strategic decisions about
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 47
Screening and confirmatory testing Using the same test multiple times
With a strategy of screening and confirmatory The same test can be used in several related
testing, as it is often used in a disease control animals at the same time (aggregate testing),
scheme, the screening test is applied to every of the same herds at different times (negative-
animal in the population to search for test- herd retesting) or in the same animal at
positives. This test should be easy to apply at different times (sequential testing). The first
a low cost. It has to be a highly sensitive test and second approach is particularly important
so that it misses only a small number of with disease control programmes, whereas the
diseased or infected animals. Its specificity third, for example, may be used as part of
should also be reasonable, so that the number research programmes.
of false positives subjected to the
Aggregate testing
confirmatory test remains economically
Most animal disease control programmes rely
justifiable. Negative reactions to the screening
on testing of aggregate populations, such as
test are considered definitive negatives, and
herds. The success of such programmes can
are not submitted to any further tests. Any
be influenced greatly by the performance of
positive screening test result is subjected to a
individual tests, but also by their
confirmatory test. This test can require more
interpretation at the herd level. It is important
technical expertise and more sophisticated
to recognise that with animal disease control
equipment, and may be more expensive,
programmes there are two units to be
because it is only applied to a reduced number
diagnosed, the herd and the individual animal
of samples. But it has to be highly specific,
within the herd. This approach takes
and any positive reaction to the confirmatory
advantage of the epidemiology of an
test is considered a definitive positive.
infectious organism, where in most cases
Comparison of combined test interpretation spread is more likely within than between
strategies herds. In practical terms, this means once an
The different strategies for combining infected animal has been detected within a
multiple test results are compared in Table 7. herd, specific measures are being taken which
are aimed at restricting the possibility of
Table 7: Comparison of methods of combining spread to other herds with susceptible
test results animals. These may involve that none of the
animals from this herd can be moved form the
Test strategy property, or that all of them are being culled,
Considerations Parallel Series Screening/ even though only one animal reacted positive
confirmation to the test.
Effect of increase SE increase SP increase SE and
strategy SP Tests with moderate sensitivity or specificity
Greatest negative test positive test both at the individual animal can be used due to the
predictive value result result aggregate-level interpretation of the test
Application and rapid diagnosis Cost – (herds). The interpretation of the test results
setting assessment when time is effectiveness;
of individual not crucial; test and
from individual animals within the herds is
patients; test and removal usually done in parallel, in order to achieve a
emergencies removal programmes
programmes
increase sensitivity at the aggregate level
Comments useful when useful when useful when
while a less than optimal specificity is often
there is an there is an large numbers considered acceptable. Detection of at least
important important of animals are one positive animal will therefore result in the
penalty for penalty for to be tested
missing false herd being classified as positive. There are
disease positives some diseases where the presence of positive
animals up to a certain number may still be
classified as negative, since test specificity
below 100% will always produce false-
positives.
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 49
It has to be taken into consideration that herd free from the condition. With decreasing
size will influence the probability of prevalence in the population, specificity
identifying infected herds, because the more becomes more important. For example in a
animals are being tested, the more likely it disease-free population, a test with 80%
becomes to detect true positives as well as specificity keeps producing a 20% prevalence.
false positives. When disease control
Sequential testing
programmes commence, prevalence of
Sequential testing is used as part of specific
infected herds can be above 20%. At this
studies where one has the opportunity to
stage, the apparent prevalence will be higher
repeatedly test the same animal over time to
than the true prevalence, as a consequence of
detect seroconversion. This technique is quite
less than 100% specificity of the diagnostic
powerful, as it does not rely on a single result
method. The lower the prevalence becomes,
for interpretation, but rather on a significant
the larger the gap will be between apparent
change in test value which may well remain
and true prevalence. Therefore, the predictive
below a cut-off which would otherwise be
value of positive test results (at aggregate or
classified as non-diseased on the basis of
individual animal level) will decrease and the
single samples. This type of testing could be
proportion of false positives will increase.
used for example during experimental
During the early phase of a control program,
infection studies.
sensitivity is more important than specificity
in order to ensure that all infected animals Using different tests for different disease
within infected herds are being detected. problems in the same animal
During this phase, there will also be a larger Batteries of multiple tests have become quite
number of herds with high prevalence common in small animal practice, where a
compared with the later phases of the control blood sample from a single animal is sent to a
program. As the prevalence decreases, laboratory for assessment of different blood
specificity becomes as important as metabolite levels. The objective is to identify
sensitivity, and it may become necessary that normal and abnormal parameters. Testing for
a second test is carried out using series allergic reactions is an example of this. The
interpretation to further increase specificity technique becomes useful if a set of different
(see screening/confirmatory testing above). It parameters is of diagnostic value for
should be noted though that sensitivity will establishing a pattern that is considered
still be important during this phase, as the suggestive of a particular disease. The
control program cannot afford missing too approach becomes questionable, however, if it
many infected animals. is part of a ‘fishing expedition’ for a
diagnosis. The clinician has to keep in mind
Negative-herd re-testing
that a cut-off for a single test is typically set
Within the context of animal disease control
such that it includes 95% of the normal
programmes, negative-herd re-testing is a
population, which means it will produce 5%
typical testing strategy. This means that only
false positives. As an example, with a battery
animals that were negative to an initial test
of 12 diagnostic tests measuring different
undergo a further test, because any positive
blood parameters, each of them will have a
animals would have been culled.
0.95 probability of diagnosing a ‘normal’
Interpretation of results is usually at the herd
animal correctly as negative. But it also
level. This increases aggregate-level (= herd
means that the overall chance of a correct
level) sensitivity because, if there are diseased
negative diagnosis on all tests is (0.95)12
animals in the herd, even a test with a
amounts to a probability of 0.54. Therefore
moderate sensitivity will eventually detect at
there is a 46% chance that a ‘normal’ animal
least one of them. The testing strategy
has at least one false-positive result among
increases the chance of finding infection
these 12 tests.
missed on previous testing rounds. In
principle, the herd is asked to ‘prove’ that it is
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 50
SURGERY
FAILURE
($400 - $150)
0.1
RECOVER
($1000 - $15)
0.8
MAGNET
ILLNESS
($400 - $15)
0.2
Epidemiological approach
In contrast, the epidemiological approach to
Epidemiological Animal investigation of herd problems removes the
Disease Information assumption of the existence of a ‘normal’
Management standard. It focuses on the comparison of sub-
groups or animals. A systematic approach is
Learning Objectives used to keep the process objective and
At the completion of this topic, you will be unbiased. With simple problems, the
able to: epidemiological will not be distinguishable
• enumerate steps to take during an from the clinical approach. In the case of new
outbreak investigation, including or complex problems, the epidemiological
description of the outbreak by animal, approach becomes the method of choice for
place and time outbreak investigations.
• understand the principles of herd health The epidemiological approach to
and productivity profiling investigations of herd problems includes the
Outbreak Investigation following series of investigational steps: First
An outbreak is a series of disease events the diagnosis has to be verified by making a
clustered in time. Outbreak investigations are definitive or tentative diagnosis, followed by a
used to systematically identify causes (risk clinico-pathological work-up. Then a case
factors) of disease outbreaks or to identify definition is established. Cases should be
patterns of disease occurrence. The defined as precisely as possible, and other
investigator asks the questions: What is the diseases should be excluded. As the next step,
problem? Can something be done to control the magnitude of the problem has to be
it? Can future occurrences be prevented? The determined. Is there an epidemic? The
problem can be approached using the cumulative incidence can be computed and
traditional clinical or the modern compared with normal or expected risks of
epidemiological approach to investigation of disease. Subsequently, the temporal pattern is
herd problems. examined which typically involves
constructing an epidemic curve. If possible,
Clinical approach incubation and exposure period are estimated
With the clinical approach, the focus is on on the basis of the data. Then, the spatial
diseased animals and they are identified on pattern is examined, for example by drawing
the basis of features which distinguish them a sketch map of the area or the layout of the
from normal animals. The clinician can come pens and the number of cases within pens.
up with a correct diagnosis given that the The investigator should inspect the drawing
“true” diagnosis is included in the list of for possible interrelationships among cases,
differential diagnoses taken into and between location and cases and other
consideration. The latter will depend on the physical features. This process is followed by
areas of special expertise, and typically an examination of animal patterns. It includes
clinicians will only include differential investigating factors such as age, sex, breed
diagnoses they are familiar with. The clinical and strain patterns. A list of potential causal
diagnosis can be derived relatively easily if it or non-causal factors associated with the
is a single clearly identifiable disease. The disease is established. Animals are
situation becomes more difficult, if multiple categorised according to the presence of each
disease determinants interact to cause a attribute. This data can be presented as
syndrome. This type of outbreak situation is frequency and attack rate tables. After
better described as a causal web than a causal collating this information it should be
chain. analysed using quantitative methods. As part
of the data analysis, factor-specific disease
risks are computed for each of the potential
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 53
risk factors. The objective is to identify the 12 months prior to the outbreak typically
highest as well as lowest disease risks, the about 7% of the litters suffered from
greatest difference between disease risks, to diarrhoea, now 40% of litters are affected.
estimate the relative and attributable risks. The immediate action taken by the vet
Information about the expected level of included submission of 3 acutely affected pigs
disease is very important in this process. The to the local diagnostic laboratory who
data should not just be looked at by diagnosed one piglet as infected with E.coli
calculating proportions, it is also important to serotype 08, but did not find any pathogenic
assess absolute numbers of animals. Risk bacteria or viruses in the other 2 pigs. The
factors associated with the epidemic are intestinal lesions found in all 3 pigs were
identified by assessing the association consistent with acute enteritis. The
between disease patterns and the distribution veterinarian decides to adopt an
of factors. The objective is to demonstrate that epidemiological approach for investigating
an observed association is not due to chance. the problem and its financial impact.
The result from this analysis should be a
General knowledge about preweaning
working hypothesis taking into account
diarrhoea in pigs indicates that most herds
potential causes, sources, mode of
have endemic neonatal diarrhoea at low levels
transmission, exposure period and population
and experience periodic outbreaks. The
at risk. The hypothesis may have to be
organisms involved are E.coli, rota virus,
revised, if it does not fit all facts. If it is
Salmonella spp., Campylobacter spp.,
possible, the hypothesis should be tested.
cryptosporidia and coccidia, as well as TGE
With complex problems not revealing any
virus in some countries. There is still
quick answers as to the source of the problem,
insufficient knowledge to explain why some
it is often advised to conduct an intensive
litters (herds) are affected and others not.
follow-up. This would involve a clinical,
Figure 62 presents the causal web of factors
pathological, microbiological and
associated with the occurrence of coli-
toxicological examination of tissues, feeds,
bacillosis in neonatal pigs.
objects etc.. Detailed diagrams of feed
preparation or movement of animals could be
Design of shed and crates
prepared and a search for additional cases on Farrowing shed Temperature
Drainage/pen moisture
other premises or outbreaks of similar nature environment
Bedding
Cleaning/disinfection
in other locations could be conducted. If it is All in - all out farrowing
parity and risk of scouring pigs in a litter. The variable to look at is the disease status of the
problem with this particular application of the sow (see Table 9). The relative risk amounts
chi-square statistic is that the expected to 2.7, but also is not statistically significant.
number of counts in at least one cell is less
than 5 observations, which violates the Table 8: Cross-tabulation of aggregated parity
assumptions of this statistical method. categories against presence of dead scouring
Modern statistical software can get around pigs in litter
this problem by using Monte Carlo sampling dead scouring pigs in litter
or exact inference techniques. Applying the parity proportion yes no total
Monte Carlo method results in a significance 1 0.80 4 1 5
level of 0.39 (95% CI 0.38 - 0.40). It should >1 0.33 7 14 21
be kept in mind, that due to the small sample totals 0.42 11 15 26
size the statistical power of this analysis is 4 5
low. RR = = 2. 4
7 21
Parity
Litters with dead scouring pigs
Proportion yes no total
χ 2 = 3. 6;1df ; p = 0. 06
1 0.80 4 1 5 Fisher’s exact test (2-tailed) p=0.13
2 0.60 3 2 5
3 0.25 1 3 4 Table 9: Cross-tabulation of disease status of
4 0.50 2 2 4
5 0.33 1 2 3
sow against presence of dead scouring pigs in
6 0.00 0 3 3 litter
7 0.00 0 1 1
8 0.00 0 1 1 dead scouring pigs in litter
Totals 0.42 11 15 26 sick sow proportion yes no total
yes 1.00 2 0 2
χ 2
= 7 . 9 ; 7 df ; p = 0 . 34
no 0.38 9 15 24
0.80
totals 0.46 11 15 24
0.70
Proportion litters affected
0.60 2 2
0.50 RR = = 2 . 67
0.40
9 24
0.30
0.20 χ 2 = 3. 0; 1df ; p = 0. 09
0.10
0.00
1 2 3 4 5 6 7 8 Fisher’s exact test (2-tailed) p=0.17
Parity
Table 10: Cross-tabulation of litter size confounded, as removal of each from the
categories against presence of dead scouring model did not change the magnitude of the
pigs in litter effect of the other factor to a meaningful
extent. Disease status of the sow was not
dead scouring pigs in litter
litter size proportion yes no total RR important as it did not produce a statistically
<=8 0.13 1 7 8 1 significant regression coefficient. The results
9 - 10 0.50 4 4 8 4 can be interpreted as suggesting that the risk
11 - 12 0.50 4 4 8 4
>=13 1.00 2 0 2 8
of diarrhoea diminishes with parity and with
totals 0.42 11 15 26 distance from the entrance to the shed. The
odds ratio of 0.34 for PARITY indicates that
χ 2 = 6. 0; 3df ; p = 0.11 an increase in one unit in the variable
Table 11: Cross-tabulation of pen location PARITY changes the odds of mortality from
categories against presence of dead scouring scouring in the litter by a factor of 0.34. An
pigs in litter increase in litter size results in an increased
risk of mortality from scouring in the litter.
dead scouring pigs in litter
pens proportion yes no total STATISTIX 4.0 COLIPIGS, 25/03/93, 8:29
analysis it can be concluded that there are LOGISTIC REGRESSION ODDS RATIOS FOR COLI
potentially multiple factors associated with PREDICTOR 95% C.I. 95% C.I.
the disease problem. In this type of situation, VARIABLES LOWER LIMIT ODDS RATIO UPPER LIMIT
--------- ----------- ---------- -----------
it is then necessary to decide on which of PARITY 0.13 0.34 0.95
these factors are most important and which CRATENO
BORN
0.43
1.06
0.66
3.70
1.01
12.88
are likely to be confounding factors such as
could be for example gilts which may be more Figure 67: Output generated by multiple
likely to be placed near the entrance. logistic regression analysis of piglet scouring
Statistical analysis methods for multivariable data
problems are appropriate with this type of The results of this data analysis have to be
data. In this case, logistic regression is the interpreted very carefully. Firstly, in this
method of choice as it allows multivariate particular case the statistical power was quite
analysis of binary outcome variables. The low due to the small number of 26 litters
outcome variable is whether E. coli diarrhoea available for analysis. Under such
was present or not present in a litter. The circumstances, the investigator could use a
result of the logistic regression analysis 10% significance level, but should be aware
indicates that parity number (PARITY) is the that this will increase the risk of detecting
most important risk factor, but litter size statistically significant associations (=type I
(BORN) and crate location (CRATENO) were error) which in reality might not exist. With
also included (see Figure 67). There was no any observational study one should be aware
statistically significant interaction between of the potential for confounding factors. In
crate and parity number which suggests that this particular data set, the gilts not being
the two risk factors are not dependent. It is familiar with the farrowing facilities may
also unlikely that the two factors were
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 57
have been more likely to be placed in pens and disinfection program on the farm.
close to the entrance to the farrowing shed. Additional samples for microbiology could be
The potential for such a confounding collected from untreated pigs with diarrhoea
relationship between parity category and crate and pigs without diarrhoea. The investigator
location was eliminated as part of the should remember that with this type of
multivariate analysis. Taking the results problem it is rather unlikely to have only one
together, risk of diarrhoea does seem to be causal agent involved and that culture has
related to parity and litter size which could poor sensitivity. This means for the laboratory
both indicate insufficient passive immunity. not to isolate any other potentially causative
The variable crate location may be an organisms does not necessarily mean that they
indicator for inadequate disinfection and have not been present. There is also the
cleaning procedures. Hence, there are two possibility that the herd has a problem with
main hypotheses which could be further post-farrowing illness in sows, but given the
investigated to come up with a set of suitable small number of only 2 cases it is unlikely to
preventive measures. To test the hypothesis of have a major impact in the current problem. In
poor passive immunity, the investigator could any case, it would be recommended to
consider conducting an intervention study of monitor piglet mortality over a period of time.
sow vaccination (E.coli). This will also allow review of the hypotheses
as more data becomes available. Finally, the
From the farmer's perspective, an economic
investigator should produce a report for the
evaluation will be important. The preweaning
farmer describing the outcome of the
mortality is currently at 19%, but used to be
investigation.
11.5% which means that 7.5% are attributable
to the current scouring episode. This excess
Assessment of Productivity and Health
mortality of 7.5% in a total of 253 pigs born
Status of Livestock Populations
alive results in 19 piglet deaths, of which 1 Traditionally, animal disease control mainly
(5%) would be expected to die after weaning, focused on monitoring of disease outbreaks
leaving 18 pigs for the remaining calculations. and movements of disease agents. Nowadays,
During the last year 7% of litters were treated, it has become important to collect information
meaning that 1.8 litters (7%) of the current 26 for setting priorities and defining actions
litters would be expected to be treated under against diseases common in livestock
normal circumstances, but in reality 8 (31%) population. Health and productivity profiling
of the 26 litters had to be treated in is one particular method that can be used for
association with the outbreak. This results in this objective. With this method of data
opportunity costs for 18 pigs of 18* $35, collection, a limited number of representative
treatment costs for 6 litters of 6 * $10 sample units such as for example a small
amounting to a total of $690 which comes to number of herds or villages are being used to
about $26.50 per sow farrowed. Sow collect detailed information on production,
vaccination costs are estimated as $5 for the management and disease. This data forms the
vaccine plus $0.30 for labour resulting in a basis of epidemiological analysis for
total of $5.30. The benefit/cost ratio is investigation of complex relationships
therefore 26.5 / 5.3 = 5. This means including interactions between multiple
introduction of vaccination could very factors and diseases. The results can be
beneficial. For every dollar spent, 5 dollars applied to defining action priorities for
could be earned. disease control programmes based on
As a consequence of this analysis, the economic or other considerations.
following action could be taken. Firstly, the In developing countries an effective
intervention study of sow vaccination with government veterinary service often is not
E.coli vaccine mentioned above could be economically sustainable, but can be
conducted. The spatial pattern could be supplemented by a primary animal health
further investigated by reviewing the cleaning
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 58
care service which employs so called Software applications used for individual
"barefoot" veterinarians to provide the most cases apply a probabilistic approach to
common veterinary services to farmers. These veterinary diagnosis and therapy. Examples of
priority services can be defined on the basis of such systems are BOVID and CANID which
a basic set of health and productivity are expert systems used for diagnoses of cattle
indicators which are being continually and canine disease problems, respectively.
monitored using village animal health and Herd health management increasingly
production monitoring systems. involves the use of computerised recording
and monitoring systems such as DairyWIN or
Theoretical Epidemiology PigWIN. Eventually these systems will
Simulation modelling is an area which will develop into decision support systems (see
become more important as part of decision Figure 68).
support in agriculture in the future. Models
are used to represent dynamic processes or comparative accounting
systems and simulate their behaviour through slaughter
analyses
check
time. They mimic the system under study
which allows testing of specific hypotheses
for example about the epidemiology of an electronic
simulation
PigWIN models
infectious process or to identify gaps in our data entry
strategies. There are two main groups of Figure 68: Components of a decision support
simulation models, those based on system for pig production
mathematical equations which are called
deterministic models and those based on Modern national disease control programs
probabilistic sampling of distributions which have computerised database management
are called stochastic models. The tools used systems as an essential component. Examples
for the development of such models include are the New Zealand national tuberculosis
computerised spreadsheets which allow non- database and the ANIMO system used by the
programmers to construct simple models. European Union to monitor animal
Computer programming languages have to be movements. More recently, geographical
used when developing more complex models. information system technology has developed
to a level where it can be used by disease
Information Technology and control managers at a farm as well as at the
Veterinary Applications regional level (see Figure 69). For these
Modern information technology will in the systems to evolve into a decision support
near future become part of the tool set which system they should integrate database
the veterinarian in the 20th century will have management, simulation modelling, decision
to work with. Computerised information analysis as well as expert system components.
retrieval systems have been developed such as The computerised emergency response system
for example the electronic edition of the for outbreaks of foot-and-mouth-disease
MERCK VETERINARY MANUAL. called EpiMAN developed in New Zealand is
Electronic literature reference databases such one of the most sophisticated such systems
as Current Contents can now be accessed currently available (see Figure 70).
through the World Wide Web. On the more
technical side, artificial intelligence methods
including knowledge-based systems (expert
systems) or neural networks are being used to
design decision support systems which can
help the veterinarian in the diagnostic process.
D.U. Pfeiffer Veterinary Epidemiology - An Introduction 59
Reactors/100,000 tested
MAPS
AGRIBASE
EPIDEMIC DATA
ACTIVITY FORMS
STANDARD USER
INTERFACE
DATA
STANDARD
INPUT & REPORTS
EPIMAN DATABASE
"fi red" ru le
rejected
EPIDEMIOLOGICAL
PARAMETERS
STATISTICAL ANALYSIS
SIMULATION
EPIDEMIOLOGIST'S
WORKBENCH
STATUS MAPS
STATISTICAL REPORTS
DECISION SUPPORT
—R— —W—
rate ratio, 23 web of causation, 9
relative odds, 23
relative risk, 22
— —
retrospective studies, 19 α- error, 25
risk, 22
risk assessment, 22
risk difference, 23
risk factor method, 44
risk factors, 22
risk ratio, 22
ROC Analysis, 45
—S—
sample, 19
sampling, 28
sampling fraction, 28
sampling frame, 28
sampling to detect disease, 36, 37
sampling to measure continuous variable, 37
sampling unit, 28
selection bias, 25
sensitivity, 39
simple random sampling, 29, 30
simulation modelling, 59
sources of error in probability sampling, 29
spatial pattern, 7, 53
specificity, 40