Professional Documents
Culture Documents
EVALUATION
EVALUATION
EVALU
EVALUATION
ION
ALUATION
EVALUATION
EVALUATIONEVALUATIO
EVALUATION EVALUATION
ATION
LUATIONEVALUAT
EVALUATIONEVALUATION
EVALUATION
TION EVALUATION
EVALU
U.S. Department of State
Washington, D.C.
Acknowledgments
This booklet was prepared by Rolf Sartorius with support from Taryn
Anderson, Michael Bamberger, Danielle de Garca, Mateusz Pucilowski and
Mike Duthie. The authors wish to thank Krishna Kumar and Peter Davis for
their useful comments and suggestions.
Copyright 2013
S O C I A L I M PA C T
Version 2.0
Social Impact
Table of Contents
Part I: Purpose and Overview......................................................................... 5
1. Purpose.......................................................................................................... 6
2. Evaluation types............................................................................................. 7
Performance and process..................................................................................... 7
Summative/ex-post.............................................................................................. 7
Impact evaluations............................................................................................... 7
Global/regional program evaluations................................................................... 8
Experience reviews and meta-analysis.................................................................. 8
Special Studies..................................................................................................... 8
Part II: Evaluation Designs.............................................................................. 9
Quantitative designs.......................................................................................... 14
Qualitative designs............................................................................................ 17
Mixed method designs....................................................................................... 18
Part III: Data Collection Methods................................................................. 25
Mining Project Records and Secondary Data..................................................... 26
Formal Surveys.................................................................................................. 28
Rapid Appraisal Methods.................................................................................. 31
Participatory Methods....................................................................................... 34
Part IV: Evaluation Tools and Approaches..................................................... 37
Results Frameworks........................................................................................... 38
Performance Management Plans........................................................................ 40
Gender Analysis................................................................................................ 42
Cost-Benefit and Cost-Effectiveness Analysis..................................................... 44
Information Communication Technology......................................................... 46
EE
Social Impact
EVALUATION
PART I
EVALUATION
EVALU
Purpose and Overview
PURPOSE
The State Department promotes evaluation to achieve the most effective foreign
policy outcomes and greater accountability to the American people.1 The findings,
conclusions and recommendations generated from evaluations should be used to
improve the effectiveness of State Department programming in the areas of foreign
assistance, diplomacy, security and other operations and to prevent mistakes from
being repeated. The Departments evaluation policy is based on the premise that
evaluations must enable evidence-based decision making and reward candor more
than superficial success stories.
The purpose of this Evaluation Overview is to strengthen awareness of evaluation
tools, methods and approaches, in order to assist the Department and its partners
in their planning and implementation of useful, high-quality evaluations. The
booklet is designed to present some of the most useful and promising evaluation
tools and methods to facilitate planning, managing and using evaluations. For
each tool and approach, you will find a summary of purpose and use, advantages
and disadvantages, relative costs, and the skillsets and time required for their
undertaking. Key references for additional information are also provided.
The tools, methods and approaches in this overview are intended to further serve the
Departments primary evaluation purposes:
AccountabilityWell designed and timely evaluations help to ensure
accountability for the USG resources spent on foreign affairs activities.
Evaluations enable program managers and leadership to determine the cost
effectiveness of programs, projects, initiatives, activities, interventions, etc., and
in the case of a program or project, quality of its planning and implementation.
Consequently, evaluation findings can provide empirical data for reports to
various stakeholders in foreign assistance planning and in larger diplomatic
activities.
LearningEvaluations document the results, impact, or effectiveness of
organizational and programming activities, thereby facilitating learning from
experience. The Department can apply such learning to the development
of new projects, programs, strategies and policies. Empirically-grounded
evaluations also aid informed decision making when considering new programs
or projects, interventions, activities, etc.
The Department distinguishes among several types of evaluation: performance/
process evaluations, summative/ex-post evaluations, impact evaluations, global/
regional program evaluations, experience reviews and special evaluation studies.
The Department anticipates that most of its evaluations will be performance
evaluations due to their ability to generate rapid and cost-effective learning.
1
Social Impact
Types of Evaluations
and impacts, but they often include
effectiveness, and they are conducted
when an effort has ended or is soon to
end. Summative/ex-post evaluations
answer questions such as: What changes
were observed in targeted populations,
organizations or policies during and at
the end of effort? To what extent can
the observed changes be attributed to
it? Were there unintended effects which
were not anticipated at the planning
stage? Were they positive or negative?
What factors explain the intended
and unintended impacts? The essence
of summative/ex-post evaluations lies
in establishing that the changes have
occurred as a result of the intervention,
or at least the latter has substantially
contributed to them.
EVALUATI
This booklet provides an overview of various tools and methods for conducting these
different types of evaluations. The list of topics included here is not intended to be
comprehensive. Some of these tools and approaches are complementary, while some
are substitutes. Although some have broad applicability, others are quite narrow
in their uses. The choice of which tools or approaches are appropriate for a given
context will depend on a range of considerations, including the intended uses of the
evaluation, evaluation stakeholders, the speed with which the information is needed,
and the resources available.
Social Impact
Department's Evaluation Policy does not value any single method over others. A
mixed method approach provides evaluations with a number of advantages:
Triangulation increases validity and credibility of findings and conclusions;
Comprehensiveness allows for collection of enough data to sufficiently answer
all evaluation questions;
ION
Clarity data collected from one method can clarify or supplement findings
collected via another method;
Key resources
`` Department of State (2012). Department of State Program Evaluation Policy.
http://www.state.gov/s/d/rm/rls/evaluation/2012/184556.htm
`` Department of State (2012). Evaluation Guidance for the Department of State.
`` Department of State (2012). Department of State Evaluation Policy: Frequently Asked Questions.
`` Department of State Diplopedia: http://diplopedia.state.gov/index.php?title=State_Program_Evaluations
`` Suggested Evaluation Resources: http://diplopedia.state.gov/images/Suggested_Evaluation_Resources.docx
`` Department of State Evaluation Community of Practice: http://cas.state.gov/evaluation
EE
10
Social Impact
EVALUATION
PART II
EVALUATION
EVALU
Evaluation Designs
11
Overview of Evaluation Designs
Evaluation designs describe the logic and the conceptual framework for answering
the evaluations key questions. DoS recognizes two main kinds of evaluation
designs, each with particular strengths and limitations in addressing specific kinds
of evaluation questionsperformance evaluation designs and impact evaluation
designs.
Performance evaluation (PE) designs are best at answering descriptive and
normative questions about programs. Illustrative descriptive questions can include:
who benefited from the program and who did not? What were strengths and
weaknesses in program implementation? How has the program made a difference?
Normative questions are those which gauge program performance against certain
agreed norms or standards: to what extent did the program achieve its target of
training 400 election supervisors? To what extent were do no harm principles
adhered to during implementation of the peace building program? To what extent
were DoS Green Building standards followed in the construction of the new
embassies? There are many PE designs and to simplify we identify three main types:
1) PE designs using primarily quantitative methods; 2) PE designs using primarily
qualitative methods; and 3) mixed methods PE design. Due to their practicality
in terms of lower costs, faster implementation and greater ability to describe the
interaction of the program and the program context, PE designs are much more
widely used in evaluating DoS programs compared to IE designs.
Impact evaluation (IE) designs are used to answer cause-and-effect questions about
DoS programsto what extent did the program cause outcomes to occur? To what
extent can the program benefits be attributed to the program? For example, an IE
design might be used to answer the question: To what extent did the youth training
program in Tunisia lead to greater employment of youth? In the DoS evaluation
policy IEs use experimental and quasi-experimental designs with a counterfactual
or control group. DoS intends to use these design only very selectively due to their
substantial cost, time and technical requirements.
These main designs are summarized in the following chart and then the advantages
and challenges in using each design type are outlined below.
12
Social Impact
Performance Evaluations
Quantitative designs
without control or
comparison groups
Qualitative designs
Designs that
systematically integrate
quantitative &
qualitative methods
Snapshot design
Before-and-after
Cross-sectional
Appreciative inquiry
Most significant
change
Case study design
Snapshot design
Before-and-after
Cross-sectional
Time Series
13
Quantitative Performance
Evaluation Designs
What are they?
Quantitative performance evaluation (PE) designs rely predominantly on the use
standardized measures and standardized data collection procedures throughout the
evaluation to ensure comparability or to measure results. There a many types of
quantitative PE designs. Several of these that are potentially most useful for DOS
are: 1) quantitative snapshot designs; 2) before and after designs 3) cross sectional
designs; and 4) time series designs. Quantitative PE designs measure program effects
at a single point in time or through repeated measures without a counterfactual
group and do not answer cause-and-effect questions with certainty. These designs are
often very practical and can be used widely for evaluating DoS programs.
ADVANTAGES:
14
Social Impact
COSTS:
Many quantitative PEs are in the middle cost range with snapshot PEs being less
expensive and times series designs being more expensive assuming DOS will pay for
the repeated data collection costs.
SKILLS REQUIRED:
A high degree of skill in survey design, sampling, survey management and analysis,
especially for larger scale survey. Quantitative PE designs using simple surveys will be
less technically demanding
TIME REQUIRED:
KEY RESOURCES
`` Bamberger, M. (2012). RealWorld Evaluation Chapter 12.
`` Gertler, P. Impact Evaluation in Practice
`` Rist, R. and Morra, L. (2009). The Road to Results: Designing and Conducting Effective Development Evaluations.
The World Bank. www.
15
Quantitative PE Designs
Quantitative Snapshot Designs look
at a group of program participants at
one point in time during or after the
intervention. These designs can be used to
answer descriptive or normative questions.
For example, the main design might use
a single simple survey of several dozen to
a few hundred program participants to
answer descriptive questions about how the
program has benefited participants, how
much they liked the program, or how they
rate the quality of program services. The
design can also be used to answer normative
questions against specific targets or criterion
such as did the program achieve its targeted
75% satisfaction rating among program
participants? These designs produce rapid
access to evaluative information and they
are relatively cheap.
Quantitative Before-and-After Designs
can be used to answer descriptive and
normative questions such as: how much
have participants learned during the
program, how income has increased or
how have morbidity or mortality rates,
or incidents of violent conflict decreased
during the program. Evaluators ask about
group characteristics before and after the
program and there is no comparison group.
For example, in a conflict management
program evaluators could do a pre-test
(before the program) and post-test (after
the program) to see how much participants
learned about mediation techniques. Or
in an income generation project evaluator
could measure participant income at
program inception and again at program
completion. These designs require a baseline
or the time and resources to recreate one.
Although each of the above designs are quantitative each could have a qualitative
counterparts. For example, qualitative designs can use the same basic logic of snapshot,
before-and-after, cross-sectional and even time series designs.
16
Social Impact
Qualitative PE Designs
What are they?
Qualitative PE designs draw on a range of primarily qualitative methods to
address evaluation questions where understanding the interplay between program
performance and the program context is central, where evaluations need to be
conducted quickly, flexibly and at low cost; or where complex or rapidly changing
programs require fast, flexible, and on-going learning. These PE designs are often
comprised of a combination of document review, key informant interviews,
focus groups interviews and participatory evaluation methods such as mapping
exercises, or other qualitative approaches such as Appreciative Inquiry or Most
Significant Change methodology. Qualitative PE designs are very practical for
evaluating many kinds of DoS programs, especially those where outcomes such as
poverty, vulnerability, security, and empowerment combine a number of different
dimensions, which can be difficult to observe and measure.
STRENGTHS
``Flexibility to evolve
``Sampling focuses on high value subjects
``Holistic focus
``Examine the broader context of the program
``Multiple qualitative sources provide understanding of complex phenomena
``Narrative reports more accessible to non-specialists
``The use of participatory approaches makes it more likely that vulnerable and
voiceless groups are heard
CHALLENGES
``A flexible, evolving design may frustrate key users of the evaluation and be less
practicable, especially for some short duration evaluations
``Lack of generalizability of findings to other programs
``Multiple perspectiveshard to reach consensus on some major themes
``Methodological challenges in boiling down large quantities of qualitative data
``Interpretivist methods appear too subjective
COSTS, TIME, SKILLS REQUIRED, RESOURCES
Cost, time and skills required will depend on the specific blend of qualitative
methods selected to support the PE design. See Part III for specifics.
17
Mixed Methods Evaluation Designs
What are they?
Mixed methods evaluation designs involve the systematic integration of different
methodologies at all stages of an evaluation. The mixed method approach normally
refers to evaluation designs that combine quantitative and qualitative methods.
It is important to distinguish systematic integration of quantitative and qualitative
methods from many evaluations that combine methods in an ad hoc manner.
The benefits and systematic and integrated use of mixed methods evaluation designs
is widely recognized in all spheres of domestic and international evaluation work.
They are generally the preferred evaluation design DoS programs for the following
reasons:
``DoS programs operate in complex and changing social, economic, ecological
and political contexts and no single evaluation methodology can adequately
describe the interactions among all these different factors.
``DoS program implementation and outcomes are affected by a wide range
of historical, economic, political, cultural, organizational, demographic and
natural environmental factors, all of which require different methodologies for
their assessment.
``DoS programs also produce a range of different outcomes and impacts, many
of which require different methodologies for measurement and evaluation.
``DoS programs change in responsive to how they are perceived by different
segments of the target (and non-target) population, and observing these
processes of behavioral change requires the application of different methods.
ADVANTAGES
Mixed methods designs, when used systematically, offer the potential to combine the
benefits of both qualitative and quantitative approaches while compensating for the
limitations of each approach when used separately. A well-designed mixed method
approach can offer a range of potential benefits:
``A well-designed mixed methods evaluation is able to draw on a much broader
range of qualitative and quantitative tools, techniques and conceptual
frameworks at all stages of the evaluation
``Normally, the design will also incorporate professionals from different
disciplines into the core evaluation team
``Mixed methods designs assist DoS in understanding how local contextual
factors can explain variations in program implementation and outcomes in
different locations
18
Social Impact
Cost, time and skills required will depend on the specific blend of mixed methods
selected to support the PE or IE design. See Part III for specifics
RESOURCES
`` Bamberger, M. (2013). The Mixed Methods Approach to Evaluation. SI Concept Note Series. No. 1., April
2013. http://www.socialimpact.com/evaluationresources
19
Case Study Design
What is it?
A Case Study Design is typically a qualitative or mixed methods evaluation. It
is a non-experimental design and does not use random selection or control and
comparison groups. The design is frequently used when DoS wants to gain in-depth
understanding of a program process, event, or situation and explain why results
occurred. It is useful when answering descriptive evaluation questions about why and
how the intervention works and it can be especially useful in portraying complex
program processes and how the program context interacts with targeted individuals,
communities or institutions to produce results or behavior changes. A key attribute
of a case study design is that it highlights why decisions were made, how decisions
were implemented and finally, with what results. Case study designs can be used
to examine program extremes (high performing or low performing examples) or a
typical intervention.
Case studies can use qualitative, quantitative or mixed methods to collect data. They
can consist of a single case or multiple cases across multiple sites or countries. For
example, a case study design could be used to describe how a program to reintegrate
former child soldiers in Sierra-Leone affected children participating in the program.
The case study would describe the program context, the history and background of
some key individual(s), how they participated in the program and how the program
affected the lives of selected children and their communities upon reintegration.
Rich learning could be gained from cases that portrayed typical children in the
program, or ones who had had experienced particular successes or failures due to
their involvement.
ADVANTAGES:
``Allows for in-depth analysis of interplay between program context and results
``Helps to establish plausible causal relationships between interventions and
outcomes
``Frequently involves a mixed-methods approach strengthening credibility of
results
``Provides a story line that may be compelling for readers
LIMITATIONS:
``Focuses on only one causal relationship, sometime leaving out other potential
relationships
``Difficult to generalize to other situations
20
Social Impact
Generally a minimum of at least 2-3 days required to collect case study data for a
single case such as an individual or a small-scale organization, plus an additional 2-3
days for analysis and report writing. More time if surveys are required. Substantially
more time of the unit of analysis is a large group (Liberian armed forces) or large
organization (Liberian Ministry of Defense), community, or region.
KEY RESOURCES:
`` USAID (2013). Evaluative Case Studies. Technical Note. USAID Monitoring and Evaluation Series. No x.
Version x. (draft)
`` Yin, Robert K. (2009). Case Study Research: Design and Methods. Fourth Edition. Thousand Oaks, CA: Sage.
`` Social Impact (2006). Monitoring, Evaluation and Learning for Fragile State and Peacebuilding Programs: Practical
Tools for Improving Program Performance and Results, pp. 40-47.
21
Impact Evaluation
What is it?
Impact evaluations utilize experimental or quasi-experimental methods to assess
the changes in development outcomes that are directly attributable to a given
intervention. To be able to isolate the impact of an intervention, evaluators must
first identify a credible and rigorously defined counterfactual a theoretical state
that predicts what would have happened to beneficiaries in the absence of the
intervention. The counterfactual is estimated by identifying a comparison group
that is as similar to the beneficiary (treatment) group as possible. Impact is then
measured by comparing the changes over time between the treatment group and
this comparison group. While comparison groups can be selected using a variety
of methodologies, randomized selection of potential beneficiaries into treatment
and control groups provides the strongest evidence of a relationship between the
intervention under study and the outcome measured. Given the complexities of
development work, however, experimental designs entailing randomized selection are
not always possible or desired. In such cases, evaluators should use the most rigorous
quasi-experimental methods available.
22
Social Impact
CHALLENGES:
Impact evaluations generally cost more than performance evaluations. For reference
see the TIPS note on Impact Evaluation Costing.
Skills required:
Varies according to design and scope of evaluation but could take multiple years.
23
Key Resources
`` Gertler, P. et. al. (2011). Impact Evaluation in Practice. The World Bank, Washington, D. C.
`` International Initiative for Impact Evaluation (3ie). http://www.3ieimpact.org/
`` Khandker, S. R., Koolwal, G. B., & Samad, H. A. (2010). Handbook on Impact Evaluation: Quantitative Methods
and Practices. Washington, DC: The International Bank for Reconstruction and Development / The World Bank.
http://go.worldbank.org/9H20R7VMP0
`` World Bank. Development Impact Evaluation Initiative. http://go.worldbank.org/1F1W42VYV0
`` Duflo, E. (2007). Using Randomization in Development Economics Research: A Toolkit. Centre for Economic Policy
Research, London. econ-www.mit.edu/files/806
24
EE
Social Impact
EVALUATION
PART III
EVALUATION
EVALU
Data Collection Methods
25
Mining Project Records and
Secondary Data
What is it?
Data mining uses project documents and records or secondary sources such as
published reports, censuses, surveys and comparative international data during the
evaluation. Project documents that can be mined include periodic project reports
(monthly, biannual, annual), baseline data, needs assessments, grant data bases,
internal and external evaluations, technical advisor input reports, field reports and
project logs and diaries kept by project personnel or beneficiaries. Mining secondary
data can include use of qualitative and ethnographic data such as posters, graffiti,
mass media reports (newspapers, TV, etc.), e-mail and social media (Facebook,
You-Tube, etc.). Examples of widely used comparative international data sets
include: MDG statistics, UN Human Development Index, Demographic and
Health Surveys, World Bank World Development Indicators and Transparency
International.
26
Social Impact
CHALLENGES:
The cost of using published data or well-documented surveys is very low. However,
using project records or data from other organizations may require significant costs
to put the data in a form that can be used for the evaluation.
SKILLS REQUIRED:
Key resources
`` http://betterevaluation.org/plan/describe/existing_documents
`` Bamberger, M. (Nov. 2010). Reconstructing baseline data for impact evaluation and results measurement.
Special Series on the Nuts and Bolts of M&E Systems, No. 4. In Poverty Reduction and Economic Management
Notes. The World Bank. http://siteresources.worldbank.org/INTPOVERTY/Resources/335642-1276521901256/
premnoteME4.pdf
`` Boslaugh, S. (2007). An Introduction to Secondary Analysis. Excerpt from Secondary Data Sources for Public
Health: A Practical Guide. New York: Cambridge University Press. http://assets.cambridge.org/97805218/70016/
excerpt/9780521870016_excerpt.pdf
27
Formal Surveys
What are they?
Formal surveys are used to collect standardized information from a carefully selected
sample of individuals or aggregated units (households, schools, etc.). Surveys often
collect comparable information for a relatively large number of people in particular
project groups.
``Findings from the right sample of respondents can be applied to the wider
target group or the population as a whole
``Quantitative estimates can be made for the size and distribution of impacts
``With the proliferation of donor surveys and national statistics agencies, there
may be good survey data to draw or build on
CHALLENGES:
28
Social Impact
COST:
Ranges from $30-60 per household for the CWIQ to $1-3 million for a full DHS.
Costs may be significantly higher with no master sampling frame for the country.
SKILLS REQUIRED:
Sound technical and analytical skills for sample and questionnaire design, data
analysis, processing and reporting.
TIME REQUIRED:
Depends on the sample size. The CWIQ can be completed in two months. Standard
DHS fieldwork generally requires between three and seven months to complete,
while data collection through the final report takes one year to 18 months.
Key resources
`` Measure DHS. Demographic and Health Surveys. http://www.measuredhs.com/start.cfm
`` Grootaert, C. and van Bastelaer, T. (2002). Understanding and Measuring Social Capital: a multidisciplinary tool for
practitioners. Washington, DC: World Bank.
`` Sapsford, R. (2011). Survey Research (2nd ed.). Newbury Park, CA: Sage Publications.
`` World Bank. Core Welfare Indicator Questionnaire (CWIQ). http://go.worldbank.org/66ELZUGJ30
`` World Bank. Living Standards Measurement Survey (LSMS). http://www.worldbank.org/lsms/
`` World Bank. Quantitative Service Delivery Surveys (QSDS). http://go.worldbank.org/MB54FMT3E0
`` World Bank. Citizen Report Card and Community Scorecard. http://go.worldbank.org/QFAVL64790
29
EE
30
Social Impact
Rapid Appraisal Methods
What are they?
Data collection methods that can be employed quickly and at a low cost to obtain
a narrow, but in-depth understanding of the conditions and needs of the targeted
group. These methods elevate the importance and relevance of local knowledge. Less
structured than classic evaluation methods (i.e., surveys, experiments), they tend to
use a smaller sample size and may therefore have less statistical accuracy.
31
Time required:
32
Social Impact
Key resources
`` McNall, M. and Foster-Fishman, P. (2007). Methods of Rapid Evaluation, Assessment, and Appraisal. American
Journal of Evaluation, 28:151. http://www.pol.ulaval.ca/perfeval/upload/publication_194.pdf
`` USAID. (2007). Using Rapid Appraisal Methods. Performance Monitoring & Evaluation TIPS, 2 Ed., No. 5.
http://www.usaid.gov/policy/evalweb/documents/TIPS-UsingRapidAppraisalMethods.pdf
33
Participatory Methods
What are they?
A collection of methods designed to facilitate ownership of M&E findings and
recommendations among the local population. Project beneficiaries play the primary
role in evaluation planning, data collection, analysis, and reporting. Follow-up
actions are decided upon and implemented locally. The methods are flexible, visual
(sometimes oral) and group-oriented; a small evaluation team facilitates but does not
dictate the process.
Social Impact
Skills required:
Varies widely according to scope of evaluation and methods selected; typically a few
days to a week per community for use of some of the rapid appraisal methods but
much longer for other methods (i.e., participant observer) and for follow-up.
Key resources
`` Harvey, E. (2005). Guide for Participatory Appraisal, Monitoring and Evaluation (PAME). Braamfontein, South
Africa: The MVULA Trust. http://www2.gtz.de/Dokumente/oe44/ecosan/en-guide-participatory-monitoringevaluation.pdf
`` Taylor-Powerll, E., Rossing, B. & Geran, J. (July 1998). Evaluating Collaboratives: Reaching the Potential.
University of Wisconsin-Extension. http://learningstore.uwex.edu/assets/pdfs/G3658-8.PDF [Creative
Expression]
`` UNDP (1997). Who are the Question Makers: A Guide to Participatory Evaluation. http://www.undp.org/
evaluation/documents/who.htm)
35
EE
36
Social Impact
EVALUATION
PART IV
EVALUATION
EVALU
37
Results Frameworks
What are they?
A results framework (RF) is a graphical representation of a development hypothesis.
It demonstrates the causal linkages between all levels of results necessary and
sufficient to achieve a specific bureau or mission goal. These results must be realistic
and achievable, one-dimensional, measurable and within the manageable interest
of the implementing Operating Unit. RFs are based on problem analysis and
information produced by technical analysis and other related assessments.
``Brings the big picture to light through a focus on intervention effects and
outcomes rather than outputs
``Solidifies linkages between program outcomes and national-level goals and
strategies
``Facilitates agreement and understanding on the design and anticipated goals
of interventions, and generates ownership amongst all mission or bureau team
members
CHALLENGES:
Very little for development of actual RF; however, the cost of preceding analyses and
assessments will vary.
38
Social Impact
Skills required:
Actual development may take only a week or two but depends on the results of
assessments and analyses (i.e., environmental, gender, economic, etc.) that may
require up to several months.
Key resources
`` Department of State (2012). Managing for Results: Department of State Project Design Guidebook
`` Department of State (2012). Functional Bureau Strategy Guidance and Instructions
`` Department of State (2012). Integrated Country Strategy Guidance and Instructions
`` Department of State (2012). Joint Regional Strategy Guidance and Instructions
`` USAID. (2010). Building a Results Framework. Performance Monitoring & Evaluation TIPS, 2nd Ed., No. 13.
http://pdf.usaid.gov/pdf_docs/PNADW113.pdf
39
Performance Management Plans
What are they?
Performance management plans (PMP) measure and track progress toward achieving
results by identifying and defining a list of project-related indicators. The plans
typically include an overview of the bureau or mission's management systems,
how the PMP was developed, the relevant results framework, a narrative on the
development hypothesis, indicator reference sheets, an indicator table, and an M&E
task schedule.
``Puts bureau or mission teams on the same page at early stages of project
development
``Forces measurement of change for critical indicators
``Helps keep M&E activities on schedule (i.e., data collection, data quality
assessment, evaluations, etc.)
``Improves knowledge, transparency, and accountability
CHALLENGES:
Minimal for PMP development; baseline data collection costs will vary.
Skills required:
Social Impact
Time required:
May take between 2-4 weeks to develop PMP document, solicit and integrate input
and have reviewed by management. Depending on timing, may decide to wait to
undertake collection of baseline data and establish targets. Prep time may be needed
to develop SOWs for any aspects of PMP development that will be contracted out.
Key resources
`` Department of State (2012). Performance Management Guidebook
41
Gender Analysis
What is it?
An approach to program planning, monitoring and evaluation that assesses the
different ways in which interventions affect men and women, boys and girls, and
people of differing civil status (single, married, divorced, widowed, etc.). This
analysis can be conducted at the project, sector and national levels and ensures that
differences are addressed in the evaluation design, sample selection, data collection
and analysis. It recognizes the limitations of conventional quantitative data
collection methods for discussing sensitive topics such as domestic violence, control
of household resources, sexual harassment, social and political participation, and
gender differences in the labor market.
``Ensures that all sectors of the target population benefit from interventions and
that resources of all sectors of the target community are mobilized
``Ensures efficiency and equity of program impact
``Addresses social, cultural, legal and political factors that limit womens
participation at the household, community, local and national levels
``Addresses sensitive human rights issues such as human trafficking, sex trade
and exposure to HIV/AIDS
CHALLENGES:
``May raise sensitive issues that governments may not wish to address and that
donors may not wish to push
``Uses frameworks and data collection techniques with which many
quantitatively trained researchers may not be familiar, sometimes causing
reluctance to use these techniques
``May require additional resources to contract staff with specialist skills
42
Social Impact
COST:
Depending on the methods used, gender analysis may increase the cost of the
evaluation by 10-20%. In cases where a stand-alone gender analysis is required, the
cost will be similar to comparable conventional evaluation studies.
SKILLS REQUIRED:
Key resources
`` Department of State (2012). Department of State Policy Guidance: Promoting Gender Equality to Achieve Our
National Security and Foreign Policy Objectives. http://www.state.gov/s/gwi/rls/other/2012/187001.htm
`` Bamberger, M. (2005). Handbook for evaluating the impacts of development policies and programs. Developed
for International Program for Development Evaluation Training Workshop. Carleton University, Ottawa. http://
bambergerdevelopmentevaluation.org [click on gender]
43
Cost-Benefit and Cost-Effectiveness
Analysis
What are they?
Cost-benefit and cost-effectiveness analysis are tools for assessing whether or not
the costs of an activity can be justified by its outcomes and impacts. Cost-benefit
analysis measures efficiency by monetizing all inputs, outputs and outcomes.
Cost-effectiveness analysis estimates inputs in monetary terms and outcomes in
non-monetary quantitative terms (such as improvements in student reading scores).
Whereas cost-effectiveness focuses on a particular outcome, cost-benefit analysis
seeks to include all outcomes, each converted to a monetary benefit.
``It is a high quality approach for estimating and comparing the efficiency of
programs and projects
``Makes explicit project assumptions that might otherwise remain implicit or
overlooked at the design stage
``Useful for convincing policy-makers and funders that activity benefits justify
its costs
CHALLENGES:
44
Social Impact
COST:
The procedures used in both types of analyses are often highly technical. They
require skill in economic analysis of programs in the sector and availability of
relevant economic and cost data.
TIME REQUIRED:
Key resources
`` Belli, P., et al. (2000). Economic Analysis of Investment Operations: Analytical Tools and Practical Applications. The
World Bank, Washington, D.C.
`` Millennium Challenge Corporation. (April 2009). Guidelines for Economic and Beneficiary Analysis. http://www.
mcc.gov/documents/guidance/guidance-economicandbeneficiaryanalysis.pdf
45
EE
Information Communication
Technology (ICT) for Evaluation
What is it?
ICT for Evaluation encompasses a broad and growing range of tools to increase
the effectiveness and efficiency of international evaluations. These tools include
use of Personal Date Assistants (PDAs), Smartphones, Netbooks, iPads, email and
web-based surveys, digital photos, audio and video and on-line focus groups to
collect data. In addition, there are technologies for enhancing the communication
of geographically dispersed evaluation teams such as WebEx or Skype. Finally, there
are tools for analyzing quantitative data such as Statistical Package for the Social
Sciences (SPSS) and Statistical Analysis Software (SAS); and tools for analyzing
qualitative data such as the Center for Disease Controls EZ-Text (free) plus
commercial products such as NVivo and Atlas-ti. NVivo and Atlas-ti are tools for
storing, coding, managing, analyzing and retrieving qualitative data.
ADVANTAGES:
46
Social Impact
``Tools like NVivo and Atlas-ti take time to learn; inputting and coding
the data can be time consuming and once categories, codes etc have been
established it may be hard to change them. In field settings capturing
qualitative data on digital recorders may not be practical and producing
quality transcripts for qualitative data analysis may add large and impractical
amounts of time to the data collection level of effort and budget.
COSTS:
ICT experts for PDA and handhelds for programming purposes; tteam members
who are skilled in using specific quantitative and qualitative software packages.
TIME REQUIRED:
KEY RESOURCES:
`` Sue, V. M., & Ritter, L. A. (2012).Conducting online surveys. (2nd ed.). Thousand Oaks, CA: SAGE Publications.
`` http://www.cdc-eztext.com/
`` http://betterevaluation.org/blog/analyzing_data_using_common_software
`` The Evaluation Exchange, Volume X Number 3, Fall 2004. Taking the Next Step: Harnessing the Power of
Technology for Evaluation. Harvard Family Research Project.
47
S O C I A L I M PA C T
S O C I A L I M PA C T
http://www.socialimpact.com