Professional Documents
Culture Documents
1
How To Use This Guide
This guidebook, which includes four key sections, is meant to serve as a helpful resource for
different types of innovators applying for Grand Challenges Canada (GCC) grants across various
stages of project negotiation and implementation.
The materials contained in this guide have been developed in collaboration with
implementation teams to build a useful, feasible, and effective monitoring, evaluation, and
learning process (aka “MEL”). The guide provides an overview of the integrated process and
system, incorporating both required and recommended stages and documents. It is meant to be
used as a workbook to help you build and strengthen these processes. Please use relevant
sections as you develop your implementation process system and strategy. You are encouraged
to work at your own pace and access certain parts depending on your needs and application
stage.
The guide can also be used to review what you may have already produced on your own and
aims to help build on your existing capacity beyond any particular grant cycle. For example, for
potential Seed innovators and TTS innovators, the guide may be helpful at the application stage
so that innovators have an opportunity to leverage the content within this guide as
recommendations or guidance while developing key sections of their funding application,
whereas for Round 2 Seed innovators, different sections may be helpful to review and
operationalize at different points (e.g., Section 1-3 during the negotiation of your award, and
section 4 during early stages of implementation).
We hope this guide will help innovators develop a systematic and consistent process and will be
useful for beginners and experienced implementation teams alike. Thus, even for those who are
already familiar with this material, we suggest that you review your existing MEL system and
processes and draw on the experience of others to strengthen your own processes. Ultimately,
the guide is meant to support you and augment your experience as you work through different
stages – please feel free to use it as it best suits your needs.
Acknowledgements
The content of this guide has developed over a long-term collaborative effort between the
Grand Challenges Canada, innovator teams and Penny Holding, from the Saving Brains Learning
Platform (hosted by TruePoint) and the Global Mental Health Learning Platform (hosted by
citiesRISE). This material has been developed to be open access and we hope this is of help to
all innovators.
2
CONTENTS
3
4
Part 1: Overview of the Integrated Monitoring,
Evaluation, and Learning Process
5
1.1 What is Monitoring, Evaluation, and Learning (MEL)?
While MEL is an important part of the reporting process, it is much more than that. It is an
integral part of decision making around your implementation. Questions to consider include:
These questions should be answered in the context of the implementation program as a whole,
its stage of development, and current future goals and objectives. The purpose of this process,
and how to plan for it, is summarized in the following document.
To be useful for decision making, MEL needs to be integrated. The next figure illustrates the
cycle of implementation and the key targets of an integrated MEL process. The process from
Design to Evaluation corresponds to key documents that Grand Challenges Canada either
requires or recommends at the start of the funding cycle or during implementation. Utilizing
them facilitates the reporting process. The 4 documents are:
1. A Theory of Change model (Recommended)
2. Project Description and Key Objective (Required submission)
6
3. A Monitoring and Evaluation Plan (Recommended for Proof of Concept projects,
required for some Transition to Scale projects)
4. Results based Monitoring and Accountability Framework - RMAF (Required submission)
As you plan your integrated MEL process, a useful approach is Measurement for Change (M4C)1.
This is an approach to decision making that places collaborative and evidenced based decision
making at the heart of building effective and durable delivery.
1
The Aspirations of Measurement for Change. (2020) Krapels J, van der Haar L, Slemming W, de Laat J, Radner J, Sanou
AS, Holding P. Front Public Health. Nov 26;8:568677. doi: 10.3389/fpubh.2020.568677. PMID: 33330315; PMCID:
PMC7732529. https://www.frontiersin.org/articles/10.3389/fpubh.2020.568677/full
7
The graphic above summarizes reflection points through which to review the effectiveness of an
MEL process. The five shaded boxes represent aspirations for moving MEL beyond
measurement of change to a process of measurement that contributes to change
(measurement for change, or “M4C”). The objective is to strengthen the cycle of
implementation, by addressing impact, sustainability and feasibility in context.
These five aspirations for a MEL process focus on flexibility and innovation in identifying and
using information in decision making. The three key questions are:
● What information is required, at this time, and place?
● How can it be collected feasibly and accurately? And,
● How will it be shared and utilized?
M4C adheres to the same principles as any approach to MEL. Decisions should be based on
information that is collected in a systematic, careful, and consistent manner. The following
figure summarises factors that influence the potential value of information collected, relevant
to any methodological process.
8
1.5 Theory of Change
Developing the MEL process begins with clarifying what change is anticipated (the objectives),
and what processes have the potential to create change. Start by visualising the key, active
ingredients of your program. Document how your program works, or will work, in a Theory of
Change Model, or flow diagram.
The figure below provides an example of key stages of the process, along with some examples
(in the grey boxes) of potential activities or targets of activities to include. Additional Resources
with further detail to guide the development and or review of a TOC model are contained in Part
2 below, which is dedicated to theories of change.
9
1.6 Organising the Underlying Structure and Level of Detail
Another aspect to consider is the level of detail of the information you are collecting. Here we
outline these levels, from the broad detail of an overarching concept (lives improved) to very
fine, or granular detail of a specific indicator (e.g. access to psychological support). We cannot
measure overarching concepts directly. How we build up the level of detail from individual
indicators will influence the accuracy and generalisability of our conclusions on achievement of
our broader goals.
10
Summary
1. The MEL process is an integral and integrated part of the implementation programme.
2. Documentation on monitoring and evaluation must be linked to documentation on
implementation, communication, and reporting.
3. Structure all these documents carefully to address four levels of detail. Be very clear to
separate out the different levels of detail, as illustrated in the figure above.
11
12
Part 2: Theory of Change
2.1 Introduction 14
2.2 Component Guide: Developing a ToC Model 22
13
2.1 Introduction
Closely linked to a ToC model is a Logic Framework or Logic Model. Both include inputs &
outputs, resources, key activities, outcomes (short, medium, and long term) of your
programme. The Logic Model will take a more focused approach, on linear progression along a
specific pathway. It ignores the complexity of the connections across pathways, in favour of
outlining the detail of a specific area of implementation. A Logic Model can be drawn up for all
the pathways of a ToC.
The Link Between the ToC, MEL Plan, and the RMAF
Each component of the ToC should be linked to an indicator or indicators that help you clarify
their contribution. The RMAF (Results based Management and Accountability Framework)
provides a systematic way of recording these indicators and tracking progress towards your
goals.
How the tracking of indicators will occur, what data collection and analysis will be carried out,
and by whom, will be documented in a Monitoring, Evaluation, and Learning (MEL) Plan.
For the process to be useful and support learning, a ToC needs to be seen as a dynamic model,
not inflexible, but one that is used to reflect on outcomes and frame future directions.
14
Making the ToC a Dynamic and Living Model
The following figure outlines the cycle of recommended activities making the ToC an integral
part of the decision making process.
15
Examples of Theory of Change Models
There is no fixed way to organise a ToC model. The various models shared below illustrate
differing levels of detail, reflecting the capacity to manage increased complexity through the
addition of additional pathways and measures.
The first graphic below depicts a Logic Model framework, and focuses only on describing a
single pathway. The last graphic is a complex ToC that summarises learning across multiple
programmes. The ToC approach provides a broader picture, and guides reflection on contextual
influences, achievements, and challenges.
When you have had a look at these examples, use the ToC checklist at the end of this guide to
develop/review your own ToC.
In the next graphic you can see the overlap between a ToC and Logic Model. You can still think
in terms of key stages (labeled in the grey boxes). You will also want to identify the activities or
components that will take place (with examples of what to consider including listed in the blue
boxes). While a Logic Model is less complex, a ToC helps to describe the interconnection
between different parts of your programme.
16
In a ToC you should also build in points of regular review, where progress and achievement is
shared, and interconnections explored. Decisions should then be made about consolidation,
change, or the additions of new components or activities to enhance the effectiveness of your
programme.
When you are building a strategy to develop a durable delivery model it is vital that you gain
insight into the interconnections between pathways. Simple sequential outlines of activities
have a limited capacity to guide you towards a transition to scale over time. The following
examples of ToC models come from a variety of programmes, showing how the interconnection
of pathways have been conceptualised.
17
Example 1:
This model illustrates the evolution of a programme, linking it to time and place in its transition to scale. Here the
focus is on spreading the core idea (the 5 Pillar Approach) through partners and other agencies. The pathways are
identified to build the shared leadership model.
18
Example 2:
This model illustrates the complexity of thinking that emerges as information on programme influences and
influencers is gathered. The overall strategy is to change the culture. The programme team has identified the
multiple, and interrelated, components and pathways that have been seen, and are expected to be seen, to
influence achievement of that overall goal. They have added to their model as the awareness and evidence of the
contributing factors has emerged.
19
Example 3:
The next model also illustrates this same complexity, but using a more linear framework. Often, a concern of teams
is how to manage a complex programme.
Make sure your programme, and model, also reflects feasibility. It may be that several of the components you see in
the next model are relevant, but that you feel you cannot manage to address them all at the same time.
A ToC model therefore helps you think through that complexity; prioritise steps; focus on what is feasible,
manageable and affordable; prune back activities that fail to contribute to effective change; and add in activities
that strengthen the process.
From Collom JRD, Davidson J, Sweet D, et al Development of a peer-led, network mapping intervention to improve the health of individuals
with severe mental illnesses: protocol for a pilot study. BMJ Open 2019;9:e023768. doi: 10.1136/bmjopen-2018-023768
20
Example 4:
This final model summarises the experience across multiple studies from the Grand Challenges Canada programme
in Early child Development - Saving Brains. Looking at the detail provides another approach for how to draw a ToC.
Of particular interest is to note that, even at the early stage of development, programmes considered the ultimate
goal (scaling of the idea).
At every stage, consideration was also given not only to outcomes, but also to the process of delivery, and to the
influence of context (time and place) on framing the pathways of change.
21
2.2 Component Guide: Developing a ToC Model
Does your model and review process capture the following components?
The list is not exhaustive, but is designed to help you adapt the models displayed in the slides
for your purpose.
Discuss Notes
Accounting For Key Components That Lead Towards Understanding Impact at Scale
Workshops
22
Publications
23
Part 3: Results Based Management & Accountability
Framework (RMAF)
24
Part 3: Results Based Management & Accountability Framework (RMAF) 25
3.1 Introduction to the RMAF 26
3.2 Building The RMAF 27
3.3 The RMAF: Good Practices 27
3.4 General Overview of the GCC RMAF 28
3.5 Case Study 35
25
3.1 Introduction to the RMAF
Using the RMAF, not just filling it in, provides a systematic structure to evaluate the practicalities
of where you plan to go, where you have reached, and where you can go next. The RMAF
provides a framework to review and reflect against clear objectives, benchmarks, and
expectations. When planning your RMAF the key questions are: what are
the objectives of the programme? And, what are the key concepts? The
details you use to fill in appropriate rows of the RMAF must relate to
these objectives.
26
● What are the weaknesses that should be ironed out?
● How sure are you that what happened in the end was as a consequence of what
came before, or are you missing some information?
B. Link this outline to the learning/research you are focused on. Just to make it clear– express
your research or learning questions through writing down the main questions you are asking
yourselves.
C. Moving now into the RMAF itself, how are you able to link the GCC RMAF to your programme
objectives, the challenges you face, and ideas around how to overcome them?
For an example of this exercise, please read the Green String Network Case Study included in this
booklet.
Look for those indicator rows in the GMH RMAF that link to your key questions.
● Which indicator row(s) map directly onto your research questions?
● Which indicator rows will need you to review your questions/processes to fit more
directly into the GCC reporting process (without losing a connection to your overall
objectives)?
● Are there indicator rows that you are really struggling with?
27
Reflect upon your own experiences, and answer the following questions…
The RMAF is a comprehensive spreadsheet in which you are required to capture, in the main,
quantitative data that tracks the progress and achievements of your program. This spread sheet
will be supplemented with the Progress Report, in which the link between both quantitative and
qualitative information will be described in more detail. The narrative in the Progress Report
should clearly reference the RMAF.
Each row in the RMAF reflects specific indicators of interest to GCC, on which you will be
required to report. Where an indicator does not appear to be relevant to your program this
needs to be explained, and the requirement may later be discussed with GCC. Please note that
the RMAF may not encompass a complete list of the indicators that your team may need to
track for your programme. If this is the case, you are encouraged to consider additional
indicators and track them for your own MEL purposes. However, you are only required to report
on the indicators in your RMAF to GCC.
There are two tabs in the spreadsheet: General info, where you can provide basic information
about the innovation and your team members. The second is the RMAF table tab. Here is a
guide to the layout:
● Indicator Description (Column E): Each row of the RMAF is dedicated to tracking
progress towards a specific indicator. In following the instructions and guiding notes
provided in Column C, you have the opportunity to define your indicators in Column E.
● Expected Results (Columns G-J): This section captures the overall targets that you will
work towards achieving throughout the entire funding period. GCC encourages
28
innovators to set targets that are realistic and achievable within the time limits of the
funding period. These targets should not be changed without prior approval from GCC.
● Achieved Results at 6, 12, 18, and 24 months (Columns L-AD): This section captures the
data that describes cumulative achievements at each reporting period. This means that
at 12 months you will be adding to the data reported at 6 months. Each reporting period
will therefore describe progress to date across the funding cycle.
● Data Source, Data Collection Methods, Team Members Responsible, Frequency, and
Additional Explanations (Columns AE-AJ): In this section the data capture and data
collection methods for each indicator should be described. It is important that you fill in
this section for each indicator. The Progress Report can be used to expand the
description of the methodology. Here are some things for you to consider when filling in
this section:
o Data Source: What data will you use to track your results, and where is this data
located? For example, hospital records, demographic health surveys, training
logs, project documents, etc.
o Data Collection Method: How will you gather your data? For example, through
pre/post-tests, surveys, or aggregating results from registers or logbooks.
o Frequency: How often will you gather your data? For example, monthly,
semiannually, annually?
o Responsibility: Which team member will be responsible for collecting the data?
For example, the program manager, field staff, or M&E officer.
o Any additional details: This section is very important for explaining any
additional items that GCC should know about your targets and methods. For
example:
▪ If you are using a sample, what is the sample size and sampling method?
▪ What information did you use to help inform your targets? For example,
demographic surveys, research publications, or monitoring and
evaluation data from previous programming.
▪ If you are using scoring tools such as WHO-5 or PHQ-9, what scores do
you think beneficiaries will report at baseline and endline?
It is strongly recommended that you use this section to think through and frame the research/
learning question that you will use data on the indicator to answer. For example, lives
improved indicators, your question will clarify your definition of lives improved, which in turn
will help understand the choice of measurement tool, time frame, etc.
● Age and Gender Disaggregation (Column F): Innovators are required to disaggregate all
results by gender and age group. This information is likely to be estimated using
population demographics when you are first establishing your targets, or may be
informed by prior experience. However, when you report actual results, you need to
track these categories with more precision.
29
Key Sections of the RMAF
30
Dimension Description Example(s): Dimensional Example(s):
Components Measurement tools
31
more distal, community
and societal levels.
32
● Be very clear how you define ‘improvement”, so that this indicator can be measured
consistently throughout the program.
● For the purposes of the RMAF, GCC is particularly interested in applying standardized
approaches to measuring knowledge change. As such, the minimum requirement for
knowledge improvement is a pre-post knowledge assessment. You may choose to design
your own or use existing tools. While the RMAF indicator lends itself to a quantitative
measure such as pre/post-test, this can be further supplemented with
qualitative/observational data.
33
○ Will you be monitoring attrition rates?
○ Have you considered rates of drop-out? Further displacement which impedes
intermediaries from continuing in the program? Changing security conditions?
Loss of interest?
● Please note that the minimum requirements for measuring knowledge improvements in
intermediaries is a pre-post knowledge assessment. This can be supplemented with
qualitative/observational data.
35
3.5 Case Study
36
37
38
39
40
41
42
43
44
45
Part 4: The MEL Plan
46
4.1 Introduction to the Monitoring, Evaluation, and Learning (MEL) Plan
Developing an MEL plan is recommended, but in most cases not required, for
Global Mental Health innovators.
The MEL plan records the what, who, when, how, and why of your monitoring, evaluation, and
learning process. You may construct it as part of an overall guide to
organizational/implementation practice, or as a separate, but connected, document. It builds
out the specifics on your activities around measurement and evaluation. It will go beyond the
requirements of a specific funder to describe the full spectrum of information that you need,
are able to collect, and will use, to support decision making, advocacy, and building partnerships
and networks for future programme delivery. While the MEL plan is not a required submission
for GCC, teams are recommended to think through and develop a MEL plan regardless for your
own processes.
The plan will build around what the MEL is serving - linking the content to the ToC.
Create an outline of what appears important to know in order to track, monitor, and evaluate
progress during your current funding cycle.
47
Highlight what level of detail this information relates to (concept, construct, indicator):
• Who the primary audience is, and
• To what stage the information collected is addressed:
• Preparation/Design
• Implementation (Process)
• Implementation (Impact)
• Reflections for scaling beyond this funding cycle
Executive Summary
Completing this activity helps focus on the main aspects of an MEL plan. What aspects of the
MEL process have you thought about sufficiently and what needs more detail?
As you consider the questions below, we suggest sitting with each question to guide your
thinking and process, and then completing the Executive Summary (Section 4.2)
For each row of the RMAF, talk through the required indicator…
• Purpose: How does it tie in with your programme objectives (which question does it
help answer)?
• Method: How was it defined?
• Management: How you will be collecting it, checking it, using it in decision making?
48
Step 3: Filling in the MEL Plan
• Now fill in the details of the MEL using the application form and process as outlined by
GMH to potential grantees.
We would like to see how your Monitoring, Evaluation, and Learning process is integrated into
the organization and delivery of your implementation, and into your system of decision
making and reporting.
For Sections 1-3 listed below (impact evaluation, process evaluation, and programme
monitoring), please include the following information:
a) List the key questions that will be addressed through your MEL process.
b) Summarize the data collection methods/approaches that will be used and describe:
i. Those data sources that are part of the regular tracking and management that
also serve to track progress and achievement.
ii. If there are any new or innovative methods that you plan to use (for instance,
other ways to engage both participants and other stakeholders to collect data in
different ways, i.e. photography); how will you evaluate this methodology?
iii. How different sources of data will be integrated/interlinked/triangulated.
e) List the RMAF row(s) that describe the relevant concepts, indicators, tools, data and
data collection you will use to address each question.
For Section 4, specify who (internally and externally) will be part of the MEL process. Describe
how they will be part of the information collection process, the analysis, and who will share in
the results- as well as how it is anticipated that they will use them.
You can use the empty boxes below as a space to enter your own thoughts and ideas as you
think through this process.
49
Information to Build Next Steps
What indicators will you use to answer your questions about impact?
How will this data be used (analyzed, disseminated, and applied in decision making)?
50
Section 2: What major questions about PROCESS/DELIVERY do you hope to answer through the
evaluation you will carry out?
What indicators will you use to answer your questions about the process?
How will this data be used (analyzed, disseminated, and applied in decision making)?
Section 3: What questions do you intend to address within the grant period through
MONITORING with a view to taking action or making changes during the course of your
programme to how the innovation is delivered or what is delivered?
What indicators will you use to monitor your programme in real time?
How will this data be used (analyzed, disseminated, and applied in decision making)?
51
The Integrated MEL Process and Data Use in Decision Making
Section 4: Briefly outline the network or people involved in the MEL process, their roles,
responsibilities, and anticipated contributions. Please use a diagram or table if you wish.
52