42
Aid Program Monitoring and Evaluation Standards 1 Aid Program Monitoring and Evaluation Standards Valid from April 2014 to April 2015 Table of Contents Introduction Standard 1: Investment Design (required features for M&E) 5 Standard 2: Initiative Monitoring and Evaluation System 11 Standard 3: Initiative Progress Reporting 18 Standard 4: Terms of Reference for Independent Evaluations 23 Standard 5: Independent Evaluation Plans 29 Standard 6: Independent Evaluation Reports 34 Standard 7: Monitoring Visits 39 Introduction to the M&E Standards Why do we need standards? This is a second paragraph of body text. High quality monitoring and evaluation products are required to ensure that information generated from initiatives has a credible basis and is suitable for use to make important programming decisions and for wider learning. In turn, this is expected to lead to improved accountability, and a more effective aid program delivered efficiently. Monitoring and evaluation products in this situation refer to: Investment design documents Initiative monitoring and evaluation systems (M&E plans and M&E products generated by the initiative management teams) Initiative progress reports Independent evaluation terms of reference Independent evaluation plans Independent evaluation reports Monitoring visit plans and reports

Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 1

Aid Program Monitoring and Evaluation Standards

Valid from April 2014 to April 2015

Table of Contents

Introduction

Standard 1: Investment Design (required features for M&E) 5

Standard 2: Initiative Monitoring and Evaluation System 11

Standard 3: Initiative Progress Reporting 18

Standard 4: Terms of Reference for Independent Evaluations 23

Standard 5: Independent Evaluation Plans 29

Standard 6: Independent Evaluation Reports 34

Standard 7: Monitoring Visits 39

Introduction to the M&E Standards

Why do we need standards?

This is a second paragraph of body text.

High quality monitoring and evaluation products are required to ensure that information

generated from initiatives has a credible basis and is suitable for use to make important

programming decisions and for wider learning. In turn, this is expected to lead to improved

accountability, and a more effective aid program delivered efficiently.

Monitoring and evaluation products in this situation refer to:

› Investment design documents

› Initiative monitoring and evaluation systems (M&E plans and M&E products generated

by the initiative management teams)

› Initiative progress reports

› Independent evaluation terms of reference

› Independent evaluation plans

› Independent evaluation reports

› Monitoring visit plans and reports

Page 2: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 2

Note: In line with DFAT guidance and terminology, this document uses the word initiative to

refer to a commitment of resources and activity of work to achieve defined outcomes. The

term initiative is generally synonymous with investment but does have a particular meaning

in the context of the aid management system. The term ‘program’ defines strategies for

thematic areas, countries, regions and/or global programs.

Over the past four years, the aid program has been reviewing and exploring factors which

account for the quality of these products. The quality of monitoring and evaluation products

has been found to be highly variable which has important implications for the basis on

which programming decisions are made. Some of the factors that have been identified that

explain this situation are that key actors in M&E do not share the same understanding of

basic concepts of M&E, or have a consistent view of what represents a quality M&E product.

This lack of agreement has been observed across the aid program and within the

development and evaluation industry more broadly.

Aid program personnel need guidance on: how to articulate their requirements fairly

consistently to M&E Practitioners and the industry more broadly; how to assess the quality

of the M&E products they receive; assurance that M&E products do meet a reasonable

standard of quality; and how to work with implementation teams and M&E practitioners to

improve the quality of products where necessary.

Equally, the suppliers of M&E products benefit from this clear articulation of what is

required, and the Standards provide a strong basis for the negotiation of the delivery and

resourcing of quality products. It also improves the efficiency of monitoring and evaluation

processes as both the demand (DFAT) and supply sides (M&E provider and implementing

partners) have a clear idea of the expectations before embarking on M&E tasks.

These Standards have been created in a way that allows M&E Practitioners the flexibility to

use whatever evaluation theory and methods they consider appropriate for the task and in

consideration of their individual skill sets. They are not meant to represent a particular

theoretical perspective, but at the same time they do reflect current preferences in the

development sector such as the inclusion of a results-based management orientation.

What is the basis of these standards?

These Standards were developed by the Indonesia Program as part of their Evaluation

Capacity Building Program - an institutional change program to improve the quality and use

of M&E and to integrate evaluative thinking into everyday work. They were integrated into

agency-wide evaluation guidance in 2012 and provide a strong tool for the articulation of

expectations of the quality expected from a range of M&E products in the aid program. They

are updated annually in response to feedback. These changes not only address technical

content, but also clarity in the language and explanations of concepts suitable for

application in a wide range of country contexts. These Standards were also subjected to a

formal peer review in September 2012.

The Standards are based on a careful assessment of aid program staff roles and functions,

and an in-depth assessment of the barriers and facilitating factors that have led to current

quality in evaluation practice. The Standards reflect:

Page 3: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 3

› the tasks carried out by the different levels of program staff in the purchasing and

oversight of M&E systems and deliverables;

› the tasks carried out by contractors and M&E practitioners who design, implement and

produce M&E systems and products; and

› the capacity of staff to absorb new information and perform new tasks within their

already busy work demands.

How Rigidly Should the Standards be Applied?

DFAT recognises that change will be incremental and take time, and that different

monitoring and evaluation systems may require some flexibility. Time will be required for

DFAT personnel to reach a good degree of consistency on how they articulate and provide

feedback on requirements, and implementation partners will need time to address supply

side issues relating to the provision of methodologically trained and experienced evaluators

able to meet the Standards.

Flexibility is required as demand and supply issues are progressively addressed. These

Standards are not meant to represent a checklist of rules that must be adhered to. They are

meant to provide guidance for demand and supply actors as to what a quality product ought

to look like. This document attempts to provide guidance on where flexibility is appropriate

and how to handle this without compromising on rigour. The two key messages around

flexibility are a) the degree of rigour should be proportionate to the importance of the

decisions being made; and b) the basis of findings must be transparent.

What Standards Have Been Developed?

The Standards presented in this series are:

Standard 1: Investment Design (required features for M&E)

Standard 2: Initiative M&E Systems

Standard 3: Initiative Progress Reporting

Standard 4: Terms of Reference for Independent Evaluations

Standard 5: Independent Evaluation Plan (Methodology)

Standard 6: Independent Evaluation Report

Standard 7: Monitoring Visits

What is the Difference Between Guidelines and Standards?

The Standards complement guidance which outline DFAT’s expectations for a given process,

by providing an additional resource to assist in the articulation and measurement of quality.

They are expected to be shared with M&E Practitioners and implementation partners

working with the department.

The Standards should be used with the following Guidelines:

Page 4: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 4

› Investment Design Overview

› Investment Design Quality Standards (IDQS)

› How do I Assess and Report on Aid Quality Checks (formerly Quality at Implementation

reports

› Evaluation Guideline

Training and Practical Support for the Using the Standards

These Standards are expected to be delivered within a broad range of support for change.

Although the Standards are designed to be a relatively stand-alone document, it is useful for

teams to undergo training or be provided with support to apply the Standards appropriately.

Such training and support may be provided by P&Q areas and/ or by the Aid Management

and Performance Branch (MPB) in the Contracting and Aid Management Division (ACD). The

development and distribution of this document is not considered an adequate intervention

to achieve behaviour change. Provision of the document without training and on-going

support would likely result in a rigid checklist application from a limited understanding of

how to apply the Standards most effectively. This would not be conducive to strengthening

the demand-supply side relationships, nor contribute to institutional change.

Page 5: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 5

Standard 1: Investment Design (required features for M&E)

Feature of the Design

Investment Design Plan

1.1 Terms of reference reflect the requirements for the M&E Practitioner/Design Team Leader

1.2 M&E Practitioner/Design Team Leader has demonstrated skills in Program Logic/ Theory of Change

1.3 Design plan allows for sufficient opportunity to gather relevant information to inform the design

1.4 Design plan allows for sufficient opportunity to meaningfully engage local partners in the design process

Investment Design Document

1.5 The goal (beyond the life of the Investment) is aligned with the priority needs of the target population, and is

aligned with relevant Sector Investment Plan (SIP) Objectives and/or MDGs (see IDQS1.a)

1.6 The design document explains the causal mechanism(s) by which the end-of-program outcomes are expected to

contribute to the broader goals (see IDQS1.a)

1.7 The beneficiaries or actors expected to change any behaviours for key outcomes are identified (see IDQS1.c and

3.b)

1.8 The end-of-program outcomes are expressed in terms of performance outcomes where possible rather than

capacity, or open-ended outcomes (see IDQS3.a)

1.9 The end-of-program outcomes are pitched at the correct level for the time, effort and resources applied (see

IDQS4.c)

1.10 Immediate and intermediate outcomes are described for key interventions where possible (see IDQS3.a)

1.11 There is a robust logic and explanation of the causal mechanism(s) that are understood to link the interventions

and end-of-program outcomes (see IDQS3.a)

1.12 The design describes important contextual factors that are likely to influence outcomes (see IDQS2.a)

1.13 The choice of delivery approaches, partnerships and implementation arrangements are appropriate to achieve

end-of-program outcomes (see IDQS2.a)

1.14 The design specifies that the design, implementation and products of the M&E system needs to be undertaken

with reference to internationally agreed standards (see IDQS3.d)

1.15 High-level results are articulated (see IDQS3.d)

1.16 The design articulates the importance of partner government mutual accountability by making provision for the

mutual assessment of progress where appropriate (see IDQS3.d)

1.17 Adequate provisions have been made for design and oversight of M&E systems for the life of the investment by

an adequately qualified M&E Practitioner.

1.18 Important evaluation questions emerging during the design are highlighted

1.19 Adequate resources (personnel and financial) for M&E are included in the design (see IDQS4.c)

Page 6: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 6

Detailed Description for Standards for Design

Note: This standard is a supplement to the Investment Design Quality Standards (IDQS). The

IDQS set out the principal requirements for the content of a completed Investment Design.

These M&E Standards build on the IDQS ensuring the design contains sufficient depth to

ensure quality M&E during implementation.

Investment Design Plan

1.1. Terms of Reference reflect the requirements for the M&E Practitioner/Design Team

Leader

Terms of reference for the M&E Practitioner/designer for the design mission adequately

articulate the required tasks to meet these Standards and the Investment Design Quality

Standards. The terms of reference also indicate the qualifications and experience required

of the individual responsible for the development of the program logic. The team leader

ought to possess design skills, but can be supported by suitable technical specialists. Terms

of reference describe the design approach that will be employed – a facilitated in-house

design, a contracted out design, or a hybrid. See also TORs for an Investment Concept and

Design Process.

1.2. M&E Practitioner/Designer has demonstrated skills in Program Logic/ Theory of

Change

The individual recruited to develop, or facilitate the development of the Program Logic/

Theory of Change has demonstrated experience in Program Logic/ Theory of Change). They

can provide evidence of program logic experience if requested. They need to have the

capacity to apply program logic in a practical way reflecting the principles of program logic,

rather than developing intricate models that cannot be easily understood by stakeholders.

The ability to deal with simple, complicated and complex designs is required. If a member

of DFAT staff has been tasked to design the investment they either can demonstrate the

required experience, or are adequately supported by an individual who possesses the

required experience.

1.3. Design plan allows sufficient opportunity to gather relevant information to inform the

design

Analyses are carried out in advance of the final design decisions to ensure that the design

decisions are based on credible information, that decisions are transparent, and to ensure

the design addresses the right development issues and priorities. Adequate time is provided

to conduct analyses to inform the design. These analyses could include: a) international

definition of key concepts or dimensions of sub-sectoral areas that should be considered; b)

international review of factors commonly inhibiting performance in the relevant area; c)

analyses of supporting or inhibiting factors encountered in the country; d) important

contextual information that may impact on the success of the Investment. Relevant cross-

cutting issues such as gender as well as safeguard issues have been addressed (refer to

current DFAT policies).

1.4. The design plan allows sufficient opportunity to meaningfully engage local partners in

the design process

Page 7: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 7

The design plan allows sufficient opportunity to meaningfully engage individuals - at the

appropriate level of decision-making - in the design process. This is sufficient to maximise

national ownership of the final design. The design plan shows the specific mechanism by

which the national partner can assume joint decision making and leadership over the

design process. For example, through adequate representational work before the final

design process is carried out, and regular participation in key decision-making points

throughout the design process. It also outlines an effective approach to dialogue during the

design process.

Features in the Design Document

Program Logic/ Theory of Change

1.5. The goal (beyond the life of the investment) is aligned with the priority needs of the

target population

The goal is aligned with the priority needs of the target population. The goal for the

investment is directly aligned with the Aid Investment Plan (AIP), the Sector Investment Plan

(SIP), ( the agreement with the partner government (Statement of Commitment, Partnership

for Development, Memorandum of Understanding etc.), local partner development priorities,

and/or the Millennium Development Goals. The goal may be drawn from the level of the

outcomes articulated in the sectoral/delivery strategies that support the Program Strategy

or AIP. The basis for the identification of priority needs is clearly provided.

1.6. The design document explains the causal mechanism(s) by which the end-of-program

outcomes are expected to contribute to the broader goals

The design clearly states how the end-of-program outcomes are linked to the higher order

outcomes (or goals). For example, in an educational program, if the end-of-program

outcome is that teachers demonstrate effective, modern teaching practices (against an

agreed standard), then the design would argue exactly how improvements in teaching

quality are expected to contribute to a higher outcome of increased completion rates for

students. The proposed mechanism is fully described and draws on either research or

experience in the literature, or another credible basis. This basis must be made explicit.

1.7. The beneficiaries or actors expected to change any behaviours for key outcomes (and

related interventions) are identified

Where appropriate, key actors in the investment that are expected to adopt and sustain new

behaviours or practices are clearly identified. The expected changes in behaviour or

practices by the last day of the investment (or beyond) are described for each key actor.

1.8. The end-of-program outcomes are expressed in terms of performance outcomes where

possible rather than capacity, or open-ended outcomes

The end-of-program outcomes are expressed in terms of performance outcomes at either

the individual, work group, organisational or institutional level. It is clear what individuals or

groups will be doing differently on the last day of the investment. Capacity outcomes are not

used (capacity outcomes refer to the potential to change behaviour - clear descriptions of

the actual behaviours to be achieved is required).

Page 8: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 8

Open-ended outcome statements such as “improved capacity in human workforce planning”

or “improved health information systems”, or “increased resources allocated to X” are not

included. The magnitude of the change expected is clearly articulated. If there is a

compelling argument for open-ended outcome statements due to a lack of knowledge then:

a) this was not due to poor analyses during the design phase; and b) there are clear plans

for systematically gathering the required information, and a future date planned for further

development of the outcome statements. Outcomes for relevant cross-cutting themes (such

as gender) and safeguards (such as child protection, environment, displacement and

resettlement and disability) are clearly articulated.

1.9. The end-of-program outcomes are pitched at the correct level for the time, effort and

resources applied

The end-of-program outcomes can reasonably be expected to be achieved in the time

frame, in accordance with the effort exerted, and the degree of resources applied to

interventions. There are provisions in the design to allow key actors to first learn, then

practice and then apply new skills independently in a supervised environment (consolidate)

before the end of the investment.

1.10. Immediate and intermediate outcomes are described for key interventions where

possible

Immediate and intermediate outcomes are described where these outcomes are known or

can be predicted with a reasonable degree of certainty. These are interim steps of

behaviour change that are pre-requisites to the expected end-of-program outcome.

Immediate and intermediate outcomes are provided for the most important interventions.

Important interventions are defined as either those that play a key role in bringing about the

end-of-program outcome, or account for a significant allocation of effort or resources, or

where it is important to ensure the Program Logic/ Theory of Change is fully described.

Note: Where a program logic is complex, it may not be feasible to develop intermediate

outcomes.

1.11. There is a robust logic and explanation of the causal mechanism(s) that are

understood to link the interventions and end-of-program outcomes

The design explains how the end-of-program outcomes are expected to be achieved. This is

based on credible arguments supported by either the literature, or generally accepted

experience of professionals with known expertise in the topic area. Where there are

competing theories or a lack of consensus on how interventions are expected to bring about

the required change, then these are documented and brought into considerations for M&E

during the life of the investment.

1.12. The design describes important contextual factors that are likely to influence

outcomes

Analysis of the important contextual factors likely to support or inhibit achievement of end-

of-program outcomes are identified and addressed in the design. Important inhibiting

factors have been integrated into risk monitoring and management section. This requires

that Standard 1.3 above has been met.

Page 9: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 9

1.13. The design provides a clear rationale for how the delivery approaches, partnerships

and implementation arrangements are appropriate to achieve end-of-program outcomes

Choices such as using partner systems, core contributions, specific purpose investments/

funds, pooled funds, projects, technical assistance, scholarships or debt relief may be

suitable for the end-of-program outcomes described.1 Further standards which must be met

in regards to the delivery approach and implementation arrangements are outlined in the

descriptors for IDQS 4.b

Enabling Quality M&E Systems (features carried over into RFT, Contract or MOU)

1.14. The design specifies that the design, implementation and products of the M&E system

needs to be undertaken with reference to relevant department standards

The design document makes it clear that prevailing department policies and other

standards (such as OECD DAC standards or JCSEE Program Evaluation Standards) are

reflected in the requirements for M&E.

1.15 High-level results are articulated

A small number of high-level results for the investment are identified. These should be at

the level of the higher order end-of-program outcomes and are likely to reflect benefits to

communities rather than only changes in systems performance. Where appropriate, these

are linked to corporate, thematic or program strategies, or performance assessment

framework or to local partner performance assessment frameworks.

1.16. The design articulates the importance of mutual accountability by making provision for

the mutual assessment of progress where appropriate

The design makes provisions for the mutual assessment of progress against the agreed

commitments and end-of-program outcomes. Where appropriate, M&E information is

generated through partner government and/or local systems, but where this is not available

or appropriate, interventions are included that enable local partners to participate in the

generation and utilisation of findings and recommendations relevant for their needs. Mutual

accountability for an investment is highly desirable and is assumed to be a feature of the

M&E system. If, however, this is not considered appropriate then a full discussion of the

rationale for this decision is presented.

Note: If capacity building in M&E is proposed for local partners then this is integrated into

the design as an intervention, and not an add-on to the M&E system or to the TORs for the

M&E Practitioner for the investment.

1Refer to the DFAT Guideline: ‘Choosing approaches and types of aid for working in partner systems’ .

Page 10: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 10

1.17. Adequate provisions have been made for design and oversight of M&E systems for the

life of the investment by an adequately qualified M&E Practitioner.

Resources have been made available for an M&E Practitioner to design and provide

oversight of investment M&E systems for the life of the investment. This requires regular

short-term inputs throughout the entire implementation period. There may be special

circumstances where a full-time M&E Practitioner is required. A single task of designing an

M&E system without on-going technical support is not proposed. Regular visits (at least

twice a year, or proportionate to the scale of the investment) have been resourced. For

larger, or priority investments, a terms of reference for the M&E Practitioner has been

attached to the design document and requested to be a named individual during the

tendering process. Qualification and professional experience requirements for this specialist

reflect the magnitude or sensitivity of the investment. (Note: sample TORs for an M&E

Practitioner are available).

1.18. Important evaluation questions emerging during the design are highlighted

Important evaluation questions that emerged during the design process are highlighted for

later inclusion in the final M&E system. These could include questions to guide

management and implementation decisions identified by key stakeholders, or questions

relating to the mechanisms that link interventions with outcomes that require exploration

during implementation. Evaluation questions have clear relevance to improvement of

investment implementation, or of interest to a wider sectoral, local partner or management

audience.

1.19. Adequate resources (personnel and financial) for M&E are included in the design (see

IDQS4.c)

Adequate budget has been allocated to monitoring and evaluation. For small investments

this is likely to be a higher proportion and a lower proportion for larger investments.

Consider also DFAT-commissioned evaluations, any technical oversight and initiative

manager monitoring visits. This allocation remains intact through the Request for Tender

(RFT) and contracting phase. For pilot or innovative investments, more resources should be

available to increase the robustness of M&E for policy implications.

Page 11: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 11

Standard 2: Initiative Monitoring and Evaluation Systems

Features of the M&E System

Design Phase to Mobilisation

2.1 Adequate resources are allocated for M&E

2.2 Terms of reference for the M&E Practitioner reflect requirements for a quality system

2.3 Adequately qualified M&E Practitioner is recruited

M&E Plan

2.4 There is an adequate basis for the development of the M&E Plan (e.g. Evaluability or Readiness Assessment)

2.5 The M&E Plan provides a summary of the overarching system design including key M&E approaches and

activities

2.6 The M&E Plan reflects current international standards for evaluation practice (e.g. Joint Committee on

Standards for Educational Evaluation – Program Evaluation Standards)

2.7 Goals and End-of-program outcomes are clearly articulated and assessed

2.8 The plan is focused around answering key evaluation questions linked to specific intended uses of the

information

2.9 The quality and reach/coverage of key deliverables are monitored and evaluated

2.10 Relevant aspects of the context are monitored

2.11 Methods are full described for sampling, data collection, management, analysis and processing

2.12 Baselines are constructed where appropriate

2.13 Responsibility is allocated to specific individuals (not organisations) for all M&E activities

2.14 Mutual accountability and joint assessment by local partners is provided for (using partner systems where

appropriate)

2.15 Individuals responsible for implementing the M&E plan have the capacity to do so (time, resources and skills)

2.16 A strategy for the utilization of information is described

2.17 A complete schedule of M&E activities shows when all key M&E activities will be carried out and information

available

2.18 The M&E plan can be easily understood by the non-specialists and key stakeholders

2.19 The M&E plan is resourced and costed

M&E Products

2.20 M&E reports are available upon request and report against planned M&E activities

2.21 Progress reports meet stakeholder needs, report against M&E plan, have a credible basis for claims, and

recommend actions to improve performance.

2.22 M&E reports demonstrate how monitoring and evaluation systems have informed learning, decision-making

and action

Page 12: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 12

Detailed Description of Standards for Initiative M&E Systems

Note: In Indonesia Program, Facilities are treated as a special case for M&E and the overall design should

reflect a suitable approach2.

Design Phase to Mobilisation

2.1. Adequate resources (personnel and financial) are allocated for M&E

Adequate budget has been allocated to monitoring and evaluation. Budget is required for

personnel, and material and financial resources for monitoring and evaluation activities.

This budget also includes considerations of DFAT-commissioned evaluations, any technical

oversight, and initiative manager monitoring visits. For small initiatives this is likely to be a

higher proportion and a lower proportion for larger initiatives. This allocation remains intact

through the RFT and contracting phase. For pilot or innovative initiatives, more resources

should be available to increase the robustness of M&E for policy implications.

2.2. Terms of reference for the M&E Practitioner reflect requirements for a quality system

The Terms of Reference (TOR) clearly articulate the required tasks including tasks for: M&E

system design, oversight of the implementation of the system, and conduct of more

advanced M&E tasks relating to outcome evaluation. The TORs also state the expectation

that the standards of M&E will reflect DFAT guidance and international evaluation practice

standards such as those described in the Joint Committee for Standards of Educational Evaluation –

Program Evaluation Standards. The TORs describe the required standards of qualifications and

experience for the position. For initiatives requiring more robust M&E systems, there is a

requirement for experience in the design and conduct of a suitable range of rigorous

methodologies.

2.3. An adequately qualified M&E Practitioner has been recruited

The M&E Practitioner selected has the required qualifications and experience described in

the TORs.

Monitoring and Evaluation Plan

2.4. There is an adequate basis for planning the M&E system

The M&E Practitioner is provided with the time and resources to conduct an Evaluability

Assessment (or similar exercise). It typically takes 10 days to conduct this assessment

where the program logic does not require major adjustments and is not an extended

process. Where the design has not maintained its relevance, or the program logic of the

2 Additional support may be required to elaborate further on this distinction; Indonesia program provides guidance on this through their P&Q

unit.

Page 13: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 13

design is weak, further time has been allocated. The basis of findings from the EA is made

available to the department if requested. The Evaluability addresses such areas as:

a) Discussions with key stakeholders to confirm a shared interpretation of the

expected long-term and end of program outcomes. Approaches to engage national partners

for mutual accountability are explained.

b) Discussions with key stakeholders to review the program logic and theory of change

and describe the extent to which the program can be evaluated including clarity of

expression of end-of-program outcomes in the documentation;

c) Identification of key evaluation questions and information of interest to

stakeholders;

d) An examination of proposed/potential data sources (including partner systems) to

ensure that data is of sufficient quality, will be collected and analysed as expected, and will

be available within the required reporting cycles;

e) An assessment of the capacity of the implementation team and partners to

participate in the design and/or conduct of M&E activities;

f) A review of the budget/resources available for M&E activities;

g) Identification of reporting requirements for users of the M&E system-generated

information. This includes progress reporting, Aid Quality Check (formerly QAI) Reporting and

Aid Program Performance Reporting (APPRs and APMRs). There should be a clear indication

of how the M&E plan will provide evidence for reporting against the Performance

Assessment Framework.

h) A review of cross-cutting policy areas that will need to be included in the M&E plan;

i) Clear identification of issues and/or constraints that will affect the design of the

M&E plan.

2.5. The M&E Plan provides a summary of the overarching system design including key M&E

approaches and activities

There is a summary of the overall design of the M&E system. This describes the key M&E

approaches and activities. The reader should be able to quickly see the broad design

features, such as the purpose of the system, the primary users of information, the scope of

what will (and will not) be addressed, what the major evaluations are, and a broad

description of the associated methods. The M&E plan does not launch into a detailed

description of the system without this orientation (i.e. this investigatory framework is similar

to the introduction to a methods section in research). Resource implications of the proposed

system are summarised.

2.6. The M&E plan reflects international standards for evaluation

The M&E plan makes it clear that international standards for M&E will be met. Universally

accepted international standards can be found at: http://www.jcsee.org/program-evaluation-

standards, or the OECD-DAC standards are also relevant for the development sector. In

particular, consideration is given to how best to balance the requirements of utility, validity,

Page 14: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 14

feasibility and ethics (these features are considered as a set of guiding principles rather

than the basis for further quality assessment in addition to these Standards). In addition

any prevailing department policies or other standards are reflected in the plan.

2.7. Goals and End-of-Program outcomes are clearly articulated and assessed

Early in the M&E plan, goals that are expected to be achieved beyond the life of the initiative

(the rationale) and end-of-program outcomes are clearly articulated. End-of-program

outcomes are pitched at the right level with regard to the resources, effort and time

available, and the interventions proposed. These outcomes reflect the intent of the initiative

and reflect robust program logic. If the Evaluability Assessment shows that end-of-program

outcomes are not well articulated in the design, then these have been adjusted to be more

appropriate for the initiative intent.

Adequacy of progress toward the goal (if quality data can be collected cost effectively) and

end-of-program outcomes are assessed through the M&E system.

Note: For complex or innovative initiatives, end-of-program outcomes may evolve over the

life of the initiative. Flexibility is required in these cases. This does not mean that end-of-

program outcomes are not articulated, rather that the M&E system is designed to provide

important contextual information or analyses that may provide a strong rationale for

adjustment of the proposed outcomes. For such initiatives end-of-program outcomes are

reviewed annually.

2.8. The plan is focused around answering key evaluation questions linked to specific

intended uses of the information

A small number of high level key evaluation questions are posed that relate to specific

information needs of stakeholders, and these meet the department’s information needs in

relation to the initiative. At a minimum, questions address quality of implementation;

progress toward end-of-program outcomes; the causal mechanisms that link interventions

with intended outcomes; important positive or negative unintended outcomes. Methods are

identified for each of the evaluation questions.

The list of questions should be as focussed as possible to allow primary information users to

make appropriate decisions about the initiative, and future programming in the country

program, or inform corporate decisions where this need has been made explicit by the

department. It is readily apparent how information generated from these questions will be

utilised.

If only indicators are presented (did we achieve our expected intermediate or end-of-

program outcomes), consider whether or not these will provide sufficient information for

initiative improvement (if not, why not – if yes, what were the factors that enabled success,

and what were the factors that would be a pre-requisite for scale up or replication).

2.9. The quality and reach/coverage of key deliverables are monitored and evaluated

The quality of key initiative deliverables is monitored and evaluated. Key initiative

deliverables are defined as those accounting for large effort and/or expenditure, those who

are likely to have a major influence on the extent to which the initiative will meet its

expected outcomes, or those that are commonly employed. If a key deliverable is likely to

Page 15: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 15

have a major influence on a country or sector, then these are also formally assessed for

quality. The coverage and reach is monitored to allow output reporting on beneficiaries

exposed to the initiative. The adequacy of that exposure to the initiative by participants is

assessed where relevant.

2.10 Relevant aspects of the context are monitored

Important indicators or evaluation questions about the context are monitored or evaluated.

Information is gathered on: the progress of the broader national sector or sub-sector in

relation to the initiative’s area of work; any relevant areas in the context which impact on

the relevant needs and priorities of the initiative actors; any factors which are impacting on

the achievement of sustained outcomes or quality, reach and coverage of the initiative

deliverables; extent to which the initiative may be contributing to changes in the sector. Key

risks are being monitored and continuing risk assessed. Only highly relevant issues are

monitored and this choice represents good value for money.

2.11. Methods are full described for sampling, data collection, management, analysis and

processing

Proposed methods are fully described, and the degree of rigour is proportionate to the

importance of the initiative and/or the decisions being made. The evaluation of any pilot

activities intended for significant scale-up is treated as rigorous. Major studies are fully

designed, or provisions made for the future design of major studies by the M&E Practitioner

at a future date. TORs are prepared for major studies that will be designed by others.

Implementation teams are not expected to design evaluation studies. Methods are fully

described for sampling, data collection, data management, and data analysis or processing.

Tools are annexed if appropriate. Where tools have not been developed during the

preparation of the M&E Plan, then the finalisation of tools is identified in the M&E

implementation plan and responsibility allocated to a suitable qualified individual. Where

tools have been annexed, these adequately address the performance indicators or

evaluation questions posed.

Where national or multi-lateral partner M&E systems are proposed to provide information,

then a statement of likely quality and availability is made. This is informed by the

Evaluability Assessment, if undertaken

Note: If 2.11 is being formally appraised, it should be done by a suitably qualified evaluator

with methodological expertise

2.12. Baselines are constructed where appropriate

Baseline data are collected where appropriate. A baseline is not a broad needs assessment,

but tightly focussed on the outcomes intended to be affected by the initiative. A rationale is

provided where a baseline has not been included in the system design.

2.13. Responsibility is allocated to specific individuals (not organisations) for all M&E

activities

For all tasks required to implement the M&E system, the responsible individuals are

identified. This is not naming organisations, but individuals by position title. For larger

Page 16: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 16

studies there will be multiple responsible individuals for design, data collection, analysis,

interpretation and reporting. All major sub-tasks should be allocated.

2.14 Mutual accountability and joint assessment by national partners is provided for (using

partner systems where appropriate)

Mutual accountability by local partners is a feature of the M&E system. Partner systems are

used where appropriate information is available and of acceptable quality. If alternate

systems are proposed, then a suitable rationale is provided. Joint assessments are planned.

Mechanisms for engaging local partners for mutual accountability are explained. If this is

not considered appropriate, a full description of the rationale is provided. Local partner

engagement in the generation and use of information is provided for where appropriate.

2.15. Individuals responsible for implementation of the M&E plan have the capacity to do so

Individuals who are (or will) be recruited to have responsibility for carrying out M&E tasks

have the time, resources and skills to do so. The M&E Practitioner is not expected to

dedicate significant resources to building implementation team or partner capacity for M&E,

rather the system design reflects the resources available. If national partner capacity for

M&E is required to be developed, then this is reflected as an intervention in the design,

rather than as a sub-set of activities in the M&E system.

2.16 A strategy for the utilisation of information is described

Information that is generated by the M&E system is expected to be utilised by primary users

(and perhaps a wider audience) of the information. There is a clear strategy described that

explains: a) how evolving evaluation questions will meet the needs of primary users; b) how

information will be presented in a suitable format to primary users; c) what approaches will

be employed to ensure information is considered during relevant decision making activities.

2.17. A schedule of M&E activities is presented

A full schedule (such as a Gantt chart) is provided that indicates when all key M&E activities

are to be carried out. Reporting is not described as a frequency (e.g. monthly, annually)

rather by date (e.g. the 25th of each month; or 25th June annually). The schedule makes it

explicit when information will be available to primary users and is compatible with national

and/or DFAT reporting cycles.

2.18. The M&E plan can be easily understood by non-specialist and key stakeholders

The M&E plan is written so that the primary users can readily understand the M&E system,

and can easily see when and how their information needs will be met. The M&E plan should

not take more than a couple of hours to fully understand. Orientation sections, detailed

annexes, and sign posts to relevant sections to different primary users are all aids used in

developing an accessible plan.

2.19. The M&E Plan is resourced and costed

The resource and cost schedule includes other costs beyond the person days of the M&E

Practitioner. These may include person days of other people, personnel associated costs,

M&E materials, software and hardware required, any respondent costs, and other resources

such as transport.

Page 17: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 17

M&E Products (see also Standard 3: Initiative Progress Reports)

2.20. M&E reports are available upon request and report against planned M&E activities

All M&E reports described in the M&E plan are available on request. This may be summary

M&E reports to inform progress reports, or reports from specific evaluations or studies. All

major evaluations or studies require a separate evaluation report to be available (though

not necessarily submitted to DFAT).

2.21. Progress reports meet the needs of primary information users and report against the

M&E plan

Progress reports are available and understandable to primary information users. They meet

the department’s expected information needs. They are not too long, and they are written

for managers and non-specialists. Progress reports provide information promised in the

M&E plan, they include a credible basis for claims made, and they include recommended

actions for initiative improvement. The latter include expected dates for management

responses to be completed. Refer to Standard 3: Progress Reports.

2.22. M&E reports demonstrate how monitoring and evaluation systems have informed

learning, decision-making and action

Progressive reports demonstrate how information generated from the M&E system is

informing learning, decision-making and action.

Page 18: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 18

Standard 3: Initiative Progress Reporting

No. Standard

Feature of Progress Report

3.1 There is an executive summary that communicates the key information required for AQC reporting.

3.2 The relevant aspects of the context are adequately described

3.3 There is a reflection on the continuing relevance of the expected end-of-program outcomes in relation to need

3.4 An assessment of the adequacy of progress toward sustained end-of-program outcomes is described

3.5 The quality, reach and coverage of key outputs or deliverables for the reporting period are described

3.6 The adequacy of progress implementing the annual plan is described

3.7 An assessment of the likely adequacy of planned inputs to meet the expected end-of-program outcomes is provided

3.8 The adequacy of progress against the budget is assessed

3.9 Key management or implementation systems are described and their performance assessed

3.10 The report provides balanced and fair reporting of positive and negative issues, achievements and challenges

3.11 For claims of achievement credible supportive evidence is provided

3.12 The report includes lessons learned from monitoring and evaluation that have potentially important implications more

broadly

3.13 Previous and/or proposed management responses or recommendations are summarized

Page 19: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 19

Detailed Description of Standards for Initiative Progress Reporting3

3.1. There is an executive summary that communicates the key information required for

AQC assessment

An executive summary provides easy access to information to inform the AQC assessment,

and meets the department’s information needs. The main focus for this information is

summary information on: the continued relevance of the expected end-of-program

outcomes; progress toward achievement of sustained end-of-program outcomes (including

relevant cross-cutting theme outcomes); quality, reach and coverage of key outputs or

deliverables; important factors (including contextual factors) impacting on adequacy of

outcomes and outputs; implications of key issues to the achievement of outcomes and

outputs, implementation of the annual plan and meeting the budget; and key management

responses proposed. Any issues with the management or implementation systems are

described. The DFAT initiative manager will explain any updated information requirements

for aid program or corporate purposes.

3.2. The relevant aspects of the context are adequately described

The relevant aspects of the context are described. Information is provided on: the progress

of the broader sector or sub-sector in relation to the initiative’s area of work; any relevant

areas in the context which impact on the relevant needs and priorities of the initiative

actors; any factors which are impacting on the achievement of sustained outcomes or

quality, reach and coverage of the initiative deliverables (discussed above under factors);

extent to which the initiative may be contributing to changes in the sector. Findings on key

risks being monitored are presented and continuing risk assessed.

3.3. There is a reflection on the continuing relevance of the expected end-of-program

outcomes in relation to need

The end-of-program outcomes from the original design document are listed. Any revisions to

these end-of-program outcomes that had previously been agreed are explained. A reflection

of the continuing relevance of these current expected end-of-program outcomes in relation

to the needs of the target population is provided with supportive rationale for any proposed

changes.

3.4. An assessment of the adequacy of progress toward sustained end-of-program

outcomes is described

Where relevant, a distinction is made between end-of-program outcomes and those

outcomes that are expected to be sustained beyond the life of the initiative. A firm

3 Further guidance is available in Indonesia, East Timor and Vanuatu programs entitled: “Suggested Report Content for Implementation Partners”

Page 20: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 20

judgement of the adequacy of progress toward these outcomes is described. This also

includes outcomes relating to cross-cutting issues such as gender and disability (see

department policies on the full range of issues). A simple list of immediate and/or

intermediate outcomes or achievements is not sufficient. A judgement of the adequacy of

this progress must be explicit, and explained with reference to appropriate criteria,

standards and evidence.

Where progress has been particularly good, or particularly poor, a full exploration of the

supportive and inhibiting factors that account for this situation is provided.

Particularly where problems or challenges have been identified, there is a full analysis of the

implications of the situation on the likely achievement of end-of-program outcomes. It is

clear from the report the extent to which the department and other stakeholders will need to

track the relevant issues, including whether there is a need to reassess the suitability of

end-of-program outcomes.

If previous management responses have been taken, then their success or failure is

described. Any options for future management responses are fully elaborated. This includes

the identification of what concrete actions stakeholders are required to take. The cost

implications of these responses are discussed.

3.5. The quality, reach and coverage of key outputs or deliverables for the reporting period

are described

Only the key outputs are described. Key outputs are defined as those outputs that are

important to the achievement of expected end-of-program outcomes, or those accounting

for a reasonable proportion of the budget or effort. Long lists of minor outputs are not

required. The adequacy of the geographical coverage of outputs is provided, the adequacy

of the reach or number of outputs or beneficiaries of those outputs enumerated, and the

quality of key outputs is assessed. An assessment of the quality of minor outputs is not

necessarily required.

3.6. The adequacy of progress implementing the annual plan is described

Progress implementing the annual plan on time is described. A firm judgement is made on

the adequacy of this progress. Where there are significant delays, the factors leading to this

have been identified, and the implications to completing the initiative on time or not are

outlined. Management responses are proposed. Where extensions are requested these are

adequately justified, and other alternatives have been considered.

3.7. An assessment of the likely adequacy of planned inputs to meet the expected end-of-

program outcomes is provided

A firm judgement on the adequacy of the planned inputs to meet the expected end-of-

program outcomes is made. The assessment makes clear the assumptions on which this

assessment has been made, particularly potential changes in implementation. Where there

are anticipated shortfalls, the report provides a well-supported argument for any proposals

for additional inputs.

3.8. The adequacy of progress against the budget is assessed

Page 21: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 21

The amount of budget actually expended to date is presented against the planned budget.

The proportion of variation is provided. For any issues identified, a full exploration of the

supportive and inhibiting factors that account for this situation is provided. Particularly

where problems or challenges have been identified, there is a full analysis of the

implications the situation will have on the likelihood of delivering on the annual plan and

budget. It is clear from the report the extent to which DFAT and other stakeholders will need

to track the relevant issues. If previous management responses have been taken, then their

success or failure is described. Any options for future management responses are fully

elaborated. This includes the identification of what stakeholders are required to take what

actions. The cost implications of these responses are discussed.

3.9. Key management or implementation systems are described and their performance

assessed

Key management or implementation systems are described where there are issues to

consider. This could include: a) strategic and annual planning; b) governance and oversight

c) monitoring, evaluation and risk management; d) financial management and value for

money; and e) staffing or human resource systems. Other relevant management systems

are included. For any issues identified, a full exploration of the supportive and inhibiting

factors that account for the situation is provided.

Particularly where problems or challenges have been identified, there is a full analysis of the

implications the situations will have on the successful management of the initiative. It is

clear from the report the extent to which DFAT and other stakeholders will need to track the

relevant issues.

If previous management responses have been taken, then their success or failure is

described. Any options for future management responses are fully elaborated. This includes

the identification of which stakeholders are required to take what actions. The cost

implications of these responses are discussed.

3.10. The report provides balanced and fair reporting of positive and negative issues,

achievements and challenges

The report appears to be a balance between statements of achievement and challenges

and issues. Statements of limited achievements would be supported by a reasonable

discussion of challenges. The report reflects the challenging nature of human development.

3.11. For claims of achievement credible supportive evidence is provided

For claims of achievement (both in terms of achievement of outcomes, and quality of

outputs or deliverables), credible supportive evidence is provided. The basis by which the

claim is made is articulated. There is not an overemphasis on using examples to

demonstrate achievement, rather the emphasis is on how we know that these examples

have been achieved or outputs are of sufficient quality (the basis of the claim).

3.12. The report includes lessons learned from monitoring and evaluation that have

potentially important implications more broadly

Page 22: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 22

The report includes lessons learned with potentially important implications for the initiative;

local partner development strategies; the country or regional program, or DFAT corporately.

Minor, well established or generic development lessons are not included.

3.13 Previous and/or proposed management responses or recommendations are

summarized

The report provides a summary of the important recommendations or management

responses proposed. Any recommendations or management responses from previous

progress reports are discussed in terms of their implementation and effectiveness.

Page 23: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 23

Standard 4: Terms of Reference for Independent Evaluations

No. Standard

Background and Orientation

4.1 A brief orientation to the initiative is provided which includes: total value; time frame; expected end-of-program outcomes; a

short summary of the key approaches employed in the initiative; a brief explanation of the initiative’s expected contribution

to the Program Strategy and local development planning.

4.2 The rationale for the request for the evaluation: including the overall purpose and primary users of the information are

described

4.3 The TOR identifies the key decisions (management, operational and/or policy) which the evaluation is intended to inform

4.4 The key issues that lead to the evaluation questions are described in neutral language

Key Evaluation Questions and Scope

4.5 The key evaluation questions are consistent with the overall purpose and management decisions of the evaluation

4.6 It is clear which questions are considered to be the priority questions for DFAT that must be answered with a good degree

of rigor

4.7 The scope of the questions is suitable for the time and resources available to the team

4.8 Sufficient supporting information is provided about Key Evaluation Questions to guide the development of an appropriate

evaluation plan

Evaluation Process

4.9 A verbal briefing of the key issues and priority information is planned

4.10 Adequate time has been allocated for document review and document appraisal

4.11 There is a requirement for an elaborated evaluation plan – the depth of planning required reflecting the importance of the

review/evaluation questions and management decisions

4.12 The submission date for the evaluation plan allows sufficient time for data collection activities to be scheduled

4.13 Proposed scheduling allows for adequate data collection and analysis to answer Key Evaluation Questions

4.14 A feed-back session to relevant information users are planned together or separately depending on the sensitivity of

findings (Aide Memoire, discussion or presentation)

4.15 There is provision for processing the information collected to enable systematic analysis and interpretation, and the

development of an evidence base

4.16 Adequate time is made available to complete the draft report

4.17 The process for commenting is efficient and allows independence of the evaluation team final report

4.18 Adequate time has been allocated to responding to comments

4.19 The roles and functions of each team member are stated

4.20 The skill sets of the evaluators reflect the priority questions of the evaluation

4.21 The reporting requirements allow DFAT to track progress of the evaluation without distracting the team from carrying out

important evaluation activities

Page 24: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 24

Detailed Description of Standards for Evaluation Terms of Reference

Note: The term evaluation is used in this document to refer to both reviews and evaluations.

Evaluation would normally refer to a piece of work with a higher degree of methodological

rigour usually requiring longer time frames and additional resources.

Background and Orientation to the Evaluation

4.1. A brief orientation to the initiative is provided

As the terms of reference are used to explore with proposed consultants whether or not they

are interested in, or to comment on the proposed evaluation, the orientation must ensure

the TORs are a stand-alone document. Important information includes: the total value; the

time frame; a summary of the expected end-of-program outcomes; a short summary of the

key approaches employed (such as training, technical advisers, secondments, provision of

infrastructure, equipment, and budget support or pooled funding etc.). The context in which

the initiative is situated is described such as the program strategy and/or sector investment

plan that the initiative aims to address as well as the partner government development

plans of relevance. The delivery mechanism is described (contracted, multi-lateral

development partner, NGO) and whether or not the initiative is a project, program or facility.

Any information which can guide the reader in quickly understanding the scope/reach of the

initiative is provided.

4.2. The purpose of the evaluation is described

The TOR clearly identifies the overall purpose(s) and shows which purposes are of most

importance – accountability, initiative improvement, knowledge generation, or

developmental4. This allows the consultant to reflect these priorities in the evaluation plan.

The primary users of the information are identified so that the consultant can collect

relevant information, contribute to deepening an understanding of the findings during the

mission, and prepare an appropriate report. Primary users are identified by title not only

organisation. For example, “DFAT” is made up of senior executive, desk officers, senior

managers and initiative managers. “The Contractor” is made up of head office personnel,

implementation managers and advisers.

Although the audience is a wider group of stakeholders than the primary users, it is clearly

articulated that the report will be published on the DFAT web-site , and there are clear

instructions on how sensitive information is to be communicated.

4 Developmental evaluation is used in highly complex situations, or in programs that are in the early stages of innovation.

See Gamble (2008) A Developmental Evaluation Primer. JW McConnell Family Foundation.

Page 25: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 25

4.3. The TOR identifies the key decisions (management, operational and/or policy) which

the evaluation is intended to inform

Any important management decisions that the primary users are expected to make are

identified and described. Management decisions are more specific than the purpose and

involve decisions such as whether or not to extend an initiative, whether or not to involve a

new partner, whether partner systems are ready for use, or whether to consider a new

modality for a future initiative.

4.4. Key Issues are identified and discussed

Any important issues that have informed the call for, or design, of the evaluation terms of

reference are identified and described. They are described in neutral language and do not

infer an expectation of findings. They are described in sufficient detail to enable the

evaluator to develop the evaluation plan to adequately explore the issues.

Key Evaluation Questions or Scope

4.5. The key evaluation questions are consistent with the overall purpose and management

decisions of the evaluation

Each of the key evaluation questions is clearly related to the stated purpose(s) of the

evaluation (and clearly related to the key management decisions). There are an adequate

range of questions to meet all the stated purposes, and to ensure DFAT’s information needs

are met. There are no additional questions unrelated to the stated purpose. Although the

DAC criteria are an important consideration for the evaluation, these have not been cut and

pasted into the TOR resulting in broad questions of ambiguous scope. There is a single list

of questions in one place in the TOR.

4.6. Priority evaluation questions are identified

Of the full list of questions, the TORs clearly show what the priority questions for the

evaluation are. This will allow the evaluator to make judgments during the evaluation of

what questions must be answered in the final report, and what questions would be

desirable. These priorities are consistent with the overall purpose of the evaluation. Ideally,

only priority questions are posed, but in some cases where stakeholders have generated

numerous questions which they want to keep in the TORs, prioritising these can be a way of

showing the evaluator exactly what are the critical questions.

4.7. The scope of the questions is suitable for the time and resources available for the

evaluation

Typically, a 12-day in-country evaluation can only address four or five broad questions.

During the development of the evaluation plan, these are broken down into a larger number

of sub-questions or information requirements. In addition, for a typical interview with a

respondent without the need for translation, only a small number of topics can be

addressed with depth. Still, this is only possible if both the interviewer is skilled in

questioning techniques and the respondents are relatively articulate and experienced in the

topic areas. In addition to collecting the information, it also needs to be processed,

interpreted and reported on. The questions posed in the TORs reflect this reality.

Page 26: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 26

4.8. Sufficient supporting information is provided about Key Evaluation Questions to guide

the development of an appropriate evaluation plan

Evaluation questions are not broad or vague or open to a wide range of interpretations.

There is clarity in either the Key Evaluation Questions, or the supportive information

provided. The evaluator will be able to break down questions and identify the specific

information requirements. For this to be successful and for the purpose(s) to be met, the

evaluator will need to be able to correctly interpret the expected information from the way

the questions are worded. The Key Evaluation Questions (and supportive information) pose

questions in a way that the evaluator can select suitable methods for the time and

resources available (for example, cause-and-effect questions are difficult to answer in a

short review without access to suitable secondary data sources).

Evaluation Process

Adequate time and resources are required to enable the evaluation to be completed with an

adequate degree of rigour. The following processes are allowed for:

4.9. A verbal briefing of the key issues and priority information is planned

A phone or face-to-face briefing is planned to discuss the back ground, issues and priorities

for the evaluation with the evaluator before the evaluation plan is developed. Sufficient time

must be allocated to allow DFAT and the evaluator to work together to clarify scope, priority

questions and issues, and general approach to methods. This may require more than one

discussion.

4.10. Adequate time has been allocated for document review and document appraisal

Time has been allocated to reviewing initiative documentation (approx. 2 days) as well as

time to appraise any key documents such as strategies or the M&E system (often a day per

document for full appraisal).

4.11. There is a requirement for an elaborated evaluation plan – the depth of planning

required reflecting the importance of the review/evaluation questions and management

decisions

The depth of planning required for an evaluation reflects the importance of the related

decisions that will be made in response to the evaluation. If important decisions are to be

made then more time is allocated. Typically for a DFAT -commissioned evaluation three days

is required to develop an evaluation plan which includes a fully elaborated methodology.

See Standard 5: Evaluation Plan for more details.

4.12 The submission date for the evaluation plan allows sufficient time for data collection

activities to be scheduled

The data collection activities proposed by the evaluation team will be set out in the

evaluation plan. This plan needs to be submitted to the evaluation manager well in advance

of the in-country visit to allow for data collection activities such as interviews and site visits

to be scheduled.

Page 27: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 27

4.13. Proposed scheduling allows for adequate data collection, processing and analysis to

answer Key Evaluation Questions

The proposed schedule in the TOR is not too detailed as this is developed after the

evaluation plan identifies suitable respondents and activities to address the evaluation

questions. There are a sufficient number of days allocated to answer all the evaluation

questions, as well as to work together as a team to process and discuss findings and

identify further requirements as the mission unfolds.

4.14. A feed-back session to relevant information users are planned together or separately

depending on the sensitivity of findings (e.g. Aide Memoire, discussion or presentation)

There is adequate time to provide detail in evaluation findings to allow contestability of

those findings, and feasibility of recommendations with key stakeholders. As the uses of

information may be different for the different primary users, a suitable range of feedback

options are offered.

4.15. There is provision for processing the information collected to enable systematic

analysis and interpretation, and the development of an evidence base

The evaluation team has been given adequate time to process information from interviews,

document reviews and appraisals, observations or other methods to provide a credible

evidence base to support findings. Typically, three days would be required for processing of

data for a 12 day in-country mission that relied strongly on interviews. More complicated

evaluations (or those with an emergent design) would require more time. This is additional

time to actual report writing. Flexibility is balanced with value for money, but final time

frames should be negotiated with the evaluator.

4.16 Adequate time is made available to complete the draft report

The number of days allocated to completing the report reflects: a) the scope of the

evaluation questions; b) the complexity of the issues that have emerged; c) the number of

people contributing to the writing of the report; d) team reviewing and discussions on the

final draft. It is recommended to allow the evaluation team sufficient time to rest after the

mission and to reflect on the mission. For example, ten days allocated to report writing

could require a three week period to deliver. It is also useful to discuss with evaluators

whether or not they expect to be working on other reports and missions during this time.

4.17 The process for commenting is efficient and allows independence of the evaluation

team final report

The process for commenting on the draft report is described and is efficient. Only relevant

individuals are invited to comment, and the focus of their comments is identified. Note that

those invited to comment on the final report would also be invited to comment on the

evaluation plan to ensure their final comments are within the scope and expectations for

the evaluation. The TOR explains that the department will either provide comments in a

consolidated form to the evaluation team, or, allows additional time to respond to a large

number of comments from all stakeholders.

Note: Be aware that where department personnel consolidate comments, there must be

transparency of decisions on what comments to include or remove. It may be necessary to

Page 28: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 28

provide comments from different stakeholders (national partner, implementation team and

DFAT) separately if there are conflicting views.

4.18. Adequate time has been allocated to responding to comments

The time allocated to the evaluation team to respond to comments reflects a) the likely

range of comments generated; and b) the possibility that comments require significant

structural change in the final report.

4.19. The roles and functions of each team member are stated

Although it is the responsibility of the team leader to produce the final report and provide

detailed direction on tasks in the evaluation plan, the TORs show DFAT expectations about

how each team member will contribute. This is especially important with respect to writing

responsibilities. There is clear guidance on the extent to which the department expects

international and national consultants to participate and assume responsibility for particular

tasks. If there is any requirement of the team leader for capacity building of team members,

then adequate time has been allocated to carry this out effectively.

4.20. Skill sets of evaluation team reflect priority questions

Unless there is a compelling reason provided, the team leader is an evaluation expert, not

only technical expert in the relevant sector or thematic area. They are supported by

technical specialist(s) who will focus on technical aspects of the evaluation. The

development of the evaluation plan is allocated to the team leader who will be responsible

for its implementation. The development of the evaluation plan is not allocated to a team

member. The tasks and balance of work of technical advisers reflects the evaluation

questions. For example, if there is a strong focus on gender, or initiative management

systems, then the number of days allocated to technical specialists from other sectoral

areas reflects this focus.

4.21. The reporting requirements allow the department to track progress of the evaluation

without distracting the team from carrying out important evaluation activities

The requirement for reports during the mission provide for a good balance between

monitoring the progress of the evaluation with allowing the team to focus on important

evaluation activities. The evaluation plan is a critical document for DFAT to ensure that the

evaluator has correctly interpreted the TORs and has made suitable plans to conduct the

evaluation to meet the TORs and reasonable standards of rigour. Other reporting

requirements to consider are a) the aide memoire (which should be short and only provide

anticipated key findings and recommendations); and b) the provision of any processed data.

If requiring the provision of processed data, it is important to consider the implications for

preserving the confidentiality of respondents.

Note: A negotiation of the TORs should be encouraged during the contract negotiations.

This allows the team leader to provide professional advice on the feasibility of the TOR in

terms of the scope of questions and the resources applied.

Page 29: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 29

Standard 5: Independent Evaluation Plans

No. Standard

5.1 The evaluation plan is based on a collaborative approach.

5.2 The primary intended users of the evaluation are clearly identified and their evaluation needs are described.

5.3 The purpose and/or objectives of the evaluation are stated.

5.4 A summary is provided to orient the reader to the overall evaluation design.

5.5 Limitations or constraints on the evaluation are described (e.g. time frame; resources; available data; political

sensitivities).

5.6 The Key Evaluation Questions are supplemented by detailed descriptions and/or sub questions

5.7 It is clear which questions are considered to be of higher priority and are expected to provide the most important

information.

5.8 There is sufficient flexibility to be able to address important unexpected issues as they emerge

5.9 The methods to collect data are described for each question (or related questions).

5.10 The proposed data collection methods are appropriate for the questions posed.

5.11 Triangulation of data collection methods is proposed to strengthen the confidence in the findings.

5.12 The sampling strategy is clear and appropriate for the evaluation questions posed.

5.13 The plan describes how data will be processed and analysed

5.14 The plan identifies ethical issues and how they will be addressed

5.15 The process for making judgments is clear

5.16 Approaches to enhance the utilization of findings are outlined (if this has been requested in the terms of reference).

5.17 The evaluation plan provides guidance on scheduling. The final schedule (if attached) reflects adequate time to

answer the posed evaluation questions.

5.18 The allocation of evaluation tasks to team members is clearly described (i.e. data collection, processing and

reporting).

Page 30: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 30

Detailed Description of Standards for Evaluation Plans

Note: The Evaluation plan is developed by the evaluator based on the ToR. It is a negotiated

document between the client and the evaluator and should provide more detail and reflect

final agreements after that negotiation. The evaluation plan should be submitted as early as

possible, to enable scheduling of site visits, interviews and other data collection activities.

The agreed Evaluation Plan ought to provide the basis by which evaluator performance is

assessed.

5.1. The evaluation plan is based on a collaborative approach

The evaluator has consulted DFAT, and the stakeholders identified as important by the

department, to develop the evaluation plan. Consultation may have been in-person, by

phone or by email. Important stakeholders have been given the opportunity to comment on

the evaluation plan before the evaluation commences. Note: This ensures that additional

information will not be requested after the data collection phase is complete.

5.2. Primary intended users of the evaluation are clearly identified

An evaluation cannot meet the needs of all interested stakeholders. Individuals (by title) in

named organisations should be identified as the primary users of the evaluation findings.

These are the people who will be using the information to make judgments and decisions.

Audience is a different concept and often refers to a broader group of people that may be

interested in, or may be affected by any decisions that result from the evaluation.

5.3. The purpose and/or objectives of the evaluation are stated.

These would normally be taken from the terms of reference. The evaluation design restates

these so that the evaluation plan is a stand-alone document.

5.4. A summary is provided to orient the reader

This is an introductory orientation of the overall design of the evaluation. It is short, about

one paragraph in length. For example, it could highlight whether the evaluation is

predominantly exploratory or descriptive, or whether a cause and effect design is proposed,

or whether or not any case studies would feature in the overall design. It would highlight the

major methods for data collection and analysis. This is called the investigatory framework in

research and evaluation terms. The evaluation plan does not go straight into detailed

descriptions of methodology without this general orientation.

5.5. Limitations or constraints for the evaluation are described

The time available for the evaluation has implications for the scope of the evaluation. If a

large number of questions are posed, but the department only wants a cursory look at many

of these, then a shorter time frame may be appropriate. The evaluator highlights any

important limitations in terms of time available, resources applied, or the expertise of the

evaluation team to deliver a credible, defensible evaluation product. Political sensitivities

are highlighted where appropriate. The implications of these limitations are discussed.

Note: A long list of limitations is not considered a substitute for a poorly negotiated TOR.

Page 31: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 31

5.6. The Key Evaluation Questions are supplemented by detailed descriptions and/or sub

questions

Although the terms of reference is where the department communicates what the

evaluation is to address, the evaluator will still need to give careful consideration to how

these larger questions will be addressed. This means that more detailed information

requirements and/or sub-questions are generated. Commonly, questions presented in a

terms of reference are broad, therefore this more detailed information allows information

users to know how the evaluator has interpreted the broader questions, and whether or not

the evaluation will generate sufficient information to meet these broader questions. It also

allows the DFAT evaluation manager to see the implications of the scope of the evaluation

described in the terms of reference. This breakdown of information requirements or

questions allows the reader to assess whether or not the original scope was realistic. The

evaluation manager needs to pay careful attention to this aspect of the evaluation plan.

5.7. High priority questions are identified

Initiative or programevaluations often have a very large number of evaluation questions that

cover a very wide number of aspects of the initiative to be evaluated. Some of these

questions will be more important than others. The evaluation plan reflects where the

emphasis will be placed, and it is clear that the department’s information needs will be met.

The evaluation team will not usually be able to answer all the questions listed for all

respondents and so will need to make decisions during interviews about what will be

dropped and what is essential. The evaluation manager needs to be confident that the

evaluator will, at a minimum, deliver information on the priority questions.

5.8. There is sufficient flexibility to be able to address important unexpected issues as they

emerge

This flexibility may be built in to the questioning technique employed during an interview. It

may be built into the schedule as a whole to allow new issues to emerge and be responded

to through additional data collection if they are important. Where new issues cannot be

adequately addressed within the schedule, there are processes to review possible trade-offs

to allow them to be addressed.

5.9. Methods for each evaluation questions are described

The evaluation plan shows how each of the evaluation questions will be answered by

describing the methods that will be used to collect the information. For most initiative or

program evaluations this is likely to include in-depth interviews, focus group

discussions/interviews, document reviews and in some cases observations of activities.

Large workshops are not usually a suitable method to gather substantive, reliable and valid

information – however, they may have other important political purposes.

For several questions there may be a number of data collection methods proposed to

strengthen confidence in the findings.

The design of major evaluation activities/studies should be annexed and include tools such

as interview guides or questionnaires. In some cases the evaluator will need to develop

these later, or adjust them as the evaluation proceeds, but there is an absolute expectation

Page 32: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 32

that the evaluator uses tools to guide each evaluation activity, and do not rely on memory of

all the evaluation questions identified in the evaluation plan. Where team members are

working in different locations then tools are available ahead of those evaluation activities so

that data is collected systematically. If flexibility on this is required, then a compelling

rationale is provided. Summary statements of methods that are not linked with specific

evaluation questions are not considered adequate.

5.10. Methods are appropriate for the evaluation questions posed

Although this takes evaluation expertise, it is still worth reviewing the questions posed and

consider if the methods described could reasonably answer the questions. For example, a

focus group discussion would be most unlikely to answer a sensitive question; a review of a

program strategy document (such as gender) would be unlikely to tell you if the initiative’s

actual gender activities were of a high quality. It would need to be supported by information

from other sources.

5.11. Triangulation of methods is proposed

Triangulation is the use of a range of methods and/or sources of information to come to a

conclusion or result. It can develop greater confidence in a finding. Given the short time

frame of most DFAT evaluations or reviews, it is difficult to employ a wide range of methods.

To deal with this, the evaluation has planned to discuss similar questions across a range of

different respondents within and across different organisations, or use a number of

methods to examine the same issue. It is not sufficient to state that triangulation will be

used if this is not demonstrated in the evaluation design.

5.12. Sampling strategy is clear and appropriate

Most evaluations will require some sort of sampling strategy across individuals, sites or time

periods. Appropriate sampling strategies are chosen and justified. For short reviews that rely

on analytical rather than statistical inference, purposeful sampling will be appropriate and

could include maximum variation, a critical case, or a typical case. Efforts should be made

to avoid relying on a convenience sample which is likely to be unrepresentative of the

population of interest. Where statistical inference will be used to generalize from the

sample, random sampling strategies are appropriate – especially stratified random

sampling which reduces the sample size required.

5.13. The plan describes how data will be processed and analysed

The evaluation plan describes how the data will be processed, including measures to check

and correct any errors in data, ensure security of storage and prepare for analysis. The plan

also describes how the data will be analysed in order to answer the Key Evaluation

Questions. This may not necessarily require advanced analytical methods, but users of the

information can determine exactly what is to be done.

5.14. The plan identifies ethical issues and how they will be addressed

For most of the aid program evaluations and reviews conducted by DFAT, this will mostly be

around privacy and confidentiality issues. The plan identifies how these will be addressed

when data are collected, stored and reported. In particular, assurances about anonymity

must be honoured and data stored and reported in ways that do not inadvertently identify

Page 33: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 33

informants, including when providing a database of the evidentiary basis to DFAT as part of

the deliverables. Other relevant ethical issues are addressed including processes for

reporting serious issues if identified during data collection.

5.15. The process for making judgments is clear

The evaluation plan makes it clear that the department requires the evaluator to make a

professional judgement based on the evidence gathered and the agreed basis by which

judgements are made (such as criteria or standards). The department’s response to the

evaluator’s judgement should be provided in the Management Response to an evaluation.

In some exceptional cases, the department may require an evaluator to report neutrally on

facts and leave DFAT to make the final judgements, in which case the plan should make it

clear how evaluative judgements will be made and by whom, as this is an important

distinction and can affect the way information is collected and presented.

5.16. Approaches to enhance utilization of findings are outlined

The importance of utilization of findings needs to be communicated to the evaluator. There

are a variety of well-tested approaches to utilization that a professional evaluator will be

familiar with (e.g. stakeholder engagement strategies for evaluation design or developing

acceptance of recommendations before the report is published). Approaches to utilisation of

findings are outlined in the evaluation plan. Utilization begins with the evaluation design

stage.

5.17. Scheduling guidance is provided

The schedule is developed by the department after the evaluation plan is submitted, and

reflects guidance from the evaluator. The most common problem is that the persons

recruited for interview are not always the best respondents for the evaluation questions

posed. Often there are many donor meetings where respondents cannot provide

substantive comment on many of the evaluation questions. Also consider the time for each

interview with the associated evaluation questions. Most 60 minute interviews with a

respondent cover no more than four or five key topics; less if translation is required.

Sufficient time is available to meet with the implementation team. As part of reviewing the

methodology DFAT negotiates the proposed list of respondents before final scheduling. The

evaluator scheduling guidance is realistic for the time frame. Sufficient time is allocated to

other methods proposed. There is sufficient time allocated to evaluation team discussions

and early data processing (not late at night).

5.18. Evaluation tasks are allocated to team members

It is very important that each team member knows before the evaluation begins what they

will be expected to do. It is not appropriate for the team leader to allocate reporting

responsibilities on the last day of the in-country mission. The evaluation plan shows what

responsibilities each team member has so they can ensure that adequate data is collected,

processed, and interpreted and they can meet a high standard during the reporting stage. It

is often useful to show which evaluation questions each team member carries responsibility

for.

Page 34: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 34

Standard 6: Independent Evaluation Reports

No. Standard

Introductions

6.1 A background to the evaluation summarizes: the total value of the initiative; the number of years of the initiative; the

stage of initiative implementation; key outcomes of the initiative; and the key issues identified in the terms of

reference

6.2 A brief summary of the methodology employed is provided

6.3 Key limitations of the methodology are described and any relevant guidance provided to enable appropriate

interpretation of the findings

6.4 The executive summary provides all the necessary information to enable primary users to make good quality

decisions.

Findings and Analysis

6.5 The evaluation report clearly addresses all questions in the Terms of Reference

6.6 The relative importance of the issues communicated is clear to the reader

6.7 There is a good balance between operational and strategic issues

6.8 The report clearly explains the extent to which the evidence supports the conclusions and judgments made

6.9 Alternative points of view are presented and considered where appropriate

6.10 Complicated and complex aspects of issues are adequately explored and not oversimplified

6.11 The role of context in initiative performance is analysed

6.12 The text uses appropriate methods/language to convince the reader of the findings and conclusions

6.13 There is an adequate exploration of the factors that have influenced the issues identified and conclusions drawn

6.14 The implications of key findings are fully explored

6.15 The overall position of the author is clear and their professional judgments are unambiguous.

Conclusions and Recommendations

6.16 The conclusions and recommendations logically flow from the presentation of findings and any associated analyses.

6.17 Individuals have been allocated responsibility for responding to recommendations.

6.18 Where there are significant cost implications of recommendations, these have been estimated (financial, human

and materials costs).

6.19 The recommendations are feasible

Page 35: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 35

Detailed Description of Standards for Evaluation Reports

Introductions

6.1. The background provides adequate information for individuals not familiar with the

initiative

The background provides adequate information to enable individuals not fully familiar with

the initiative to interpret the report. It summarizes: the total value of the initiative; the

number of years of the initiative; the stage of initiative implementation; the delivery

mechanism; key expected outcomes of the initiative; and the key issues identified in the

terms of reference.

6.2. A brief summary of the methodology employed is provided

Although a fully elaborated methodology was developed before the evaluation, a summary

of the significant details is included. Sufficient information is required to enable the reader

to quickly understand the evidentiary basis of the evaluation. The evidentiary base must be

convincing and in proportion to the resources invested in the evaluation. The full

methodology is annexed. Important aspects of the strategy to ensure findings are utilised

are summarised here.

6.3. Key limitations of the methodology are described and any relevant guidance provided

to enable appropriate interpretation of the findings

Key limitations are summarised in the evaluation report to enable the reader to make

appropriate decisions. Where necessary the author has provided specific guidance of where

the reader ought to be cautious about the findings.

6.4. The executive summary provides all the necessary information to enable primary users

to make good quality decisions.

The executive summary provides all the necessary information to enable primary

stakeholders, especially senior management to make good quality decisions without reading

the entire document. It is not a simple cut and paste of the main body of the report. It

summarises the key findings, provides sufficient analyses and arguments, and presents

final conclusions and recommendations. Resource implications of recommendations are

summarised. The length of the executive summary is proportionate to the length of the

report (e.g. two to three pages for short uncomplicated reports, and up to five or six pages

for more lengthy reports with complex issues).

Findings and Analyses

6.5. The evaluation report clearly addresses all questions in the Terms of

Reference/Evaluation Plan

Note: As the Evaluation Plan supersedes the Evaluation Terms of Reference, the Plan is the

appropriate document to assess whether or not the evaluation has delivered on

expectations. In the absence of an Evaluation Plan, the Terms of Reference should be used.

Page 36: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 36

It is relatively easy to identify where each of the questions in the Evaluation Plan are

addressed. The report does not need to be a mechanical presentation of these questions,

but it should be relatively easy to negotiate the report and find relevant information about

specific questions in the Evaluation Plan. Where there are gaps, these have been explained.

The department’s information needs, as set out in the Terms of Reference and Evaluation

Plan, have been met.

6.6. The relative importance of the issues communicated is clear to the reader

The report makes it clear what issues are priority issues to consider. Minor issues are not

set out mechanically against the terms of reference and given the same depth of treatment

as more important issues. The breadth of description, depth of analysis and attention in the

recommendations can indicate the degree of priority. The author may simply state the

relative importance of issues.

6.7. There is a good balance between operational and strategic issues

The report addresses the full range of issues identified in response to the TOR and other

critical issues that have emerged. There will be technical, managerial or operational issues

that are very important to consider and are often at the core of many important challenges.

The strategic direction or any higher order issues of the initiative have been given adequate

space, and minor technical issues are treated in a more limited fashion. Flexibility is

required where the TOR evaluation questions demonstrate that this balance was not

required.

6.8. The report clearly explains the extent to which the evidence supports the conclusions

and judgments made

For key findings, the basis of the findings and related conclusions is communicated clearly.

This includes reporting the degree to which views are shared across respondents. The

information is brought together from a range of sources, but communicated as a coherent

whole. Evaluator opinions that are based on limited evidence are proposed as suggestive

only.

6.9. Alternative points of view are presented and considered where appropriate

Alternative views must be presented, especially for important, controversial or disappointing

findings. They are not immediately dismissed, but are seriously considered. Key stakeholder

views such as those of the implementation team must be given sufficient attention, and

balanced by national partners, the department or other important stakeholder views.

6.10. Complicated and complex aspects of issues are adequately explored and not

oversimplified

The report adequately acknowledges complicated aspects of issues, such as multiple

contributing factors, or emergent challenges and opportunities. The report does not present

simple solutions to these types of situations. The findings are presented fairly so that

specific stakeholders are not held fully accountable for problems when multiple factors are

involved. Human development is challenging, and the report recognises that

implementation teams and national partners are often facing multiple challenges.

Page 37: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 37

6.11. The role of context in initiative performance is analysed

The report identifies relevant aspects of the context within which activities are implemented.

These might include geographic, cultural, political, economic or social context. Sufficient

information is presented to allow the reader to understand the relationship between the

initiative and its context. The report addresses: a) how the context may have affected the

achievement of outcomes (both supportive and inhibiting); and b) the extent to which the

initiative may have had any effect on the context.

6.12. The text uses appropriate methods/language to convince the reader of the findings

and conclusions

Arguments presented do not use emotive word choices in an effort to appeal to the

emotions of the reader. The method used to convince readers is the presentation of

evidence or a credible basis for the finding. Using the international literature to build the

credibility of the report can be effective. The report handles political issues with sensitivity. A

good report considers the expected positions of the important stakeholders – if findings are

unexpected then this is carefully communicated and explained in the text.

6.13. There is an adequate exploration of the factors that have influenced the issues

identified and conclusions drawn

It is not sufficient to simply describe a situation. A full analysis of the likely factors that have

led to the situation is necessary. Factors that enable progress or achievement are just as

important as factors that inhibit them. These factors should be generated from a range of

data sources. A range of causes should be considered rather than regularly offering a single

cause for major and/or complex issues.

6.14. The implications of key findings are fully explored

Aid initiative managers, senior management and other stakeholders need some direction on

the implications of the findings if this is not immediately apparent. Implications to achieving

initiative objectives, implementation for meeting time frames, expenditure projections, or

sustainability are often important considerations.

Conclusions and Recommendations

6.15. The overall position of the author is clear and their professional judgments are

unambiguous.

The task of the evaluator is to evaluate. They must make their position clear (and as early as

possible in the report) unless the TORs have required the evaluator to report on findings

with neutrality. The report does not simply state findings and expect the department to

interpret them and draw their own conclusions. The report presents the authors view

unambiguously. Has the initiative made adequate progress or not? Are the factors that have

accounted for the limited achievements been unavoidable or are they due to poor

management. Unambiguous judgements also present findings and conclusions sensitively

and constructively.

6.16. The conclusions and recommendations logically flow from the presentation of findings

and any associated analyses.

Page 38: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 38

It is possible to trace issues through the text from description, to analysis, to conclusion and

recommendation. No recommendation appears at the end that is not supported by

descriptive and analytical work in the text. There are no important inferred

recommendations buried in the text that have not been drawn into the conclusion or list of

recommendations at the end.

The “chain of evidence” is evident. This is where all questions in the methodology have data

that has been collected, analysis conducted, findings presented, interpretation carried out

and reported. If questions in the methodology have not been addressed then an explanation

has been given.

6.17. Individuals have been allocated responsibility for responding to recommendations.

Where appropriate, job titles, rather than organisations, have been allocated responsibility

for all recommendations for action. If it is not appropriate or possible to identify the

individual, then the relevant work group is identified. If some recommendations are for

broader partner government, or sectoral or corporate learning within the department then

these are identified separately.

6.18. Significant cost implications of recommendations have been estimated

If recommendations imply human, financial or material costs, these are estimated. If

recommendations for additional technical support are made, then the number of days input

is estimated. For important technical assistance positions proposed, the key content to

consider for the terms of reference is annexed.

6.19 The recommendations are feasible

Recommendations, in the most part, are acceptable to relevant stakeholders

(recommendations that stakeholders do not agree with rarely get implemented – coming to

acceptability is dealt with by the utilisation strategy). Recommendations are feasible from a

resourcing and cost perspective. Recommendations are likely to be effective to rectify a

situation, or to achieve an expected outcome.

Page 39: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 39

Standard 7: Monitoring Visits

Note: A Monitoring Visit is a visit conducted by DFAT personnel on an initiative in their

portfolio. It can also be conducted jointly with implementation partners. There is no

department-wide approach to how monitoring visits should be undertaken – this should be

determined in accordance with the context and information needs of the initiative manager.

This standard provides one possible format only for monitoring visits.

No. Standards

The Monitoring Visit Plan

7.1 The number of the monitoring visit for the initiative is recorded

7.2 The broad purpose of the visit is described

7.3 The key questions for the visit are listed and are related to the purpose of the visit

7.4 The scope of the questions is suitable for the time and skills available for the visit

7.5 How information will be collected is described (interviews, observations, informal interaction, reading documents)

7.6 The proposed ways of collecting information are suitable for the questions posed

7.7 A proposed schedule of expected activities and persons to meet is prepared. The date sent to host is recorded.

7.8 The proposed schedule allows sufficient time to address the visit questions

7.9 Roles and responsibilities for each team member are clearly described (including joint missions)

Recording and Managing Information

7.10 Key points or issues are recorded after each activity in line with visit questions (content and impressions of the monitor)

Visit Report and Response

7.11 A background to the visit is presented which describes: a) the purpose of the visit; b) the visit number; c) the visit

questions; d) the dates of the visit; e) the participating visit members (DFAT and joint visit with other partners); e) the cost

(expenditure and person days)

7.12 A brief summary of findings for each visit question posed is provided. The four analytical questions are integrated into the

questions where appropriate

7.13 Important lessons learned or insights of general interest to the sector, the program, or the department corporately are

listed

7.14 Information relevant to updating the AQC is presented in language suitable for pasting into an AQC or SAQC report

7.15 Management responses are summarized according to who is required to take responsibility for the action

(Implementation Partner, Initiative Manager, other relevant Managers/ Counselors). Management responses are given a

time frame for action, steps required, and resource implications highlighted where relevant.

7.16 Feedback to respondents or participants is provided where appropriate

7.17 The date the report is sent to the management team for oversight of management responses is recorded.

7.18 Follow Up section is completed as each management response is completed. Report is closed when last management

response is taken. If a management response will not be taken, then the reason is recorded.

Page 40: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 40

Detailed Description of Standards for Monitoring Visits

Note: If several department staff travel together on a Monitoring Visit, it is only necessary to

compile one, joint monitoring plan and report. In such cases, one person will be nominated

as the visit leader and they will prepare the plan and write the report. For joint missions with

other implementation partners, the plan and report are developed together to ensure all

information needs are addressed.

The Visit Plan

The visit plan is sent to relevant management staff, on completion.

7.1. The number of the monitoring visit is recorded

Over the life of each initiative there will be a number of visits. The visit plan shows what

number the visit is, and the number of months the initiative has been under

implementation.

7.2. The broad purpose of the visit is described

Monitoring visits are not expected to meet some list of pre-determined questions or

purposes. In the most part, they are useful to fill in gaps in knowledge about the initiative

that are important for AQCs and general management and decision making. They can also

be used to learn more about the context in which the initiative is operating. Therefore the

broad purpose can address a range of relevant information needs.

7.3. The key questions are listed and are related to the purpose

Like in an evaluation, the visit is guided by key questions that need to be answered. It is

clear how the information generated from these questions will be used, and that this is

consistent with the purpose.

7.4. The scope of the questions is suitable for the time and skills available for the visit

Developing an in-depth understanding of a few focused issues will be more advantageous

than trying to cover a wide range of issues poorly. The scope of the questions is quite

focused on information needs of DFAT and others involved in the visit. The questions do not

require complex methods to collect and process the information. The questions are simple

questions that a generalist using good common sense has a reasonable expectation of

answering.

7.5. How the information will be collected is described

The visit plan shows how the information will be collected for each question. Who will be

interviewed, what will be observed, what documents will be reviewed, and where informal

interactions are expected. For interviews, short interview guides are developed; for

observations, what will be observed is briefly described; for documents, what will be read

will be described. It is not a lengthy description, but provides sufficient guidance to ensure

that the visit will deliver the expected information without major gaps.

7.6. The proposed ways of collecting the information are suitable for the questions posed

Page 41: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 41

Consideration is given to the best way of collecting information to answer the questions

posed. For example, it may be more effective to observe training, or review a training

agenda to discover the quality of training, than ask the participants if it was good quality.

However, interviewing participants or their supervisors is more effective if you want to learn

about the application of training.

7.7. The proposed schedule of activities and/or meetings is presented

Once the ways of collecting information are fully described, a schedule of activities (such as

observations of initiative activities in progress) or meetings is developed. The date is

recorded when this was sent to hosts (either government or implementation partners).

7.8. The proposed schedule allows sufficient time to address the visit questions

The schedule allows sufficient time to answer the questions to a reasonable degree of

depth. The schedule reflects the scope of the visit.

7.9. Roles and responsibilities of visit team members are described (including for joint

missions)

The designations and roles and responsibilities of the different team members are clear. It

is poor practice to allocate roles and responsibilities during the mission. For example, if an

initiative manager is expected to act as a translator, then they cannot also collect

information, take notes and be able to process information after the meeting. Whoever is

responsible for writing the report from the visit needs to be able to control the information

that is collected so that they can answer the questions posed in the visit plan.

Recording and Managing Information

7.10. Key points are recorded after each visit activity

The DFAT staff member responsible for the interview(s) or other information collection

activity ensures that there are dot points recorded against the key questions after each

activity. These recorded notes provide the basis for the final report. Dot points address: a)

the visit questions in terms of actual content provided by respondents, or observed; and b)

the impressions of the DFAT staff member.

Visit Report

7.11. A background to the visit is presented

This describes: a) the purpose of the visit; b) the number of the monitoring visit for the

initiative; c) the visit questions; d) the dates of the visit; e) participating visit team members

(this could be DFAT and joint visits with other partners); f) the cost (person days and

expenditure). If there are any important issues that led to the visit, these should be very

briefly summarized.

7.12. A brief summary of findings is provided

A summary is provided for each question. It is as brief as possible as these reports may be

read by a wide range of colleagues. The four analytical questions are reflected in the text

where relevant: a) what is the current situation; b) what are the factors that have led to the

situation; c) what are the implication to initiative success; d) what are the responses that

have been taken thus far (and their success), and what future responses are still required.

Page 42: Aid Program Monitoring and Evaluation Standards · 2017. 2. 22. · Aid Program Monitoring and Evaluation Standards 3 › the tasks carried out by the different levels of program

Aid Program Monitoring and Evaluation Standards 42

7.13. Important lessons learned or insights are listed

The report includes lessons learned with potentially important implications for the initiative;

local partner development strategies; the sector investment plan; the DFAT aid program, or

DFAT corporately. Minor, well established or generic development lessons are not included.

This section can be left blank if there is nothing of note.

7.14. Important information for AQC assessment is recorded

This information is presented in very brief language in a format that can be cut and pasted

into an AQCor SAQC assessment. This information is arranged in accordance with AQC

heading.

7.15. Management responses required are summarised

All the proposed management responses are listed according to who is required to take

responsibility for the recommended action (e.g. Implementation Partner, initiative manager

and relevant management staff, Counsellor, and/or local partner). Each management

response has a time frame for action, a description of the steps required, and whether or

not it has been negotiated with partners if they are responsible. Resource implications are

highlighted where relevant.

7.16 Feedback to respondents or participants is provided where appropriate

Any feedback provided to participants in the monitoring visit is recorded. This feedback can

be provided to implementation teams, local partners or beneficiaries where appropriate

7.17. The date the report is sent to relevant management staff is recorded

Management staff will need to be aware of the management responses that are generated

from each initiative they are responsible for. They are responsible for supervising the follow-

up required from each visit.

7.18. The follow up section is completed

All management responses are expected to be followed up, or a reason recorded for non-

follow up. As each management response is completed, the date is recorded. When all the

management responses for a visit have been completed, then the visit report is “closed”

and the updated version forwarded to relevant management staff. If a management

response is not taken then this is explained before closing the report.

Feedback If you have any comments or feedback about this Tool, or related business processes

information, please contact [email protected]. A log of all feedback received will be

kept and used to feed into a process of annual review.