15
35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning Mercy Corps

35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

Embed Size (px)

Citation preview

Page 1: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

35Measuring Agency-Level Results Part II: Mission Metrics

April 10, 2013

Presented by Barbara WillettDirector of Monitoring, Evaluation and Learning

Mercy Corps

Page 2: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

What is Mission Metrics

An agency initiative to

understand how we – collectively –

are achieving our mission.Mission StatementMercy Corps exists to alleviate suffering,

poverty and oppression by helping people build secure, productive and just

communities.

Page 3: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

3

Page 4: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

Mission Indicators: How They Work

• Program indicators are aligned with Mission Indicators

• Results are reported• Data is reviewed for quality and

methodology• Results aggregated, analyzed,

discussed, presented• Recommendations for action• Monitoring of actions and

learnings

Program Indicators

Mission Indicator

AGENCY RESULTS

Page 5: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

35

Mission Metrics Learning Cycle

Reflect/ DiscussAnalyze

Revi

ew

Resu

lts/L

earn

Make

Decisions

Take Action

Incoming Data

Cycle of Learning

Page 6: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning
Page 7: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning
Page 8: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

10.2-Accountable Governance

10.1-Communities Engaged in Governance9.3-Improved Relationships

9.2-Mechanisms to Manage Conflict9.1-Improved Peace/Security

8.1-Sectors Interacting7.2-Participation of Marginalized

7.1-Improved Capacity - Marginalized6.2-Jobs Created

6.1-Improved Markets

5.2-Partner Capacity5.1-Community Action

4.1-Household Prosperity3.2-Improved Health

3.1-Access to Health2.2-Reduced Vulnerability

2.1-Mechanisms to Manage Disasters1.2-Change in Emergency Condition

1.1-Basic Needs

0 10 20 30 40 50 60 70 80 90 100

Number of Results by Mission Indicator

2012 2011

Page 9: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

1.1 - Emergency Relief

1.2 - Change in Condition

2.1 - Mechanisms to Manage Disasters

2.2 - Reduced Vunerability to Disasters

3.1 - Access to Health

3.2 - Improved Health

4.1 - Household Prosperity

5.1 - Collective Action

5.2 - Partner Capacity

6.1 - Improved Markets

6.2 - New Jobs Created

7.1 - Marginalized Capacity

7.2 - Marginalized Participation

8.1 - Sectors Interacting

9.1 - Improved Peace and Security

9.2 - Mechanisms to Manage Conflict

9.3 - Improved Relationships

10.1 - Communities Engaged

10.2 - Accountable Government

0 20 40 60 80 100 120 140 160 180 200

2012 MM Indicator Results - Reported vs. Aligned

Reported Aligned

Page 10: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

35

10

Emergency Beneficiaries

2010 2011 20120

1,000,000

2,000,000

3,000,000

4,000,000

Water & Sanitation Cash Health Shelter/NFI Food

Page 11: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

35

11

Results - Economic Development

Page 12: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

35

12

Results - Economic Development

Page 13: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

35

Benefits and Strengths - Agency

- Demonstrates commitment to accountability and critical self-

reflection

- Fosters shared language and values between Agency and field

programs

- ‘Real data’ behind the meaningful change we affect

- Means for assessing strategic impact when reviewing and planning

- Highlights gaps we might not have seen otherwise

Page 14: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

35

Benefits and Strengths - Field

- Connection to agency Mission and other program achievements

- Opportunity to review collective results at country and regional

level

- Lens for reflecting on measurement quality and practices

- Platform for standardization and improvement of indicators and

tools

- Higher profile for M&E staff and responsibilities

Page 15: 35 Measuring Agency-Level Results Part II: Mission Metrics April 10, 2013 Presented by Barbara Willett Director of Monitoring, Evaluation and Learning

Lessons Learned

Focus early and often on use – what questions do you want to answer?

Engage everyone you can in the process – all stages

Link closely to program M&E processes - no extra data

Allocate quality IT and data management systems

Give it time – it may take years to develop and fully embed

Limit expectations as you go – can’t be all things to all people

Keep track of the successes and examples of how it is used