1 Facilitators: Lorna McCue and Pam Kinzie, HC Link Guest Presenter: Andrew Taylor, Taylor Newbury...

Preview:

Citation preview

1

Facilitators: Lorna McCue and Pam Kinzie, HC Link

Guest Presenter: Andrew Taylor, Taylor Newbury Consulting

Results-Based Accountability for community organizations and networks

Session 2: Performance Accountabilityfor Programs, Agencies and Service Systems

Web/Teleconference Tips Synchronize your voice with the webconnection

Web/Teleconference Tips

Mute/Un-mute: *1

ON HOLD

Web/Teleconference Tips Private and group chat

Web/Teleconference Tips Size of screen

If the whole page is not displayed an image of the the page will appear for navigation

Disconnected? See instructions Screen saver or Power saver

Contact Us

www.hclinkontario.ca

416-847-1575 Toll-free: 1-855-847-1575Fax: 416-408-4843

info@hclinkontario.ca

Introductions

Who’s Online?

Facilitators:• Pam Kinzie, HC Link• Lorna McCue, HC Link• Andrew Taylor,

Taylor Newbury Consulting

Please indicate your:• Name• Organization

Session 2 Agenda1. Welcome and Introductions

2. Learning Objectives & Agenda Review

3. Recap from Session 1

4. Performance Measures – 4 Quadrants• Choosing Headline Measures• Comparing Performance• Turn the Curve Report: Performance• Performance accountability questions

5. Performance Measures – Real life Examples

6. “Homework” Assignment

7. Q&A

8. Wrap-Up

Learning Objectives

After participating in this webinar you will be able to:

• Define performance accountability;• Describe performance measures in each of the 4

quadrants• Identify how performance accountability may be

useful to your organizations• Take the next steps to find out more about RBA

Results Accountabilityis made up of two parts:

Performance Accountabilityabout the well-being of

CLIENT POPULATIONS

For Programs – Agencies – and Service Systems

Population Accountabilityabout the well-being of

WHOLE POPULATIONSFor Communities – Cities – Counties – States - Nations

“All performance measures that have ever existed for any

program in the history of the universe involve answering two sets of interlocking questions.”

HowMuchdid we do?

( # )

HowWelldid we do it?

( % )

Quantity QualityPerformance Measures

EffortHow hard did we try?

EffectIs anyone better off?

Performance Measures

HowMuch

HowWell

Performance Measures

Effort

Effect

Quantity

How much did we do?

Education

Quality

How well did we do it?

Effe

ct

Ef

fort

Is anyone better off?

How much service did we deliver?

How welldid we deliver it?

How much change / effect did we produce?

What quality of change / effect did we produce?

Quantity

How much did we do?

Education

Quality

How well did we do it?

Effe

ct

Ef

fort

Number ofstudents

Student-teacher ratio

Number ofhigh schoolgraduates

Percent ofhigh schoolgraduates

Is anyone better off?

Quantity

How much did we do?

Education

Quality

How well did we do it?

Effe

ct

Ef

fort

Number ofstudents

Student-teacher ratio

Is anyone better off?Number of 9th graders who graduate on time and enter college or employment after graduation

Percent of 9th graders who graduate on timeand enter college oremployment after graduation

Lay

Definition

All Data Have Two Incarnations

Technical

Definition

HS Graduation Rate % enrolled June 1 who graduate June 15

% enrolled Sept 30 who graduate June 15

% enrolled 9th grade who graduate in 12th grade

RBA Categories Account for All Performance Measures (in the history of the universe)

Quantity Quality

Efficiency, Admin overhead, Unit costStaffing ratios, Staff turnoverStaff morale, Access, Waiting time, Waiting lists, Worker safety

Customer Satisfaction(quality service delivery& customer benefit)

Cost / Benefit ratioReturn on investment

Client results or client outcomes

EffectivenessValue addedProductivity

Benefit value

Product Output Impact

Process Input

Effe

ctEf

fort

Cost

TQM

Effectiveness

Efficiency

Quantity Quality

Efficiency, Admin overhead, Unit costStaffing ratios, Staff turnoverStaff morale, Access, Waiting time, Waiting lists, Worker safety

Customer Satisfaction(quality service delivery& customer benefit)

Cost / Benefit ratioReturn on investment

Client results or client outcomes

EffectivenessValue addedProductivity

Benefit value

Process Input

Effe

ctEf

fort

Cost

TQM

Product Output Impact

RBA Categories Account for All Performance Measures (in the history of the universe)

Quantity Quality

Efficiency, Admin overhead, Unit costStaffing ratios, Staff turnoverStaff morale, Access, Waiting time, Waiting lists, Worker safety

Customer Satisfaction(quality service delivery& customer benefit)

Cost / Benefit ratioReturn on investment

Client results or client outcomes

EffectivenessValue addedProductivity

Benefit value

Process Input

Effe

ctEf

fort

Cost

TQM

1. Did we treat you well?

2. Did we help you with your problems?

*

Product Output Impact

RBA Categories Account for All Performance Measures (in the history of the universe)

* World’s simplest completecustomer satisfaction survey

Not All Performance Measures Are Created Equal

Quantity Quality

Efficiency, Admin overhead, Unit costStaffing ratios, Staff turnoverStaff morale, Access, Waiting time, Waiting lists, Worker safety

Customer Satisfaction(quality service delivery& customer benefit)

Cost / Benefit ratioReturn on investment

Client results or client outcomes

EffectivenessValue addedProductivity

Benefit value

Process Input

Effe

ctEf

fort

Cost

TQM

Product Output Impact

How much did we do?

Not All Performance Measures Are Created Equal

How well did we do it?

Is anyone better off?

LeastImportant

Quantity Quality

E

ffect

E

ffort

MostImportant

Least

Most

AlsoVery Important

How much did we do?

The Matter of Control

How well did we do it?

Is anyone better off?

Quantity Quality

Effe

ct

Effo

rt

LeastControl

PARTNERSHIPS

MostControl

The Matter of Use

1. The first purpose of performance measurement is to improve performance.

2. Avoid the performance measurement equals punishment trap.•Create a healthy organizational environment.

•Start small.

•Build bottom-up and top-down simultaneously.

1. To Ourselves Can we do better than our own history?

2. To Others When it is a fair apples/apples comparison.

3. To Standards When we know what good performance is.

Comparing Performance

2. To Others When it is a fair apples/apples comparison.

3. To Standards When we know what good performance is.

1. To Ourselves First

Can we do better than our own history?

Using a BaselineCHART ON THE WALL

Comparing Performance

1. To Ourselves First Can we do better than our own history?

2. To Others When it is a fair apples/apples comparison.

Reward?

Punish?

3. To Standards When we know what good performance is.

Comparing Performance

1. To Ourselves First Can we do better than our own history?

2. To Others When it is a fair apples/apples comparison.

Comparing Performance

3. To Standards When we know what good performance is.

The Matter of Standards

Quantity

Effe

ctEf

fort1. Quality of Effort Standards aresometimes WELL ESTABLISHED

• Child care staffing ratios• Application processing time• Handicap accessibility• Child abuse response time

2. Quality of Effect Standards arealmost always EXPERIMENTAL

• Hospital recovery rates • Employment placement and

retention rates• Recidivism rates

3. Both require aLEVEL PLAYING FIELD and an ESTABLISHED RECORDof what good performance is.

BUT

AND

Advanced Baseline Display

Your Baseline

Comparison Baseline

Goal (line)Target or Standard

Instead:Count anything better than baseline as progress.

Avoid publicly declaringtargets by year if possible.

Create targets only when they are:FAIR & USEFUL

How much did we do?

Separating the Wheat from the ChaffTypes of Measures Found in Each Quadrant

How well did we do it?

Is anyone better off?

# Clients/customers served

# Activities (by type of activity)

% Common measurese.g. client staff ratio, workload ratio, staffturnover rate, staff morale, % staff fully trained, % clients seen in their own language,worker safety, unit cost

% Skills / Knowledge (e.g. parenting skills)

#

% Attitude / Opinion (e.g. toward drugs)

#

% Behavior (e.g.school attendance)

#

% Circumstance (e.g. working, in stable housing)

#

% Activity-specific measures

e.g. % timely, % clients completing activity, % correct and complete, % meeting standard

Point in Time vs. Point to Point Improvement

How much did we do?

Choosing Headline Measures and the Data Development Agenda

How well did we do it?

Is anyone better off?

Quantity Quality

E

ffect

E

ffort

# Measure 1 ----------------------------

# Measure 2 ----------------------------

# Measure 3 ----------------------------

# Measure 4 ----------------------------

# Measure 5 ----------------------------

# Measure 6 ----------------------------

# Measure 7 ----------------------------

#1 Headline

#2 Headline

#3 Headline

#1 DDA

#2 DDA

#3 DDA% Measure 8 ----------------------------

% Measure 9 -----------------------------

% Measure 10 ---------------------------

% Measure 11 ---------------------------

% Measure 12 ---------------------------

% Measure 13 ---------------------------

% Measure 14 ---------------------------

# Measure 15 ----------------------------

# Measure 16 ----------------------------

# Measure 17 ----------------------------

# Measure 18 ----------------------------

# Measure 19 ----------------------------

# Measure 20 ----------------------------

# Measure 21 ----------------------------

% Measure 15 ----------------------------

% Measure 16 ----------------------------

% Measure 17 ----------------------------

% Measure 18 ----------------------------

% Measure 19 ----------------------------

% Measure 20 ----------------------------

% Measure 21 ----------------------------

Program: _______________Performance Measure

PerformanceMeasureBaseline

Story behind the baseline

Partners

Three Best Ideas – What Works 1. --------------------------- 2. --------------------------- 3. --------- No-cost / low-cost

One Page Turn the Curve Report: Performance

4. --------- Off the Wall

Performance Accountability

1. Who are our customers?2. How can we measure if we are delivering services

well?3. How can we measure if our customers are better off?4. How are we doing on the most important of these

measures?5. Who are the partners that have a role to play in doing

better?6. What works, what could work, to do better?7. What do we propose to do?

for programs, agencies and service systems

Recommended