1
Alignment of Program Lifecycles and Evaluation Lifecycles 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 Evaluation Lifecycle Program Lifecycle N=50 10/19/09 9 2 1 2 4 1 3 5 4 7 3 4 4 1 Program lifecycles: Programs mature as part of an ongoing trial-and-error learning process, through multiple implementations. Any given program might move forward and backward through and between the lifecycle phases as needed, Not all program survive through all the lifecycle phases, many times a program is discontinued or dramatically revised because it does not show evidence of affecting outcomes. Hebbard, C; Trochim, W., Urban, J.; Archibald, T.; Casillas, W., Cathles, A.; Hargraves, M.; Hertzog, S.; Johnson, M. Cornell Office for Research on Evaluation. Department of Policy Analysis and Management, Cornell University, Ithaca, NY 14853 Programs in different life-cycle phases have different needs from evaluation, thereby implying parallel life-cycle phases for evaluation*. In the ideal, program development and program evaluation are a symbiotic co-evolutionary relationship, where evaluation changes as the program matures, Evaluation systems must be engineered in such a way that each stakeholder group’s incentive to participate in the evaluation is well understood. Phase appropriate evaluation encourages growth and maturation of the program. Phases of Clinical Trials *Evaluation Lifecycles are drawn from traditional medical clinical trials. Phase I clinical trials are small sample exploratory studies that look at early evidence of effectiveness and dosage levels, implementation difficulties and potential side-effects of treatment. May use healthy participants. Phase II trials are somewhat larger trials and use actual patients. The trials test efficacy - examining whether treatment yields measurable changes - and look for relationships among observed outcomes. “Blinded," randomized trials may begin here. About one-third of experimental drugs successfully complete both Phase I and Phase II studies. In Phase III trials large-scale testing provides a more thorough understanding of the effectiveness of the treatment, the benefits and the range of possible adverse reactions. Most (70% to 90%) drugs that enter Phase III studies successfully complete this phase of testing. Phase IV trials are concerned with fidelity of transfer and implementation of treatments in uncontrolled real-world contexts. The trial is used to compare a drug with other drugs already in the market, to monitor a drug's long-term effectiveness and impact on,quality of life; and to examine its cost effectiveness. Researchers gather information to evaluate the overall benefit-risk relationship of the treatment and identify limitations of indicators for use. Work supported by NSF Award # 0814364 Program Phase III: Maturity. Evidence-based programs. Not significantly, changing. Few implementation surprises. Relatively few programs reach this level. Phase III A Consistent change in outcomes. Aware of similar program. Written protocol or process guide. No longer dependent upon particular individuals for implementation. Phase III B Shown to be effective when compared with an alternative program or no program Program Phase IV: Implementation and Dissemination. Adapted for wide implementation Focus on extending the program to new diverse contexts and populations Phase IV A programs Implemented in multiple sites for the purpose of assuring that effectiveness holds in different contexts. Phase IV B programs In wide distribution with fidelity and with comparable effects. Program Phase II: Development. Consistent implementation Empirical evidence of changes in outcomes Maybe standardized procedures or protocols. Phase IIA programs implemented accurately with model Consistent over multiple implementations. Adapted to local context Participant experience relatively Phase IIB programs Formalized written procedures (e.g., preliminary protocol or implementation plan) Program Phase I: Initiation. Conceptualization through piloting. Have procedures that tend to change from one implementation to the next Issues of initiation include: identifying and training program staff, localiziing to context, reacting to the unanticipated problems . etc. Phase I A programs in their initial implementation(s) either as a newly conceived program or as an existing program adapted from another context or from basic research. Phase I B programs new programs that have been revised and are being re-implemented.. Evaluation Phase I: Process and Response Flexible and dynamic Implementation and process assessment Rapid feedback Integrated aspect of the program implementation., Staff debriefings Informal observations Phase IA evaluations Use post-only participant feedback and satisfaction Staff monitoring and observation. Phase IB evaluations Use post-only assessments, Focus on substantive outcomes, Assessment of internal consistency (reliability) of outcome measures. . Evaluation Phase II: Change Co-related assessment of changes in outcomes (e.g., knowledge, skills, attitude, behavior, performance) Design of observational procedures and measures, Assessing the consistency and construct Pre-post differences relationships among different observations. Phase IIA evaluations Use unmatched pretest and posttest of outcomes Assessment of consistency (reliability) and validity of measurement. Phase IIB evaluations Use matched pretest and posttest. Verify reliability and validity of change. Human subjects review and protection (informed consent, anonymity or confidentiality) is increasingly formalized. Evaluation Phase III: Comparison and Control Focuses on the causation and effectiveness. Use comparison groups or variables and statistical controls for adjusting for uncontrolled factors. Realm of experimental and quasi-experimental designs and of more structured and comparative qualitative approaches. Phase III A evaluations Show consistent change in outcomes. Use of design and statistical controls and comparisons (control groups, control variables or statistical controls). Phase III B evaluations Use of experimental or quasi-experimental designs (randomized experiment; regression-discontinuity) for assessing the effectiveness Compare with standard expectation of performance or with outcomes of people who participate in alternative programs or no programs at all. Evaluation Phase IV: Generalizability and Synthesis Concerned with generalizability or external validity. Examine consistency of outcomes across different contexts. Realm of secondary and meta-analysis or of program review approaches. Phase IVA evaluations Multi-site integrated assessments yielding large data sets over multiple waves of program implementation. Phase IVB evaluations Formal assessment across multiple program implementations that enable general assertions over diverse contexts (e.g., meta-analysis). Programs in NSF study Self-report Lifecycle alignment refine the program model, “debug” the procedures, identify barriers to high-quality adoption, and assesses participant response to the program. Phase IIA results tend to be utilized for management and internal accountability of the program, Phase IIB increasingly used for public accountability seek general inferences about the transferability of the program Program stabilizes. Collecting supporting data, strong program.

Alignment of Program Lifecycles and Evaluation LifecyclesPrograms in different life-cycle phases have different needs from evaluation, thereby implying parallel lifecycle phases for

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Alignment of Program Lifecycles and Evaluation LifecyclesPrograms in different life-cycle phases have different needs from evaluation, thereby implying parallel lifecycle phases for

Alignment of Program Lifecycles and Evaluation Lifecycles

0

1

2

3

4

5

6

7

8

9

0 1 2 3 4 5 6 7 8 9

Eval

uati

on L

ifecy

cle

Program Lifecycle

N=5010/19/09

9

2 1

2

4

1

35

4

7

3

4

4

1

Program lifecycles: Programs mature as part of an ongoing trial-and-error learning process, through multiple implementations.Any given program might move forward and backward through and between the lifecycle phases as needed, Not all program survive through all the lifecycle phases, many times a program is discontinued or dramatically revised because it does not show evidence of affecting outcomes.

Hebbard, C; Trochim, W., Urban, J.; Archibald, T.; Casillas, W., Cathles, A.; Hargraves, M.; Hertzog, S.; Johnson, M. Cornell Office for Research on Evaluation. Department of Policy Analysis and Management, Cornell University, Ithaca, NY 14853

Programs in different life-cycle phases have different needs from evaluation, thereby implying parallel life-cycle phases for evaluation*. In the ideal, program development and program evaluation are a symbiotic co-evolutionary relationship, where evaluation changes as the program matures, Evaluation systems must be engineered in such a way that each stakeholder group’s incentive to participate in the evaluation is well understood. Phase appropriate evaluation encourages growth and maturation of the program.

Phases of Clinical Trials*Evaluation Lifecycles are drawn from traditional medical clinical trials.

Phase I clinical trials are small sample exploratory studies that look at early evidence of effectiveness and dosage levels, implementation difficulties and potential side-effects of treatment. May use healthy participants.

Phase II trials are somewhat larger trials and use actual patients. The trials test efficacy - examining whether treatment yields measurable changes - and look for relationships among observed outcomes. “Blinded," randomized trials may begin here. About one-third of experimental drugs successfully complete both Phase I and Phase II studies.

In Phase III trials large-scale testing provides a more thorough understanding of the effectiveness of the treatment, the benefits and the range of possible adverse reactions. Most (70% to 90%) drugs that enter Phase III studies successfully complete this phase of testing.

Phase IV trials are concerned with fidelity of transfer and implementation of treatments in uncontrolled real-world contexts. The trial is used to compare a drug with other drugs already in the market, to monitor a drug's long-term effectiveness and impact on,quality of life; and to examine its cost effectiveness. Researchers gather information to evaluate the overall benefit-risk relationship of the treatment and identify limitations of indicators for use.

Work supported by NSF Award # 0814364

• Program Phase III: Maturity.

• Evidence-based programs. • Not significantly, changing.

• Few implementation surprises. • Relatively few programs reach this

level.

• Phase III A• Consistent change in outcomes.

• Aware of similar program. • Written protocol or process guide.

• No longer dependent upon particular individuals for implementation.

• Phase III B• Shown to be effective when compared

with an alternative program or no program

• Program Phase IV: Implementation and Dissemination.

• Adapted for wide implementation• Focus on extending the program to new

diverse contexts and populations

• Phase IV A programs• Implemented in multiple sites for the

purpose of assuring that effectiveness holds in different contexts.

• Phase IV B programs• In wide distribution with fidelity and with

comparable effects.

• Program Phase II: Development.

• Consistent implementation

• Empirical evidence of changes in outcomes

• Maybe standardized procedures or protocols.

• Phase IIA programs• implemented accurately with model

• Consistent over multiple implementations.• Adapted to local context

• Participant experience relatively

• Phase IIB programs• Formalized written procedures (e.g., preliminary protocol or implementation

plan)

• Program Phase I: Initiation.

• Conceptualization through piloting.

• Have procedures that tend to change from one implementation to the next

• Issues of initiation include: identifying and training program staff, localiziing to context, reacting to the unanticipated problems . etc.

• Phase I A programs• in their initial implementation(s) either as a

newly conceived program or as an existing program adapted from another context or from basic research.

• Phase I B programs• new programs that have been revised and are

being re-implemented..

Evaluation Phase I: Process and ResponseFlexible and dynamic

Implementation and process assessmentRapid feedback

Integrated aspect of the program implementation.,Staff debriefings

Informal observations

Phase IA evaluations

Use post-only participant feedback and satisfaction Staff monitoring and observation.

Phase IB evaluations

Use post-only assessments, Focus on substantive outcomes,

Assessment of internal consistency (reliability) of outcome measures.

.

Evaluation Phase II: Change

Co-related assessment of changes in outcomes (e.g., knowledge, skills, attitude, behavior, performance) Design of observational procedures and measures, Assessing the consistency and construct Pre-post differences relationships among different observations.

Phase IIA evaluations

Use unmatched pretest and posttest of outcomes Assessment of consistency (reliability) and validity of measurement.

Phase IIB evaluations

Use matched pretest and posttest. Verify reliability and validity of change. Human subjects review and protection (informed consent, anonymity or confidentiality) is increasingly formalized.

Evaluation Phase III: Comparison and Control

Focuses on the causation and effectiveness. Use comparison groups or variables and statistical controls for adjusting for uncontrolled factors. Realm of experimental and quasi-experimental designs and of more structured and comparative qualitative approaches.

Phase III A evaluations

Show consistent change in outcomes. Use of design and statistical controls and comparisons (control groups, control variables or statistical controls).

Phase III B evaluations

Use of experimental or quasi-experimental designs (randomized experiment; regression-discontinuity) for assessing the effectiveness Compare with standard expectation of performance or with outcomes of people who participate in alternative programs or no programs at all.

Evaluation Phase IV: Generalizability and Synthesis

Concerned with generalizability or external validity.

Examine consistency of outcomes across different contexts.

Realm of secondary and meta-analysis or of program review approaches.

Phase IVA evaluations

Multi-site integrated assessments yielding large data sets over multiple waves of program implementation.

Phase IVB evaluations

Formal assessment across multiple program implementations that enable general assertions over

diverse contexts (e.g., meta-analysis).

Prog

ram

sin

NSF

stu

dy

Self-report Lifecycle alignment

refine the program model, “debug” the procedures, identify

barriers to high-quality adoption, and assesses participant response to

the program.

Phase IIA results tend to be utilized for

management and internal accountability of the program, Phase

IIB increasingly used for public accountability

seek general inferences about the transferability

of the program

Program stabilizes. Collecting supporting data, strong

program.