14
RESEARCH AND EVALUATION © LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON

Chapter3

Embed Size (px)

DESCRIPTION

Research Methods in Education 6th Edition

Citation preview

Page 1: Chapter3

RESEARCH AND EVALUATION

© LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON

Page 2: Chapter3

STRUCTURE OF THE CHAPTER

• Research and evaluation: similarities and differences

• Research, politics and policy making

Page 3: Chapter3

DEFINING EVALUATION

The provision of information about specified issues upon which judgements are based and from which decisions for action are taken.

Page 4: Chapter3

SIMILARITIES BETWEEN EVALUATION AND RESEARCH

• Evaluation can examine the effectiveness of a program or policies, as can research;

• They share the same methodologies (styles, instrumentation, sampling, ethics, reliability, validity, data analysis techniques, reporting and dissemination mechanisms).

Page 5: Chapter3

DIFFERENCES BETWEEN EVALUATION AND RESEARCH

(Smith, M. and Glass, G. (1987) Research and Evaluation in the Social Sciences. New Jersey: Prentice Hall)

• The intents and purposes of the investigation;• The scope of the investigation;• Values in the investigation;• The origins of the study;• The uses of the study;• The timeliness of the study;• Criteria for judging the study;• The agendas of the study.

Page 6: Chapter3

DIFFERENCES BETWEEN EVALUATION AND RESEARCH

(Norris, N. (1990) Understanding Educational Evaluation. London: Kogan Page)

• The motivation of the enquirer;• The objectives of the research;• Laws versus description;• The role of explanation;• The autonomy of the enquiry;

Page 7: Chapter3

DIFFERENCES BETWEEN EVALUATION AND RESEARCH

(Norris, N. (1990) Understanding Educational Evaluation. London: Kogan Page)

• Properties of the phenomena that are assessed;• Universality of the phenomena studied;• Salience of the value question;• Investigative techniques;• Criteria for assessing the activity;• Disciplinary base.

Page 8: Chapter3

CONFORMATIVE EVALUATION(Stronach, I. and Morris, B. (1994) Polemical notes on educational evaluation in an age of ‘policy hysteria’.

Evaluation and Research in Education, 8 (1&2), pp. 5-19).

• Short-term, takes project goals as given;• Ignores the evaluation of longer-term outcomes;• Gives undue weight to the perceptions of

programme participants who are responsible for the successful development and implementation of the programme: ‘over-reports’ change;

• Neglects/‘under-reports’ the views of some practitioners and critics;

Page 9: Chapter3

CONFORMATIVE EVALUATION

• Adopts an atheoretical approach, and regards the aggregation of opinion as the determination of significance;

• Involves a tight contractual relationship with programme sponsors that disbars public reporting or encourages self-censorship to protect future funding;

• Risks implicit advocacy of the programme in its reporting style.

Page 10: Chapter3

MODELS OF EVALUATION

• Survey;• Experiment;• Illuminative;• The CIPP model:

– Context, Input, Process, Product;– Look for congruence between what was intended

to happen and what actually happened in these four areas.

• Objectives:– How far have the objectives been achieved.

Page 11: Chapter3

 STAKE’S MODEL OF EVALUATIONCongruence between intentions & observations – what actually

happened 

INTENTIONS 

 Congruence

 OBSERVATIONS

Intended antecedents

 

Actual antecedents

Intended transactions

 

Actual transactions

Intended outcomes

 

Actual outcomes

Antecedents = initial conditionsTransactions = processes, what takes place during the program

Page 12: Chapter3

RESEARCH, POLITICS & POLICY MAKING

Politics, research and evaluation are inextricably linked in respect of:

– Funding– Policy-related research– Commissioned research– Control and release of data and findings– Dissemination of research– How does research influence policy?– Who judges research utilization?– Consonance with political agendas

Page 13: Chapter3

RESEARCH, POLITICS & POLICY MAKING

Researchers and policy makers may have conflicting:• Interests• Agendas• Audiences• Time scales• Terminology• Concern for topicality

Page 14: Chapter3

RESEARCH, POLITICS & POLICY MAKING

Policy makers like:• Simple Impact Model

• Superficial facts

• Unequivocal data

• Short term solutions

• Simple, clear remedies for complex, generalized social problems

• Certainty

• Positivist methodologies

Researchers work with:• Complex models

• Complex data

• Uncertain findings

• Longer-term time scales

• Subtle and provisional data on complex and multi-faceted issues

• Conjecture

• Diverse methodologies (fitness for purpose)