21
IMPLEMENTING AN EMAIL REQUEST TASK FOR ACADEMIC WRITING ASSESSMENT: A STUDY OF RELIABILITY AND TEST DEVELOPMENT Randall Rebman [email protected] Northern Arizona University

Talk given at AAAL 2013

Embed Size (px)

Citation preview

Page 1: Talk given at AAAL 2013

IMPLEMENTING AN EMAIL REQUEST TASK

FOR ACADEMIC WRITING ASSESSMENT: A

STUDY OF RELIABILITY AND TEST

DEVELOPMENT

Randall Rebman

[email protected]

Northern Arizona University

Page 2: Talk given at AAAL 2013

LITERATURE REVIEW

• Pragmatics is part of most models of communicative

competence (Bachman & Palmer, 2010; Canale, 1983;

Canale & Swain, 1980).

• The testing of pragmatic competence is an

underexplored area in second language assessment

(Roever, 2011).

• Tasks assessing the construct of academic writing can

include integrated (reading/writing or listening/writing),

independent, and situational-based tasks (Cumming et

al., 2000).

Page 3: Talk given at AAAL 2013

TARGET DOMAIN OF ACADEMIC WRITING

• It is important to select writing tasks from the

Target Language Use (TLU) domain that can be

developed for assessment tasks (Bachman &

Palmer, 2010).

– Select academic writing tasks that:

• occur in introductory level academic courses that

require writing (Cumming et al., 2000).

• occur in academic communication situations

between peers and faculty and in general on

campus (Youn,S.J., 2009).

Page 4: Talk given at AAAL 2013

TEST PURPOSE

• To place L2 students in different levels of writing ability.

• To decide if students matriculate into the university from the

intensive English program.

• To include a representation of writing tasks that second language

writers will be required to produce in university contexts.

• To explore how a broader coverage of the construct can be

attained using a pragmatic task for writing assessment

Page 5: Talk given at AAAL 2013

WHY AN EMAIL TASK?

• Needs

– EAP students struggle with using the proper conventions of

email for communication (Youn,S.J., 2009)

• Washback

– Potential for positive washback (Crusan, 2010) by

encouraging more coverage of pragmatic features of emails

(indirect and direct request strategies, genre markers) in

course instruction

• Adding an email task to a writing test can expand the range of the

construct of academic writing that is assessed

Page 6: Talk given at AAAL 2013

LIMITED TEST DOMAIN

Situational-based Task:

Request to a Professor

Integrated Task:Summary of a

chart

Independent Task: Prompt-based

Argumentative essay

Page 7: Talk given at AAAL 2013

RESEARCH QUESTIONS FOR EMAIL TASK (RQ 1&2)

1. Can the raters produce consistent ratings of an

email writing task using a new rubric?

2. Is the email response task testing academic

writing ability in a different way than the

integrated and independent writing tasks?

Page 8: Talk given at AAAL 2013

PARTICIPANTS

• n= 103

• International students mainly representing China,

Saudi Arabia, Japan, Korea, and Kuwait

• Ages ranged from 18-25

• All were pre-university students required to take

the English placement test to determine level

placement into the intensive English program or

into Northern Arizona University

Page 9: Talk given at AAAL 2013

METHODS

• University IEP students were given the prototype email task as

part of a placement/exit test.

• A holistic scale for the prototype email task was created by the

IEP assessment team using the empirical method (Weigle, 2002)

during the piloting of the task.

• A 6-point rubric operationalized the construct of writing ability on

the email task. This resulted in a score of 0-5 given by a single

rater.

• The summed score between two raters was used to give a score

ranging from 0-10.

Page 10: Talk given at AAAL 2013

METHODS

• The Spearman rho is used to measure and inter-rater reliability

for RQ1.

• A Spearman rho is used to measure the internal consistency

between the email task, integrated task and independent task for

RQ 2.

Page 11: Talk given at AAAL 2013

METHODS

Email task rubric scoring criteria developed through the empirical

method covers the following features:

• Language use

• Grammatical and lexical features

• Register awareness, including appropriate forms of address

• Genre markers specific to emails

• Topical relevance

• Task completion

Page 12: Talk given at AAAL 2013

TEST TASK CHARACTERISTICS

Task Prompt

Ddirections: Read the question below. Plan, write, and revise an email.

Use the space below to prepare writing your email. You may begin now.

Question: You are new at XXX University. You do not know what

classes to take. Write an email to Professor Smith to do the following:

1) introduce yourself

2) explain your problem

3) ask for advice

Page 13: Talk given at AAAL 2013

RESULTS: RQ1

Ratings

Correlation Coefficient N

Email Rating 1 -- 103

Email Rating 2 .92 103

Correlation Coefficient for Inter-rater reliability

Note. df = 172; alpha .05; rho critical = .364; N = Number of pairs;

95% CI

Ratings M SD N LL UL

Email Rating 1 3.44 1.13 103 3.22 3.66

Email Rating 2 3.47 1.14 103 3.24 3.69

Descriptive Statistics

Note. CI = confidence interval; LL = lower limit; UL = upper limit

Page 14: Talk given at AAAL 2013

RESULTS: RQ2

Note. CI = confidence interval; LL = lower limit; UL = upper limit

95% CI

Writing Task M SD N LL UL

Email 6.90 2.22 103 6.46 7.33

Independent 4.91 2.02 103 4.51 5.30

Integrated 4.66 1.87 103 4.23 5.03

Correlations Between Writing Tasks

Note. df = 172; alpha .05; rho critical = .364; N = Number of pairs;

Page 15: Talk given at AAAL 2013

DISCUSSION

• Students had higher mean scores on the email task than on the

other two task types.

• What are the limitations for implementing the new task?

– The test scores on the email task did not distinguish students by

writing ability.

– The task appears to be lacking complexity or the scale made it too

easy to get a high score.

Page 16: Talk given at AAAL 2013

IMPLICATIONS FOR ONGOING TEST TASK

DEVELOPMENT

• The email task could be improved by adding complexity to the

task design.

– Give more input for learners to respond to

• Ex: a sample email from a professor to which they must respond to

– involve a higher imposition request to the professor

Page 17: Talk given at AAAL 2013

IMPLICATIONS FOR ONGOING TEST TASK

DEVELOPMENT

• The scale must be revised to better distinguish criteria expected

for different bands of rubric.

– A sample of emails to faculty members could be gathered to

identify pragmatic features lacking in current task design.

– We could take out any criteria in rubric that does not apply to

pragmatic features.

Page 18: Talk given at AAAL 2013

REVISED TASK

• Writing Task 1: Email a Professor (10 minutes total)

• Planning (2 minutes): Read the situation below. Plan and write an

email. Use the space below to prepare writing your email.

• Situation: You are a student at Northern Arizona University. You

have to turn in an essay in five days, but your writing has many

problems. Write an email to Professor Smith and do the following:

– fill in the subject line

– introduce yourself

– explain your problem

– request an appointment to get help

– ask for a reply

Page 19: Talk given at AAAL 2013

THANK YOU FOR COMING

Questions?

Page 20: Talk given at AAAL 2013

IMPLEMENTING AN EMAIL REQUEST TASK

FOR ACADEMIC WRITING ASSESSMENT: A

STUDY OF RELIABILITY AND TEST

DEVELOPMENT

Randall Rebman

[email protected]

Northern Arizona University

Page 21: Talk given at AAAL 2013

CLICK TO EDIT MASTER TITLE STYLE

CORRECTION FOR ATTENUATION

Correlations Corrected for Attenuation