Upload
others
View
2
Download
0
Embed Size (px)
Citation preview
Language teaching methodology
256
AN INVESTIGATION INTO TEACHERS’ PRACTICE
OF EFL WRITING ASSESSMENT
Truong Thi Anh*
Nguyen Ho Hoang Thuy
University of Foreign Languages, Hue University
Abstract: This study aimed to explore the teachers’ practice of EFL writing
assessment at some upper secondary schools in Quang Tri province, Vietnam. For this
purpose, the study adopted a mixed-methods approach. Sixty EFL teachers from
sixteen schools responded to a five-point Likert scale questionnaire, ten of them
participated in subsequent interview sessions and five teachers shared their students’
assessment writing papers. Empirical data indicated that teachers often used five
criteria of content, organization, language use, vocabulary and mechanics in
evaluating the students’ written works, and frequently responded to student writing by
giving marks and feedback. However, there remained some major issues: linguistic
features were prioritized, sources of self-peer assessment were poorly utilized, success
criteria were not explicitly framed as part of feedback, and written feedback lacked
diagnostic information about students’ strengths and weaknesses in writing. The study
revealed many impediments constraining teachers’ implementation of classroom
writing assessment.
Key words: EFL writing assessment, assessment criteria, feedback.
1. Introduction
National Foreign Languages Project 2020, which was launched by Vietnam’s
Ministry of Education and Training (MOET) and has now been extended to 2025, has
made a huge reform in curriculum design, teaching methodology and assessment. In terms
of language testing and assessment, Vietnam’s MOET requires that the assessment of four
language skills namely reading, listening, speaking and writing be comprehensively
conducted at schools of all levels (Official Document No 5333/BGDĐT-GDTrH) so that
students, on completing upper secondary education, will have achieved level three of the
Foreign Language Proficiency Framework for Vietnam, the equivalent of CEFR level B1
(Hoang, Hoang, Dang, Phan, Hoang, Kieu, Vu, & Dao, 2014). This means that there is an
emphasis on assessing language skills in the teaching-learning process. Nevertheless, in
practice, many teachers keep using paper-pencil tests mainly for grading purposes while
assessing language skills is often ignored. How to assess writing effectively seems to
* Email: [email protected]
Proceedings of the 4th International Conference
Interdisciplinary Research in Linguistics and Language Education
257
receive little attention from high school teachers. A number of teachers hold a belief that
assessing writing is giving learners a topic to write and marking their papers. Nonetheless,
writing assessment is not that simple. Khongput (2010) asserts writing as “an art” which
requires “the mixture of criteria, judgment and experience” as an assessment method (p. 7).
It is even more demanding when writing assessment aims to enhance the teaching and
learning of writing.
A number of previous studies have investigated teachers’ practice of writing
assessment in several EFL contexts such as Thailand, Hongkong, Ethiopia, Saudi Arabia,
etc. Teachers’ assessment literacy was surveyed in Crusan, Plakans and Gebril (2016); the
concerns as well as obstacles teachers faced were explored in Obeid (2017); controversies
around formative assessment were found in Guadu and Boersma (2018) and Lee and
Coniam (2013); the criteria raters focused on were discussed in Ezza (2017) and Khongput
(2010). Some studies conducted in Vietnam have dealt with rater consistency in essay
evaluation (Nguyen, 2016), and self-assessment in writing class (Nguyen, 2015). It can be
noted that these studies have examined certain aspects of EFL writing assessment at
tertiary education contexts while few studies have examined these issues at secondary
education settings, where there are significant differences in instructional programs,
teacher experiences, student levels, etc. Also, it is problematic that none of the
aforementioned studies provides empirical evidence of what criteria school teachers
employ in assessing writing and how they respond to student writing. This leaves a critical
gap for the study to explore. The current study therefore sets out to investigate Vietnamese
EFL teachers’ implementation of writing assessment at some high schools in Quang Tri,
Vietnam and highlight its relevant issues.
2. Literature review
2.1. Definition of assessment
Assessment is defined as the process of gathering and synthesizing information to
discover and document students’ strengths and weaknesses, plan and enhance instruction,
or evaluate progress and make decisions about students (Cizek, 1997). Assessment can be
used for formative or summative purposes. Formative assessment aims to provide learners
and instructors with information on learners’ present performance to improve it in the
future (Andrade & Cizek, 2010). Brown (2004) emphasizes that the key to such
information is delivering immediate and appropriate feedback on performance, with an eye
toward the future continuation of learning. Summative assessment differs from formative
assessment in a fundamental way; it aims to measure, or summarize student’s achievement
at the end of a course or unit of instruction (Brown, 2004).
Language teaching methodology
258
2.2. Nature of writing ability
English writing in many ways is perceived as the most difficult language skill.
Compared to speaking as a so-called communicative skill, speaking setting provides
interlocutors with prosodic features such as pitch, rhythm, intonation and non-verbal
features that enhance meaning; in contrast, the writing context is created by words alone
without having direct interaction between the writer and the reader (Weigle, 2002). Due to
its non-interactive nature, writing is a distinct mode of communication involving writers’
attempts to convey meaning with an absent reader in both time and space distances.
Writing is also seen as a cognitive phenomenon. Nunan (1989) strongly states that writing
is a complex cognitive process involving the writer controlling a number of variables both
inside and outside sentence level. This means that writers must control content, format,
vocabulary, punctuation, spelling, and letter formation simultaneously. Also, they must be
able to organize and integrate their ideas into cohesive and coherent paragraphs. The nature
of writing is related to not only linguistic knowledge but also sociolinguistic knowledge.
As Weigle (2002) states, it is important to view writing not solely as the product of an
individual, but as a sociocultural act. In other words, writing does not simply make surface
of the text, but how to use language appropriately in social contexts is more significant.
Consequently, writing is the most difficult language skill to teach, and when it comes to
assess, it is much more difficult.
2.3. Product and process approaches to writing
In teaching and assessing writing, the product and process approaches see writing in
different perspectives. The product approach focuses on the learner’s final product with
error-free performance (Nunan, 1999). This means the written product must be a
grammatically correct and coherent text, with appropriate use of vocabulary, grammar and
language devices. While the product-oriented approach places the emphasis on accurate
language form in the final text, the process approach concerns more on the process of how
students develop ideas and formulate them into effective written works. The process
approach sees writing as a non-linear sequence in which learners go recursively through such
stages as planning, drafting, editing, revising and publishing, and it puts special emphasis on
audience and interaction with peers and teachers (Hyland, 2003). However different those
explanations are, there is an important point on which they all agree; that is, a good product
depends on good process.
2.4. Writing assessment criteria
In assessing writing, to avoid bias from raters and yield consistent assessment, there
needs to be a written rubric that outlines the criteria. The most widely used analytic scale in
assessing compositions created by Jacob, Zingkgraf, Wormuth, Hartfiel, and Hughey (1981)
Proceedings of the 4th International Conference
Interdisciplinary Research in Linguistics and Language Education
259
presents a set of criteria including content, organization, vocabulary, language use, and
mechanics. In this profile, content receives more weight than the other indicators, which
means evaluating content of a written product is considered the most important. A highly-
rated composition should have its content relevant to topic, develop the thesis thoroughly,
and address task requirements appropriately. Organization is also invariably significant. A
well-organized composition focuses on coherence and cohesion. Coherence is concerned
with the overall interpretation of a text as a unified piece of discourse while cohesion deals
with formal link by using cohesive devices including lexical devices (e.g., repetition,
synonym, antonym, superordinate) and grammatical devices (e.g., reference, substitution,
ellipsis, conjunction). Additionally, vocabulary is considered the foundation of writing and
one of important features of good writing. Regarding language use, it refers to the use of
grammar accurately and appropriately to convey intended meaning. A composition with
good use of language should have effective constructions, few errors of agreement, tense,
word order, articles, pronouns, and prepositions. The last feature is mechanics (punctuation,
spelling, and capitalization), which receives the least emphasis in Jacob et al’.s (1981) scale.
2.5. Scoring a written product
Scoring a piece of writing can be either holistic or analytic. Holistic scoring is a
method by which raters give a single score based on their overall impression of a piece of
writing. The obvious benefit of this type is that it requires a short span of time to evaluate a
paper since only a single score is reported and raters do not comment the student’s work.
However, reducing writing to a single score means teachers gain little information about
students’ specific strengths and weaknesses, thus cannot feedback into their instruction
(Hyland, 2003). Analytic scoring, on the other hand, is a method by which raters evaluate
different aspects of writing performance separately. It is suggested that while holistic
scales are faster, analytic scales tend to be more reliable, and certainly provide more useful
feedback, as scores on different aspects of writing can tell students what their strengths and
weaknesses are (Weigle, 2007). As a result, analytic rubric is useful as a diagnostic
teaching tool. As Brown (2004) claims, in classroom settings where a teacher wishes to
adapt their teaching to the needs of students, much more information across sub-skills is
desirable than a single score.
2.6. Feedback in writing assessment
Despite the importance of timed writing, a single script written within limited time is
insufficient to measure a student’s ability, because it provides a very restricted picture of
what the student is able to do with writing in real world (Weigle, 2002). That is why
Brown (2004) suggests an assessment type which is informal, formative and replete with
washback be conducted in the tutored relationship of teacher and student, and in the
Language teaching methodology
260
community of peers. In the same manner, Lee (2017) claims that feedback delivered during
stages of writing process lies at the heart of classroom writing assessment. In the article
The Power of Feedback, Hattie and Timperley (2007) stress “feedback is one of the most
powerful influences on learning and achievement” (p. 81). They suggest that effective
feedback be divided into three stages, feed up (Where am I going?), feedback (How am I
going?), and feed forward (Where to next?). In the feed up stage, the teacher shares
learning goals and success criteria with students and provides instructional scaffolding. In
the feedback stage, descriptive and diagnostic feedback linking planned criteria is delivered
to students, informing them of what progress they are making toward the goal. The last
stage, feed forward, provides information about what students need to do next to move
forward in their learning.
Feedback to student writing can come from both teachers and students. Hyland
(2003) points out that teachers’ response takes a variety of forms, such common ones as
commentary, rubrics, minimal marking, and electronic feedback. According to Lee (2017),
written feedback should be followed by oral feedback so that teachers can explain
ambiguities and students have opportunities to ask questions. Lee (2017) asserts that
teachers should use feedback to inform students of their major strengths and areas for
improvement with reference to the learning goals. Additionally, peer feedback is used
synonymously with peer response or peer assessment, which has potential in enhancing L2
student writing through creating meaningful interaction among peers (Lee, 2017). Andrade
and Cizek (2010) claim that students are in favour of feedback on their work provided by
peers rather than by the teacher because peer feedback can be more immediate, timely and
individualized. Lee (2017) stresses that if students engage in peer feedback meaningfully,
they not only help their peers improve their writing but also develop their ability to review
and critique their own writing. Likewise, self-assessment is important in life-long learning
and reduces teacher’s burden (Butt, 2010). Students should be responsible for their own
learning, achieve self-awareness of their work, rather than teachers constantly tell them
what is good or bad. Self-assessment is an emphasis of student-centered approach which
places learners as the center of teaching and learning (Iraji, Enayat & Momeni, 2016). To
sum up, apart from using scoring methods, classroom teachers can make the best use of
peer/ self-assessment to assess student writing. By evaluating student work done in both
timed and untimed settings, teachers will be able to get a full picture of what their students
are capable of doing in writing.
3. Research methodology
3.1. Participants and contexts
The number of participants in this study was 60 EFL teachers (11 males and 49
Proceedings of the 4th International Conference
Interdisciplinary Research in Linguistics and Language Education
261
females) who were from 16 upper secondary schools in Quang Tri province, making up
31.9% of the population. The samples’ teaching experience ranged from one to over twenty
years, 8 had a Master’s degree and 52 had a Bachelor’s degree in EFL methodology.
3.2. Data collection instruments and procedure
In order to investigate teachers’ practice of EFL writing assessment, a mixed-
methods design was adopted (Creswell, 2012). The study started with collection and
analysis of quantitative data from questionnaire to gain an overall understanding of
emerging problems and followed by qualitative data to obtain further information.
The full questionnaire consisted of 48 close-ended statements exploring teachers’
perceptions and practice of EFL writing assessment. For the purpose of this article, 5 items
concerning teachers’ use of assessment criteria and 12 items inquiring teachers’ responding
to student writing were chosen for analysis. The teachers were asked to indicate their
opinion on a 5-point Likert scale (5 = always, 4 = often, 3 = sometimes, 2 = rarely, and 1 =
never). The questionnaire was written in Vietnamese to ensure clarity of meaning for the
participants and distributed online by means of Google form via email and Facebook.
Following initial analysis of the quantitative data, individual semi-structured interviews
were carried out in Vietnamese with 10 participants. Each individual interview lasted from
15 to 25 minutes and was audio-taped with the consent of the interviewees. Collecting
students’ assessment papers was also employed to examine instructors’ real experiences in
assessing writing. The researcher made contact with the teachers and five of them agreed
to send writing papers soon after they conducted an assessment in their classes.
3.3. Data analysis
Descriptive analysis was used to transform a set of numbers into indices to describe the
survey data (Creswell, 2012). Statistical numbers, including percentages, means and standard
deviations were computed using EXCEL. The results were then presented in figures and
tables to make the interpretations comprehensible. The interview data was transcribed and
organized into major themes which were identified based on the messages of word repetition
or similar phrases to elaborate the questionnaire data. Regarding assessment of writing
samples, the researcher read assessment writing papers carefully to have a general view of
the teachers’ responding to student writing. The analysis centered on the assessment criteria
in focus and the way feedback was given in the margins of papers.
4. Findings and discussion
4.1. Criteria in writing assessment
Quantitative data concerning criteria which surveyed what teachers focused on while
evaluating students’ writing are summarized in Table 1 below.
Language teaching methodology
262
Table 1. Frequency of criteria used in writing assessment (AW = Always, O = Often,
S = Sometimes, R = Rarely, N = Never, M = Mean, SD = Standard Deviation)
Questions AW
%
O
%
S
%
R
%
N
% M SD
I evaluate content of my student’s composition. (1) 25.0 61.7 10.0 3.3 0.0 4.08 0.70
I evaluate organization of my student’s composition. (2) 35.0 51.7 11.7 1.7 0.0 4.20 0.71
I evaluate language use of my student’s composition. (3) 26.7 65.0 6.7 1.7 0.0 4.17 0.62
I evaluate vocabulary of my student’s composition. (4) 30.0 56.7 13.3 0.0 0.0 4.17 0.64
I evaluate mechanics of my student’s composition. (5) 23.3 46.7 26.7 3.3 0.0 3.90 0.80
As shown in Table 4.1, there are substantial proportions of participants reporting
always and often using the listed criteria to evaluate students’ composition, which indicates
comparatively high frequencies of using assessment criteria among participants. The table
also shows that participants did not assign equal weighting to every feature of writing but
prioritized some over others. The rank of the writing assessment criteria according to mean
scores from the highest to the lowest is illustrated in Figure 1.
Figure 1. Rank of criteria used in writing assessment
As shown in Figure 4.1, organization was the most frequently used criterion with the
highest mean score (M=4.20), followed by language use and vocabulary, at the same
frequencies as illustrated by equal mean scores (M=4.17), content ranked fourth and
mechanics is at the lowest range of the list. Interestingly, participants weighed
organization, language use and vocabulary more than content. In other words, form took
priority over meaning. This may be due to the fact that most of the teaching follows a
structural approach that privileges form over meaning in EFL classes in Vietnam (Le,
2011). According to the interviewees, competence in grammar and vocabulary would
Proceedings of the 4th International Conference
Interdisciplinary Research in Linguistics and Language Education
263
determine the quality of writing outcome. The students having problems with grammar and
vocabulary might have low writing ability, as a teacher commented “Some of my students
have poor structure and incorrect vocabulary usage which limit their ability to complete a
piece of writing”. Moreover, teachers’ choice of criteria might depend on the writing tasks
they had to teach and then assess. In fact, paragraphs, letters and essays are some popular
genres that school students are exposed to. Accordingly, the teachers assessed their
students’ accumulation of sub-skills taught in these writing lessons, emphasizing steps
required to produce a well-organized text, as shared by an interviewee:
I normally assess what students have learnt in writing lessons, especially whether
they organize a paragraph with a topic sentence, supporting ideas, and a concluding
sentence, or whether they follow an appropriate format of a letter.
It could be inferred from the findings above that many participants considered
vocabulary and grammar, and text structure important while fewer of them were concerned
about content or meaning embedded in the written texts. This implies that the teachers paid
more attention to linguistic features than communicative ones in the assessment of student
writing performance.
Compared to relevant literature, the reported findings both confirm and contradict
those in previous studies. That mechanics ranked lowest is similar to Khongput’s (2010)
study and Jacob et al’.s (1981) profile. In Jacob et al. (1981) and Khongput (2010), content
received more weight than the other indicators, which is in marked contrast to the current
results. Differently, participants in Ezza’s (2017) study employed ideas, spelling, punctuation
and text structure as top criteria in writing exams. In fact, writing is a demanding skill in its
nature and assessing writing is seen as a complicated activity in which raters’ subjectivity
and experience may play an important role (Weigle, 2002). This more or less affects the
various ways in teachers’ use of criteria; therefore, although certain criteria are employed for
rating writing, some variability between practices still occurs. This is plausible when Ezza
(2017) concluded that the use of rubrics is affected by some variables, including instructional
experience, academic levels at which writing courses are offered and the type of writing
being assessed.
4.2. Teachers’ responding to student writing
This section deals with data concerning how teachers responded to students’
writings. Table 2 details percentages, means and standard deviations of respondents’ rating
to every single item.
Language teaching methodology
264
Table 2. Frequency of teachers’ activities in assessing writing
Questions AW
%
O
%
S
%
R
%
N
% M SD
I involve my students in assessment process. (6) 20.0 46.7 25.0 8.3 0.0 3.78 0.87
I assess student ability during stages of writing
process. (7) 31.7 26.7 23.3 15.0 3.3 3.68 1.17
I inform students of criteria before assessing their
works. (8) 15.0 53.3 21.7 10.0 0.0 3.73 0.84
I apply rubrics in my writing classes. (9) 8.3 43.3 31.7 11.7 5.0 3.38 0.98
I provide students with marks on their papers. (10) 28.3 56.7 11.7 3.3 0.0 4.10 0.73
I provide students with comments on their papers. (11) 30.0 58.3 11.7 0.0 0.0 4.18 0.62
Oral feedback is used to assess students’ work. (12) 13.3 41.7 33.3 11.7 0.0 3.57 0.87
Written feedback is used to assess students’ work. (13) 21.7 48.2 26.7 1.7 1.7 3.87 0.83
I inform students of their strong points in writing. (14) 23.3 30.0 43.3 3.4 0.0 3.73 0.86
I inform students of their weak points in writing. (15) 15.0 36.7 43.3 5.0 0.0 3.62 0.80
I apply peer assessment in my writing classes. (16) 16.7 43.3 31.7 8.3 0.0 3.68 0.85
I apply self-assessment in my writing classes. (17) 1.7 28.3 35.0 25.0 10.0 2.87 1.00
As indicated in Table 2, two outstanding items are 10 (M = 4.10) and 11 (M = 4.18).
Their comparatively high mean scores indicate that the participants regularly gave marks
and comments on students’ papers. That the mean score of item 11 is a bit higher than that
of item 10 implies they tended to give more comments than marks. Item 17 got the lowest
mean score (M=2.87) which indicates that self-assessment was least utilized. Responses to
the rest of items were fairly homogeneous in a slight variation within 0.49 and the mean
range from 3.38 to 3.87. These medium mean scores show that activities including giving
feedback, informing progress, involving students, assessing writing process, informing
criteria, applying rubrics, and peer assessment were moderately conducted. The table also
reveals that both oral feedback and written feedback were utilized by many teachers. This
is convincing when we see around 60% of surveyed teachers reported always and often
doing so in items 12 (M = 3.57) and 13 (M = 3.87). Such practice is quite good because
oral feedback can make sure students understand the feedback deeply by asking questions
if there is any misunderstanding and written feedback allows them to go back and read the
comments later on (Lee, 2017). These analyzed items also tell that although both kinds of
Proceedings of the 4th International Conference
Interdisciplinary Research in Linguistics and Language Education
265
feedback were valued, written feedback was used more frequently. Even though the
teachers claimed to provide feedback either regularly or often, it is not how much feedback
is provided but how effectively it is delivered. This raised issues for further exploration about
teacher responding to student writing.
The first issue is teacher’ assessing student writing seemed limited to writing
process. As illustrated in item 16 (M = 3.68, SD = 1.17), 23.3% of surveyed teachers
reported that they sometimes assessed student during the writing process and nearly 20%
of them rarely and never had such practice. The process approach is believed to encourage
student thinking and revising, thereby enhances writing skills (Hyland, 2003). However,
interviewed results indicate that limited class time prevented teachers from giving frequent
and subsequent feedback to student writing. An interviewee asserted like this:
45 minutes were too short for many activities, teacher teaching students how to
write, students practicing writing and teacher giving feedback. I just have a few final
minutes to evaluate some writings and no time to do this with multiple drafts from 42
students.
Understandably, time is a big challenge to all school teachers when they are
mandated to cover three or more writing tasks as defined in the syllabus within a 45-
minute lesson with a class size of around 40 students. Previously stated, self-/peer
assessment techniques are viewed not only to reduce teachers’ workload but also to
accelerate student writing (Butt, 2010; Lee, 2017). Unexpectedly, the collected data clearly
indicate that the teachers did not frequently involve students in assessment process (item 6,
M = 3.78). This anticipation can be also seen in items 16 (M = 3.68) and 17 (M = 2.87)
which reveal peer assessment was only sometimes utilized and rarely was self-assessment
applied. Students’ low proficiency and time constrains were two primary reasons for the
teachers to skip or spend little time on self-/ peer feedback. As for self-assessment, some
opinions were cited “However good a student is, he or she cannot self-assess their written
work”, “From my experience, students do not often realize their mistakes until someone
shows them”. Undoubtedly, it is not easy for L2 learners to assess their own work because
writing requires sub-skills within and beyond sentence level (Nunan, 1989). By contrast,
research by Iraji et al. (2016) shows that students are capable of analyzing and responding
to their own writing, so self-assessment can boost learner autonomy, promote intrinsic
motivation, and make them prepared for life-long learning. Many teachers still kept doubts
about student ability, so narrowed their chance of assessing writing on their own, which is
in harmony with Nguyen’s (2015) finding. Certainly, teacher feedback should be valued;
however, if students rely too much on the teacher, their proactive involvement in learning is
limited, resulting in being passive consumers of feedback.
Language teaching methodology
266
Additionally, rubric is considered a useful tool to inform criteria and give feedback,
based on which students would have a sense of what criteria needed for the desired
performance (Butt, 2010; Hyland, 2003; Lee, 2017). Conversely, it was not often
implemented, as illustrated in item 9 (M = 3.38). Although it was reported that the teachers
evaluated writings by using certain criteria, rubric did not seem to work well when there was
a significant variation among participants’ responses. Interviewed teachers admitted telling
their students the requirements of a composition in the pre-stage of a writing lesson or on the
outset of assessment process, but gave no evidence of assessment forms such as a checklist
or rubric. This implies that the teachers did not develop any criteria sheets to share with
students. It would thus be reasonable to say that criteria for assessment tasks were not
communicated to the students explicitly and transparently. The results reveal that the
teachers employed certain criteria in assessing writing, yet they did not seem to implement
them effectively inside the classroom. These findings are in accordance with those found in
Crusan et al. (2016) that teachers are inadequately prepared in the use of rubrics. Obeid
(2017) shares the same point that rubric is a major obstacle that teachers encounter in
practicing writing assessment. How to design and apply rubrics effectively is a really
remaining issue in EFL contexts, as Myskow (2011) argues that for most EFL teachers who
are already burdened with many responsibilities, an investment on a single rubric is not
realistic.
Ultimately, data from assessment papers inform written comments seemed not
diagnostic enough to help students realize their strengths and weaknesses in writing. In
some cases, the teachers wrote some words “good”, “quite good” on high-mark papers.
These praises could motivate students as they rewarded their successful performance but
lacked explanation about what they did well. In many cases, the teachers made direct
corrections of grammar and spelling mistakes by crossing over wrong words and replacing
them with correct ones, whereas in some others they made indirect comments by
underlining or circling the part with errors. Seemingly, teachers seemed to be attracted by
observable mistakes in spelling, vocabulary and grammar while comments on other
features such as organization and content were almost absent. Even some papers were left
merely with numerical values which implied little diagnostic information. Accordingly,
such manners of teacher giving feedback were not descriptive enough to inform students of
what they could and could not do with writing and how to tackle similar problems next
time. The most common observation emerging from all assessed writings is that feedback
was provided at the same time as the grade. Especially, the teachers gave a single score to
the overall quality of writing rather than giving scores to distinct features. If a writing
paper is returned to the student with just a grade and without informing strengths and
weaknesses, it might have negative washback because he/she does not identify his/her
knowledge gaps (Brown, 2004). Consequently, data from assessed papers reveal that
Proceedings of the 4th International Conference
Interdisciplinary Research in Linguistics and Language Education
267
assessment information is mainly used for grading and reporting purposes which are the
typical features of the traditional testing system.
Hattie and Timperley (2007) have suggested that effective feedback should inform
students of learning goals, the progress they are making toward the goal, and how to move
forward to close their learning gaps. The data analysis above, however, indicates that the
feedback provided by participants did not seem to be a powerful tool to enhance writing
because of the restriction of student involvement, lack of explicit explanation of rubrics,
and limitation of subsequent feedback. Written feedback also provided little diagnostic
information about students’ strengths and weaknesses in writing, and was delivered as the
same time as the grade, thereby mainly served summative purposes. Time pressure seemed
the most prominent obstacle affecting the quality of teachers’ feedback because reading
and giving comments on students’ papers are considered time-consuming tasks, as cited
“Assessing writing takes me a lot of time to read and give feedback”, “I often need a lot of
time to mark and write comments for each student writing”. It is even more challenging
when the teachers had to tackle with a large class. This could explain why teachers’ written
feedback tended to be brief and general. Another negative factor might be examination
pressure. In fact, Vietnamese education is very much embedded within examination-
oriented culture which puts pressure on students to learn for grading (Ho, 2015).
According to the teachers, because writing skill is not the focus of standardized tests in
secondary education, many students found it tedious to take part in writing assessment
activities; instead, they were more concerned about graded tasks and summative scores. As
a result, the teachers were inclined to grade rather than assess their students formatively to
get insights into the teaching and learning of writing. The findings confirm Lee and
Coniam’s (2013) study that the examination-oriented system that valued summative scores
could pose severe obstacles to giving feedback in writing classes in Hong Kong schools.
Likewise, this practice is consistent with Guadu and Boersma’s (2018) study that formative
assessment was not promising in Ethiopia due to time constraint to finish the course and
unmanageable class size.
5. Conclusion
The results obtained from quantitative and qualitative data indicated that the teachers
frequently employed assessment criteria and prioritized linguistic features in evaluating
written work. Moreover, they often gave marks and feedback to student writing, yet did not
utilize the tool of rubric, written feedback and self-/peer assessment in order to enhance
student writing. These remaining issues seem to reflect the common situation in many EFL
contexts, particularly in Quang Tri province. This is because there are many setbacks
hindering the effective implementation of EFL writing assessment including student low
proficiency, large class size, syllabus and examination pressure, and class time constraints.
Language teaching methodology
268
How to tackle these problems is still an open question because the context in which the
teachers work does not always allow them to bring their desire to fruition; however, they
should be taken into consideration in congruence with the implementation of EFL writing
assessment.
References
Andrade, H., & Cizek, G.J. (2010). Handbook of formative assessment. UK: Routledge.
Brown, H.D. (2004). Language assessment: Principles and classroom practices. New
York: Pearson Education.
Butt, G. (2010). Making assessment matter. London: Bloomsbury Publishing.
Cizek, G.J. (1997). Learning, achievement, and assessment: Constructs at a crossroads.
In G.D. Phye. Handbook of classroom assessment: Learning, achievement and
adjustment (pp. 2-29). California: Academic Press.
Creswell, J.W. (2012). Educational research: Planning, conducting and evaluating
quantitative and qualitative research (4th edition).. Upper Saddle River, New Jersey:
Pearson Education, Inc.
Crusan, D., Plakans, L., & Gebril, A. (2016). Writing assessment literacy: Surveying
second language teachers’ knowledge, beliefs, and practices. Assessing Writing, 28, 43-56.
Ezza, E.S.Y. (2017). Criteria for assessing EFL writing at Majma’ahuniversity. In S. Hidri
& C. Coombe (2017), Evaluation in foreign language education in the Middle East and
North Africa (pp. 185-200). Springer International Publishing.
Guadu, Z.B., & Boersma, E.J. (2018). EFL instructors’ beliefs and practices of formative
assessment in teaching writing. Journal of Language Teaching and Research, 9(1), 42-50.
Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational
Research, 77(1), 81-112.
Ho, T.N. (2015). An exploratory investigation of the practice of assessment for learning in
Vietnamese higher education: Three case studies of lecturers' practice. Unpublished
doctoral dissertation. Queensland University.
Hoang, V.V., Hoang, T.X.H., Dang, H.G., Phan, H., Hoang, T.H.H., Kieu, T.T.H., Vu,
T.L., & Dao, N.L. (2014). Tieng Anh 10 (Teacher book - Volume 1). Ha Noi: Vietnam
Education Publishing House.
Hyland, K. (2003). Second language writing. Cambridge, UK: Cambridge University
Press.
Iraji, H.R., Enayat, M.J., & Momeni, M. (2016). The effects of self-and peer-assessment on
Iranian EFL learners' argumentative writing performance. Theory and Practice in
Language Studies, 6(4), 716-722.
Proceedings of the 4th International Conference
Interdisciplinary Research in Linguistics and Language Education
269
Jacobs, H., Zingkgraf, S., Wormuth, D., Hartfiel, V., & Hughey, J. (1981). Testing ESL
composition: A practical approach. Rowley, MA: Newbury House.
Khongput, S. (2010). EFL writing assessment practices: Teachers’ perspectives.
Proceedings of the 36th Conference on International Association for Educational
Assessment (pp. 1-8). Thailand.
Le, V.C. (2011). Form-focused instruction: A case study of Vietnamese teachers’ beliefs
and practices. Unpublished doctoral dissertation. University of Waikato, Hamilton, New
Zealand.
Lee, I. (2017). Classroom writing assessment and feedback in L2 school contexts. Singapore:
Springer Nature.
Lee, I., & Coniam, D. (2013). Introducing assessment for learning for EFL writing in an
assessment of learning examination-driven system in Hong Kong. Journal of Second
Language Writing, 22(1), 34-50.
Ministry of Education and Training, V.N. (2014). Dispatch 5333/BGDĐT-GDTrH on
implementing language assessment policy to develop learners' English proficiency at
secondary education level, starting from the academic years 2014-2015. Vietnam, Hanoi:
Ministry of Education and Training.
Myskow, G. (2011). L2 writing assessment in context: The case of a Japanese EFL
secondary school. Toyo University Institute of Human Sciences, 13, 17-33.
Nguyen, T.H.T. (2015). An investigation into students’ self-assessment practice in writing
classes in QuangBinh University. Unpublished master thesis. University of Foreign
Languages, Hue University.
Nguyen, T.Q.Y. (2016). Rater consistency in rating L2 learners’ writing task. VNU Journal
of Foreign Studies, 32(2), 75-84.
Nunan, D. (1989). Designing tasks for the communicative classroom. Cambridge, UK:
Cambridge University Press.
Nunan, D. (1999). Second language teaching and learning. Boston, Mass: Heinle&Heinle
Publishers.
Obeid, R. (2017). Second language writing and assessment: Voices from Within the Saudi
EFL Context. English Language Teaching, 10(6), 174-181.
Weigle, S.C. (2002). Assessing writing. Cambridge, UK: Cambridge University Press.
Weigle, S.C. (2007). Teaching writing teachers about assessment. Journal of Second
Language Writing, 16(3), 194-209.
Language teaching methodology
270
NGHIÊN CỨU THỰC TẾ GIÁO VIÊN
ÁP DỤNG ĐÁNH GIÁ KỸ NĂNG VIẾT TIẾNG ANH
Tóm tắt: Nghiên cứu được thực hiện để điều tra việc giáo viên áp dụng đánh giá kỹ
năng viết tiếng Anh ở một số trường THPT tại Quảng Trị, Việt Nam. 60 giáo viên
tiếng Anh từ 16 trường trả lời phiếu khảo sát, 10 giáo viên tham gia phỏng vấn và 5
giáo viên chia sẻ bài đánh giá. Kết quả cho thấy, giáo viên thường sử dụng năm tiêu
chí, nội dung, tổ chức, sử dụng ngôn ngữ, từ vựng và chính tả, và thường đánh giá bài
viết học sinh bằng điểm và nhận xét. Tuy nhiên, một số vấn đề tồn tại trong đánh giá,
kiến thức ngôn ngữ được ưu tiên, học sinh tự đánh giá và đánh giá chéo ít được áp
dụng, tiêu chí đánh giá không được thể hiện rõ ràng trong phản hồi và nhận xét viết
thiếu thông tin chẩn đoán về điểm mạnh và điểm yếu của học sinh. Nghiên cứu chỉ ra
nhiều yếu tố cản trở giáo viên thực hiện đánh giá viết hiệu quả trong lớp học.
Từ khóa: Đánh giá kỹ năng viết tiếng Anh, tiêu chí đánh giá, phản hồi.