Assessing and reporting performances on pre-sessional EAP courses
Assessing and reporting performances on pre-sessional EAP courses
Assessing and reporting performances on pre-sessional EAP courses
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
J. Banerjee, D. Wall / Journal of English for Academic Purposes 5 (2006) 50–69 63<br />
student self-assessment. Since, the exit assessment report re<strong>pre</strong>sented high stakes for the<br />
students they might well have described themselves in a way that they thought appropriate in<br />
order to successfully complete the course. However, the tutors we interviewed felt that their<br />
students had largely been forthright in their self-assessments <str<strong>on</strong>g>and</str<strong>on</strong>g> individual tutorials. In fact,<br />
<strong>on</strong>e tutor commented that the students “were more h<strong>on</strong>est than perhaps we give them credit<br />
for.” Indeed many students encouraged their tutors to be “truthful” so that they could set<br />
learning goals for the rest of the year.<br />
Nevertheless, a number of issues also emerged from the interviews. Most of them are clearly<br />
training issues which need to be addressed as so<strong>on</strong> as possible:<br />
1. The checklist is criteri<strong>on</strong>-referenced <str<strong>on</strong>g>and</str<strong>on</strong>g> each student is to be judged against the can-do<br />
statements. However, at least <strong>on</strong>e tutor not <strong>on</strong>ly c<strong>on</strong>sidered whether the student had the stated<br />
ability or had had difficulty but also tried to rank his students. For example, in the case of the<br />
writing item “Can produce grammatically correct text” he ticked that a student had this<br />
ability not because she performed well in this area but because he wanted to distinguish her<br />
from another good student who did not.<br />
2. During the c<strong>on</strong>sultati<strong>on</strong> <str<strong>on</strong>g>and</str<strong>on</strong>g> piloting phase (see Secti<strong>on</strong> 3.1, above), the tutors specifically<br />
requested a column entitled “we have no evidence” to cover cases where students did not<br />
dem<strong>on</strong>strate a particular ability. The purpose of this column was discussed extensively at the<br />
time yet, during the implementati<strong>on</strong> of the checklist, at least <strong>on</strong>e tutor ag<strong>on</strong>ised over using<br />
the “no evidence” column, worrying that it would imply that the student did not have that<br />
language ability.<br />
3. The exit assessment checklist includes a summary judgement <strong>on</strong> the final page. This was<br />
included at the request of the course tutors (see Secti<strong>on</strong> 3.1, above). However, the interviews<br />
revealed that the route by which the tutors arrived at their summary judgements differed. One<br />
tutor said that she had a global sense of whether a student had made satisfactory progress or<br />
not <str<strong>on</strong>g>and</str<strong>on</strong>g> based her judgement <strong>on</strong> that rather than <strong>on</strong> the distributi<strong>on</strong> of ticks across the report.<br />
Another tutor based her judgement <strong>on</strong> the number of ticks the student received in the “has<br />
had some difficulties” column as well as her judgement of the gravity of these ticks. This last<br />
practice is particularly interesting for it introduces a level of detail not explicit in the report<br />
form. This has implicati<strong>on</strong>s for the relati<strong>on</strong>ship between the detailed checklist <str<strong>on</strong>g>and</str<strong>on</strong>g> the<br />
summary secti<strong>on</strong> of the report form <str<strong>on</strong>g>and</str<strong>on</strong>g> is again an issue to be dealt with in training.<br />
Though there is still much to be d<strong>on</strong>e to help tutors adjust to the new <str<strong>on</strong>g>reporting</str<strong>on</strong>g> system <str<strong>on</strong>g>and</str<strong>on</strong>g> to<br />
orientate new users of the checklist, the remaining c<strong>on</strong>cern relates to how the tutors applied the<br />
“can-do” statements. It is important to remember that the tutors had discussed their<br />
inter<strong>pre</strong>tati<strong>on</strong>s of items during the c<strong>on</strong>sultati<strong>on</strong> <str<strong>on</strong>g>and</str<strong>on</strong>g> piloting stage (described earlier) <str<strong>on</strong>g>and</str<strong>on</strong>g> were<br />
therefore familiar with the checklist. Indeed, they had c<strong>on</strong>tributed to its wording. Yet, as Lumley<br />
(2002) points out, when making their judgements all raters have to rec<strong>on</strong>cile the rating scale (in<br />
this case the can-do statements) <str<strong>on</strong>g>and</str<strong>on</strong>g> their observati<strong>on</strong> <str<strong>on</strong>g>and</str<strong>on</strong>g> inter<strong>pre</strong>tati<strong>on</strong> of student<br />
<str<strong>on</strong>g>performances</str<strong>on</strong>g>. This proved to be an issue for the tutors we interviewed. For instance, <strong>on</strong>e<br />
tutor commented <strong>on</strong> the word “satisfactory” in the statement “can take satisfactory notes.” Since,<br />
notes are generally for pers<strong>on</strong>al use, this tutor w<strong>on</strong>dered whether she was the best judge of<br />
whether a student’s notes were “satisfactory” for the student’s own purposes. Much can be d<strong>on</strong>e<br />
to discuss with the course tutors the behaviours that c<strong>on</strong>stitute evidence of a particular ability (as<br />
well as the behaviours that suggest difficulty). However, the tutor is still the ultimate arbitrator<br />
<str<strong>on</strong>g>and</str<strong>on</strong>g>, as Lumley (2002) argues, it is not possible to cover all eventualities during rater training.