learning - Academic Conferences Limited
learning - Academic Conferences Limited learning - Academic Conferences Limited
Rosario Kane-Iturrioz When compared with the mean test scores (summative) for language skills over the same three years (Figure 5), it is apparent that these show the opposite distribution to the mean Likert values for Writing in 2008-09. 80 70 60 50 Mean scores 40 30 20 10 0 Reading Listening Writing Oral Figure 5: Mean test scores (summative) for language skills 4. Discussion The study set out to explore the effectiveness and learners’ perception of formative assessment and feedback in the context of a blended learning module where online learning was the main feature. There were three questions that the study sought to answer, dealing with students’ perceptions regarding the effectiveness of the formative assessment strategy, their evaluation of the formative feedback provided and the students’ views relating to their improvement in language skills. Regarding the first question, the results of the study clearly demonstrated that the Progress Tests were perceived as an effective strategy. These results corroborated other research findings which have reported that learners valued the availability of online formative assessments because this allowed them to take control of their learning (Lyons and Thorpe, 2009) by facilitating reflexion on the gap between their performance and the level of attainment needed (Gibbs and Simpson, 2004). Some students reported that the Progress Tests had contributed to the development of better independent learning skills involving self-discipline which appeared to indicate that they were on their way to develop a better self-regulation of their learning (Nicol and Mcfarlane-Dick, 2006). Additionally, students stated in the 2008-09 survey that the Progress Tests were the third most popular element of the online provision only surpassed by the convenience of working at their own pace and that multiple attempts were allowed (for both the online exercises and the Progress Test). The latter result confirms that of Felix (2001) who reported that students felt that having multiple attempts promoted language learning. The second question focused on the students’ views regarding the formative feedback received. The results indicated that the respondents associated the provision of an immediate and comprehensive feedback with an increase in their motivation and also in their sense of efficacy. Similar findings were reported by Segarra and Zapata (2008) and Furnborough and Truman (2009). Computer-based feedback was included in the majority of the Progress Tests’ questions from the second year of the study onwards in response to students’ requests. In the case of some learners, the 352 2007-08 2008-09 2010-11
Rosario Kane-Iturrioz instant feedback led to spending “many unanticipated hours in front of the computer in pursuit of the perfect score” (Segarra and Zapata, 2008). A few of the Progress Tests questions retained tutorbased feedback in order to provide students with semi-structured activities that were too complex for automatic computer feedback and required detailed explanation from the tutor. Students found the combination of computer and tutor feedback very useful in subsequent years because this method integrated the benefits of timely guidance with constructive feedback and information on how to improve their performance (Felix, 2003). The voice tools were integrated into the module’s homepage (VLE) and considered easy to use by both students and the tutor. They were used by 40% to 50% of the students because it provided opportunities to practice their oral skills through an online environment and they could also keep their audio files for future reference by downloading them onto their computers or mobile phones. For the tutor, the recorded audio files provided opportunities for comprehensive feedback. From a practical perspective, it also facilitated internal and external moderation of students’ summative oral assessments. However, voice tools for oral practice were not universally welcomed. Approximately one third of the students each year revealed reluctance about using computers to practice their oral skills. The consistent feedback from these students was that they preferred to practice oral conversation in a class-based setting with ‘real’ people. A further drawback mentioned by some students was that they did not like being overheard by other students when using voice tools on-campus. These perceptions showing a dislike by some students to computers as a medium for language learning and the discomfort felt whilst recording on campus were also key features in a study of learner drop-out reported by Stracke (2007). A larger proportion of the students only used the voice eMail tool for the oral activities contained in the Progress Tests as preparation for the summative assessment. In an earlier paper regarding learners’ experiences and perceptions of blended learning for languages (Kane-Iturrioz, 2010), it was reported that the students’ learning experience was still mainly determined by a traditional class-based model that did not foster independent learning skills. In the current study, students referred to their difficulties in adjusting to a regular weekly online workload very different from the learning pattern used in other modules where most of their study was linked to assessments and exams. This corroborates the results of Murday et al. (2008) who also report students’ problems regarding ‘self-regulation’ and its effect on timely study. The final research question sought to determine whether the summative test corroborated students’ perceptions regarding the effectiveness of the module in improving their language skills. The results reported indicate that students’ perceptions of such improvements in 2008-09 seem to mirror the only change introduced in that year, that is, the reorganization of the Progress Tests. Those in 2010-11 could be attributed to the enhanced feedback provided in both materials and Progress Tests and also to the change in the module’s schedule which gave students more time to develop their skills. Which of these changes contributed more to the students’ perceptions cannot be ascertained from the available data. The comparison of the trend in the mean scores and perception values seems to indicate that they closely resemble each other in their incremental direction. A minority of students expressed doubts regarding the appropriateness of an e-blended approach to language learning. However, one of these students who had little confidence in using computers stated, “it was incredibly hard ...but overall I did enjoy it and, you know, I have learnt a lot more Spanish that I thought I would do”. Some students expressed the opinion that “.. for a language module I think online is not the way”. Nevertheless, one of these found the module useful in helping him to become a better independent learner, “I think that I know where I stand as an independent learner, which really helps”. 5. Conclusion Students highly appreciated the module’s assessment strategy designed to provide regular opportunities for reflection on their online learning. The range of formative feedback offered was also welcomed by students because of the combination of timeless provision of immediate computerbased comments/hints and detailed commentary from the tutor. The first type of feedback encouraged many students to engage with the online materials and also improved motivation through a better understanding of the language and the test requirements. The second type, helped students to appreciate their strengths and weaknesses and facilitated a deeper knowledge of the subject. 353
- Page 328 and 329: Danny Glick and Roni Aviram Bernard
- Page 330 and 331: Andrea Gorra and Ollie Jones to hel
- Page 332 and 333: Andrea Gorra and Ollie Jones Howeve
- Page 334 and 335: Andrea Gorra and Ollie Jones Figure
- Page 336 and 337: Andrea Gorra and Ollie Jones Studen
- Page 338 and 339: Rose Heaney and Megan Anne Arroll A
- Page 340 and 341: Rose Heaney and Megan Anne Arroll l
- Page 342 and 343: Rose Heaney and Megan Anne Arroll
- Page 344 and 345: Rose Heaney and Megan Anne Arroll J
- Page 346 and 347: Amanda Jefferies learning was furth
- Page 348 and 349: Amanda Jefferies way in which the o
- Page 350 and 351: Amanda Jefferies their teaching mat
- Page 352 and 353: A Methodology for Incorporating Usa
- Page 354 and 355: Anne Jelfs and Chetz Colwell To try
- Page 356 and 357: Anne Jelfs and Chetz Colwell We wor
- Page 358 and 359: The Virtual Learning Environment -
- Page 360 and 361: John Jessel 2.1 An outline framewor
- Page 362 and 363: John Jessel teachers who agreed to
- Page 364 and 365: John Jessel ‘“reduce the clicks
- Page 366 and 367: Mutlimodal Teaching Through ICT Edu
- Page 368 and 369: Paraskevi Kanari and Georgios Potam
- Page 370 and 371: Paraskevi Kanari and Georgios Potam
- Page 372 and 373: Rosario Kane-Iturrioz Regarding lan
- Page 374 and 375: Rosario Kane-Iturrioz Figure 2: Exa
- Page 376 and 377: Rosario Kane-Iturrioz Tests very us
- Page 380 and 381: Rosario Kane-Iturrioz Although the
- Page 382 and 383: Jana Kapounova et al. eLearning is
- Page 384 and 385: Jana Kapounova et al. Each dimensio
- Page 386 and 387: Jana Kapounova et al. project, conn
- Page 388 and 389: Acknowledgments Jana Kapounova et a
- Page 390 and 391: Andrea Kelz skills and competences
- Page 392 and 393: Andrea Kelz web-based activities in
- Page 394 and 395: Andrea Kelz system. Most other univ
- Page 396 and 397: Open Courses: The Next big Thing in
- Page 398 and 399: Kaido Kikkas et al. However, in the
- Page 400 and 401: Kaido Kikkas et al. generation of w
- Page 402 and 403: Kaido Kikkas et al. Occasional gue
- Page 404 and 405: John Knight and Rebecca Rochon guid
- Page 406 and 407: Evaluation of Quality of Learning S
- Page 408 and 409: Eugenijus Kurilovas et al. Essalmi
- Page 410 and 411: Eugenijus Kurilovas et al. (LOs), l
- Page 412 and 413: Eugenijus Kurilovas et al. Then hie
- Page 414 and 415: Eugenijus Kurilovas et al. If we lo
- Page 416 and 417: Models of eLearning: The Developmen
- Page 418 and 419: Stella Lee et al. Converging (AC a
- Page 420 and 421: Stella Lee et al. knowledge. Meta k
- Page 422 and 423: Stella Lee et al. Figure 3: Home pa
- Page 424 and 425: Stella Lee et al. Azevedo, R., Crom
- Page 426 and 427: Jake Leith et al. opportunities ble
Rosario Kane-Iturrioz<br />
instant feedback led to spending “many unanticipated hours in front of the computer in pursuit of the<br />
perfect score” (Segarra and Zapata, 2008). A few of the Progress Tests questions retained tutorbased<br />
feedback in order to provide students with semi-structured activities that were too complex for<br />
automatic computer feedback and required detailed explanation from the tutor. Students found the<br />
combination of computer and tutor feedback very useful in subsequent years because this method<br />
integrated the benefits of timely guidance with constructive feedback and information on how to<br />
improve their performance (Felix, 2003).<br />
The voice tools were integrated into the module’s homepage (VLE) and considered easy to use by<br />
both students and the tutor. They were used by 40% to 50% of the students because it provided<br />
opportunities to practice their oral skills through an online environment and they could also keep their<br />
audio files for future reference by downloading them onto their computers or mobile phones. For the<br />
tutor, the recorded audio files provided opportunities for comprehensive feedback. From a practical<br />
perspective, it also facilitated internal and external moderation of students’ summative oral<br />
assessments.<br />
However, voice tools for oral practice were not universally welcomed. Approximately one third of the<br />
students each year revealed reluctance about using computers to practice their oral skills. The<br />
consistent feedback from these students was that they preferred to practice oral conversation in a<br />
class-based setting with ‘real’ people. A further drawback mentioned by some students was that they<br />
did not like being overheard by other students when using voice tools on-campus. These perceptions<br />
showing a dislike by some students to computers as a medium for language <strong>learning</strong> and the<br />
discomfort felt whilst recording on campus were also key features in a study of learner drop-out<br />
reported by Stracke (2007). A larger proportion of the students only used the voice eMail tool for the<br />
oral activities contained in the Progress Tests as preparation for the summative assessment.<br />
In an earlier paper regarding learners’ experiences and perceptions of blended <strong>learning</strong> for languages<br />
(Kane-Iturrioz, 2010), it was reported that the students’ <strong>learning</strong> experience was still mainly<br />
determined by a traditional class-based model that did not foster independent <strong>learning</strong> skills. In the<br />
current study, students referred to their difficulties in adjusting to a regular weekly online workload<br />
very different from the <strong>learning</strong> pattern used in other modules where most of their study was linked to<br />
assessments and exams. This corroborates the results of Murday et al. (2008) who also report<br />
students’ problems regarding ‘self-regulation’ and its effect on timely study.<br />
The final research question sought to determine whether the summative test corroborated students’<br />
perceptions regarding the effectiveness of the module in improving their language skills. The results<br />
reported indicate that students’ perceptions of such improvements in 2008-09 seem to mirror the only<br />
change introduced in that year, that is, the reorganization of the Progress Tests. Those in 2010-11<br />
could be attributed to the enhanced feedback provided in both materials and Progress Tests and also<br />
to the change in the module’s schedule which gave students more time to develop their skills. Which<br />
of these changes contributed more to the students’ perceptions cannot be ascertained from the<br />
available data. The comparison of the trend in the mean scores and perception values seems to<br />
indicate that they closely resemble each other in their incremental direction.<br />
A minority of students expressed doubts regarding the appropriateness of an e-blended approach to<br />
language <strong>learning</strong>. However, one of these students who had little confidence in using computers<br />
stated, “it was incredibly hard ...but overall I did enjoy it and, you know, I have learnt a lot more<br />
Spanish that I thought I would do”. Some students expressed the opinion that “.. for a language<br />
module I think online is not the way”. Nevertheless, one of these found the module useful in helping<br />
him to become a better independent learner, “I think that I know where I stand as an independent<br />
learner, which really helps”.<br />
5. Conclusion<br />
Students highly appreciated the module’s assessment strategy designed to provide regular<br />
opportunities for reflection on their online <strong>learning</strong>. The range of formative feedback offered was also<br />
welcomed by students because of the combination of timeless provision of immediate computerbased<br />
comments/hints and detailed commentary from the tutor. The first type of feedback encouraged<br />
many students to engage with the online materials and also improved motivation through a better<br />
understanding of the language and the test requirements. The second type, helped students to<br />
appreciate their strengths and weaknesses and facilitated a deeper knowledge of the subject.<br />
353