learning - Academic Conferences Limited
learning - Academic Conferences Limited learning - Academic Conferences Limited
Rosario Kane-Iturrioz Tests very useful as a self-assessment facility to check their progress. A breakdown of students’ perception by year can be seen in Table 1. Table 1: Students’ perceptions on the usefulness of the progress tests Statement: I have found the Progress Tests useful to keep track of how much I have learnt online. 2007-08 (N=13) 2008-09 (N= 31) 2009-10 (N=28) 2010-11 (N=17) Strongly agree 3 (23.08%) 12 (38.71%) 11 (39.29%) 5 (29.41%) Agree 10 (76.92%) 15 (48.39%) 15 (53.57%) 10 (58.83%) Neutral 0 (0.0%) 4 (12.90% 1 (3.57%) 1 (5.88%) Disagree 0 (0.0%) 0 (0.0%) 1 (3.57%) 1 (5.88%) Strongly disagree 0 (0.0%) 0 (0.0%) 0 (0.0%) 0 (0.0%) In 2008-09, students were also asked to select which aspect they liked most about the module’s online provision from a list of ten items. Progress Tests were the students third favourite preference (63%) after “work at my own pace” and “multiple attempts allowed”. The open question (questionnaires) and interviews (2008-09) corroborated these results and many students stated that they “would rather do the Progress Test than the voice mail or discussion”. Others liked this tool because it “...help(s) us to revise as well as prepare for the face-to-face sessions.” or, “I could see how I got to learn more really and you could always go back”. Although many students associated the Progress Test with an improvement in their independent learning skills: “..this experience was good and I have got a lot of advantage out of it, like self-discipline and time management.”; they also indicated that they had faced difficulties: “I ran out of time and so my main aim was to learn the topic material” or “there was nobody pushing me”. In one case the problem was fear of failure “Didn’t do most of them (Progress Tests) because wasn’t confident enough”. One of the aims of the Progress Test was to familiarise students with the module’s summative assessment format. A computer-based test was chosen in order to be consistent with the module’s main delivery mode. Students were asked about their perceptions regarding the computer-based summative assessment at two points during the four years of this study, 2008-09 and 2010-11. The students’ answers indicate a clear change in perceptions in the intervening year. 52% and 74% respectively “found the computer-based assessment better than the traditional assessment (including face-to-face oral tests). The breakdown of the results can be seen in Table 2. Table 2: Students’ perceptions of computer-based assessments Statement: “I have found the computer-based assessment better than the traditional assessment (including face-toface oral tests). 2008-09 (N = 31) 2010-11 (N =19) Strongly agree 6 (19.35%) 5 (26.31%) Agree 10 (32.26%) 9 (47.37%) Neutral 9 (29.03% 3 (15.79%) Disagree 4 (12.90%) 2 (10.53%) Strongly disagree 2 (12.90%) 0 (0.0%) 3.3 Students’ perceptions about formative feedback In 2008-09 and 2010-11, students were asked to select which aspect they preferred most about the module’s online coursework from a list of ten items. Immediate computer feedback was the 6 th favourite preference in 2008-09 and the third in 2010-11 after “work at my own pace” and “variety of activities”. Many students referred to the feedback provided for activities and exercises linking them to positive effects on enjoyment: “That is the bit I enjoy the most in the module, the whole interactiveness of it, it made it seem like fun”. Others reflected on the perceived motivational help provided by the activities feedback (mainly computer-based): “ It feels you being there with me’’. This statement seems to indicate that a comprehensive feedback is seen by students as an indication of ‘caring behaviour’ on the part of the tutor. The Progress Test feedback both computer-based and tutor provided was in general perceived as highly useful (84.62%) and motivating: “What really got me motivated was getting the results straight back; to have a look through at what the mistakes were so it helps us”. A few students would have 350
Rosario Kane-Iturrioz preferred all the feedback computer-based because of the perceived effect of the immediacy of a response, “with an immediate answer (I) maybe do an extra half and hour”. Other students preferred the mixture of computer-based and tutor feedback as being very appropriate to prepare them for the summative assessments. The use of Wimba voice tools for oral practice was received with mixed reactions. The students more familiar with technology seemed to welcome the use of state-of-the-art tools, “Also I like the Wimba voice mail, is quite new technology, I can record my speaking, listen it, if I don’t like it, I can record again”. Other students found this tool interesting and fun: “The Wimba… I find it really interesting; it was fun as well because there were things I have never used before”. Although appreciating the benefits of Wimba’s voice tools, approximately 50% of the students did not use this resource very often for a number of reasons. Sometimes microphones were unavailable on campus or the students did not like practising on campus with other people listening to them. A larger proportion of the students (60% to 70%) submitted the voice eMails linked to the oral activities in the Progress Tests. Further insights came from the questionnaires’ open questions in which many students stated that they only engaged in activities that had a direct relevance to the summative assessments. The audio and text feedback provided via voice mail technology was well received by those students who submitted the recordings corresponding to the oral activities. There were two aspects highly valued by students, the first was a greater effectiveness of the feedback due to the synergies achieved in the integration of audio and text commentary, “The written feedback was helpful but the voiceMail feedback help me to understand better were I went wrong with my pronunciation”. Students reported that the audio feedback had reinforced the written feedback because it was more detailed and personalised. The second benefit referred to the effective delivery of feedback comments, “The feedback for my voice eMail submissions was great as I received it straight into my mail inbox”. 3.4 Language skills Students’ perceptions about the module being effective in increasing their language skills in the first two years and in the final year of the study can be seen in Figure 4. This figure shows that the mean values for Reading, Listening and Oral skills increased substantially in 2008-09 and again in 2010 in the case of Reading. The highest values recorded corresponded to Writing and these values remained high in all three years. Mean Likert values (1-5) 2.5 2 1.5 1 0.5 0 Reading Listening Writing Oral Figure 4: Perceptions of improvements in language skills 351 2007-08 2008-09 2010-11
- Page 326 and 327: Danny Glick and Roni Aviram opinion
- Page 328 and 329: Danny Glick and Roni Aviram Bernard
- Page 330 and 331: Andrea Gorra and Ollie Jones to hel
- Page 332 and 333: Andrea Gorra and Ollie Jones Howeve
- Page 334 and 335: Andrea Gorra and Ollie Jones Figure
- Page 336 and 337: Andrea Gorra and Ollie Jones Studen
- Page 338 and 339: Rose Heaney and Megan Anne Arroll A
- Page 340 and 341: Rose Heaney and Megan Anne Arroll l
- Page 342 and 343: Rose Heaney and Megan Anne Arroll
- Page 344 and 345: Rose Heaney and Megan Anne Arroll J
- Page 346 and 347: Amanda Jefferies learning was furth
- Page 348 and 349: Amanda Jefferies way in which the o
- Page 350 and 351: Amanda Jefferies their teaching mat
- Page 352 and 353: A Methodology for Incorporating Usa
- Page 354 and 355: Anne Jelfs and Chetz Colwell To try
- Page 356 and 357: Anne Jelfs and Chetz Colwell We wor
- Page 358 and 359: The Virtual Learning Environment -
- Page 360 and 361: John Jessel 2.1 An outline framewor
- Page 362 and 363: John Jessel teachers who agreed to
- Page 364 and 365: John Jessel ‘“reduce the clicks
- Page 366 and 367: Mutlimodal Teaching Through ICT Edu
- Page 368 and 369: Paraskevi Kanari and Georgios Potam
- Page 370 and 371: Paraskevi Kanari and Georgios Potam
- Page 372 and 373: Rosario Kane-Iturrioz Regarding lan
- Page 374 and 375: Rosario Kane-Iturrioz Figure 2: Exa
- Page 378 and 379: Rosario Kane-Iturrioz When compared
- Page 380 and 381: Rosario Kane-Iturrioz Although the
- Page 382 and 383: Jana Kapounova et al. eLearning is
- Page 384 and 385: Jana Kapounova et al. Each dimensio
- Page 386 and 387: Jana Kapounova et al. project, conn
- Page 388 and 389: Acknowledgments Jana Kapounova et a
- Page 390 and 391: Andrea Kelz skills and competences
- Page 392 and 393: Andrea Kelz web-based activities in
- Page 394 and 395: Andrea Kelz system. Most other univ
- Page 396 and 397: Open Courses: The Next big Thing in
- Page 398 and 399: Kaido Kikkas et al. However, in the
- Page 400 and 401: Kaido Kikkas et al. generation of w
- Page 402 and 403: Kaido Kikkas et al. Occasional gue
- Page 404 and 405: John Knight and Rebecca Rochon guid
- Page 406 and 407: Evaluation of Quality of Learning S
- Page 408 and 409: Eugenijus Kurilovas et al. Essalmi
- Page 410 and 411: Eugenijus Kurilovas et al. (LOs), l
- Page 412 and 413: Eugenijus Kurilovas et al. Then hie
- Page 414 and 415: Eugenijus Kurilovas et al. If we lo
- Page 416 and 417: Models of eLearning: The Developmen
- Page 418 and 419: Stella Lee et al. Converging (AC a
- Page 420 and 421: Stella Lee et al. knowledge. Meta k
- Page 422 and 423: Stella Lee et al. Figure 3: Home pa
- Page 424 and 425: Stella Lee et al. Azevedo, R., Crom
Rosario Kane-Iturrioz<br />
Tests very useful as a self-assessment facility to check their progress. A breakdown of students’<br />
perception by year can be seen in Table 1.<br />
Table 1: Students’ perceptions on the usefulness of the progress tests<br />
Statement: I have found the Progress Tests useful to keep<br />
track of how much I have learnt online.<br />
2007-08<br />
(N=13)<br />
2008-09<br />
(N= 31)<br />
2009-10<br />
(N=28)<br />
2010-11 (N=17)<br />
Strongly agree 3 (23.08%) 12 (38.71%) 11 (39.29%) 5 (29.41%)<br />
Agree 10 (76.92%) 15 (48.39%) 15 (53.57%) 10 (58.83%)<br />
Neutral 0 (0.0%) 4 (12.90% 1 (3.57%) 1 (5.88%)<br />
Disagree 0 (0.0%) 0 (0.0%) 1 (3.57%) 1 (5.88%)<br />
Strongly disagree 0 (0.0%) 0 (0.0%) 0 (0.0%) 0 (0.0%)<br />
In 2008-09, students were also asked to select which aspect they liked most about the module’s<br />
online provision from a list of ten items. Progress Tests were the students third favourite preference<br />
(63%) after “work at my own pace” and “multiple attempts allowed”. The open question<br />
(questionnaires) and interviews (2008-09) corroborated these results and many students stated that<br />
they “would rather do the Progress Test than the voice mail or discussion”. Others liked this tool<br />
because it “...help(s) us to revise as well as prepare for the face-to-face sessions.” or, “I could see<br />
how I got to learn more really and you could always go back”. Although many students associated the<br />
Progress Test with an improvement in their independent <strong>learning</strong> skills: “..this experience was good<br />
and I have got a lot of advantage out of it, like self-discipline and time management.”; they also<br />
indicated that they had faced difficulties: “I ran out of time and so my main aim was to learn the topic<br />
material” or “there was nobody pushing me”. In one case the problem was fear of failure “Didn’t do<br />
most of them (Progress Tests) because wasn’t confident enough”.<br />
One of the aims of the Progress Test was to familiarise students with the module’s summative<br />
assessment format. A computer-based test was chosen in order to be consistent with the module’s<br />
main delivery mode. Students were asked about their perceptions regarding the computer-based<br />
summative assessment at two points during the four years of this study, 2008-09 and 2010-11. The<br />
students’ answers indicate a clear change in perceptions in the intervening year. 52% and 74%<br />
respectively “found the computer-based assessment better than the traditional assessment (including<br />
face-to-face oral tests). The breakdown of the results can be seen in Table 2.<br />
Table 2: Students’ perceptions of computer-based assessments<br />
Statement: “I have found the computer-based assessment better<br />
than the traditional assessment (including face-toface<br />
oral tests).<br />
2008-09 (N = 31) 2010-11 (N =19)<br />
Strongly agree 6 (19.35%) 5 (26.31%)<br />
Agree 10 (32.26%) 9 (47.37%)<br />
Neutral 9 (29.03% 3 (15.79%)<br />
Disagree 4 (12.90%) 2 (10.53%)<br />
Strongly disagree 2 (12.90%) 0 (0.0%)<br />
3.3 Students’ perceptions about formative feedback<br />
In 2008-09 and 2010-11, students were asked to select which aspect they preferred most about the<br />
module’s online coursework from a list of ten items. Immediate computer feedback was the 6 th<br />
favourite preference in 2008-09 and the third in 2010-11 after “work at my own pace” and “variety of<br />
activities”. Many students referred to the feedback provided for activities and exercises linking them to<br />
positive effects on enjoyment: “That is the bit I enjoy the most in the module, the whole<br />
interactiveness of it, it made it seem like fun”. Others reflected on the perceived motivational help<br />
provided by the activities feedback (mainly computer-based): “ It feels you being there with me’’. This<br />
statement seems to indicate that a comprehensive feedback is seen by students as an indication of<br />
‘caring behaviour’ on the part of the tutor.<br />
The Progress Test feedback both computer-based and tutor provided was in general perceived as<br />
highly useful (84.62%) and motivating: “What really got me motivated was getting the results straight<br />
back; to have a look through at what the mistakes were so it helps us”. A few students would have<br />
350