01 NRDC Dyslexia 1-88 update - Texthelp
01 NRDC Dyslexia 1-88 update - Texthelp
01 NRDC Dyslexia 1-88 update - Texthelp
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
80<br />
Research Report<br />
devising. There might be assignment bias, which can occur when members of the intervention<br />
group differ from controls in ways that predict a better response to the intervention, perhaps<br />
by having higher motivation or greater ability. Although there are procedures to avoid these<br />
kinds of bias, they are not always followed in intervention studies.<br />
A fourth group of challenges is associated with shortcomings in the administration of the<br />
intervention, such as insufficient treatment duration, differential withdrawal from the<br />
programme on the part of successful and unsuccessful participants, or lack of follow-up to<br />
determine whether the intervention has had a lasting effect. Inappropriate statistical<br />
analyses, especially those which violate the assumptions of the tests used, are not<br />
uncommon. These violations can suggest that differences in outcome are statistically<br />
significant when they could have occurred by chance. Alternatively, differences in outcome<br />
might be statistically significant but trivial to the policy-maker.<br />
A further question can be asked of any apparently-effective intervention, namely: ‘Will it be<br />
effective for others?’ This is an especially pertinent question to ask of the lateral visual<br />
masking study of ‘four severe adult dyslexics’ by Geiger et al. (1992). Even if the initial<br />
diagnosis was accurate (and if the intervention addressed the participants’ problem<br />
successfully), no finding from such a small-scale study can be generalised with confidence to<br />
a larger population. For future researchers, studies like this one can be important, but for<br />
policy-makers they offer little guidance.<br />
A different question can be asked of the study of thirty college students (Guyer, Banks, &<br />
Guyer, 1993): ‘Is the superior intervention the most effective one available?’ One of the two<br />
interventions in the comparison was an explicit, structured remedial spelling programme<br />
using analytic and synthetic phonics. The comparison programme ‘taught spelling using a<br />
non-phonetic approach’. In effect, the programmes were horses for different courses, one for<br />
regularly-spelled words and one for irregular words. The outcome was that the ‘regular’ horse<br />
ran faster than the ‘irregular’ horse, since most spellings are regular. Readers of a specialist<br />
dyslexia journal might be disposed to believe that the ‘regular’ horse (in the form of an Orton-<br />
Gillingham programme) would have won the gold cup in any comparison, but in this<br />
comparison the race was fixed and the finding offers no answer to the question.<br />
There are many pitfalls in evaluation research.<br />
How can we know what really works?<br />
What kinds of intervention study are likely to offer reliable guidance to tutors? This is not an<br />
idle question. The design of intervention studies is exacting in any circumstances; it is<br />
especially so when the participants are hard to reach. Perhaps for these reasons, there are<br />
many design limitations in existing reading intervention research (Abadzi, 1994; Beder, 1999;<br />
Lyon & Moats, 1997; Simmerman & Swanson, 20<strong>01</strong>; Torgerson et al., 2002).<br />
Nevertheless, a good intervention study should satisfy a number of criteria. They include:<br />
■<br />
■<br />
■<br />
a focused research question, specifying a theoretically plausible mechanism of change;<br />
a design that permits investigators to minimise the effects of extraneous variables, such as<br />
teacher effects and differences in motivation;<br />
a large enough sample to reduce the possibility that the outcomes of the intervention might<br />
have occurred by chance;