RESEARCH METHOD COHEN ok

RESEARCH METHOD COHEN ok RESEARCH METHOD COHEN ok

12.01.2015 Views

RESEARCH AND EVALUATION 41 particular experiences. In response to this, some feminist researchers (p. 40) suggest that researchers only have the warrant to confine themselves to their own immediate communities, though this is acontentiousissue.Thereisvalueinspeakingfor others, not least for those who are silenced and marginalized, and in not speaking for others for fear of oppression and colonization. One has to question the acceptability and appropriateness of, and fidelity to, the feminist ethic, if one represents and uses others’ stories (p. 41). An example of a feminist approach to research is the Girls Into Science and Technology (GIST) action research project. This took place over three years, involving 2,000 students and their teachers in ten coeducational, comprehensive schools in one area of the United Kingdom, eight schools serving as the bases of the ‘action’, the remaining two acting as ‘controls’. Several publications have documented the methodologies and findings of the GIST study (Kelly 1986; 1989a; 1989b; Kelly and Smail 1986; Whyte 1986), described by its co-director as ‘simultaneous-integrated action research’ (Kelly 1987) (i.e. integrating action and research). Kelly is open about the feminist orientation of the GIST project team, seeking deliberately to change girls’ option choices and career aspirations, because the researchers saw that girls were disadvantaged by traditional sexstereotypes. The researchers’ actions, she suggests, were a small attempt to ameliorate women’s subordinate social position (Kelly 1987). case of ‘categorically funded’ and commissioned research – research which is funded by policymakers (e.g. governments, fund-awarding bodies) under any number of different headings that those policy-makers devise (Burgess 1993). On the one hand, this is laudable, for it targets research directly towards policy; on the other hand, it is dangerous in that it enables others to set the research agenda. Research ceases to become open-ended, pure research, and, instead, becomes the evaluation of given initiatives. Less politically charged, much research is evaluative, and indeed there are many similarities between research and evaluation. The two overlap but possess important differences. The problem of trying to identify differences between evaluation and research is compounded because not only do they share several of the same methodological characteristics but also one branch of research is called evaluative research or applied research. This is often kept separate from ‘blue skies’ research in that the latter is open-ended, exploratory, contributes something original to the substantive field and extends the frontiers of knowledge and theory whereas in the former the theory is given rather than interrogated or tested. Onecandetect many similarities between the two in that they both use methodologies and methods of social science research generally, covering, for example (see http://www.routledge.com/textbooks/ 9780415368780 – Chapter 1, file 1.9. ppt), the following: Chapter 1 Research and evaluation The preceding discussion has suggested that research and politics are inextricably bound together. This can be taken further, as researchers in education will be advised to pay serious consideration to the politics of their research enterprise and the ways in which politics can steer research. For example, one can detect a trend in educational research towards more evaluative research, where, for example, a researcher’s task is to evaluate the effectiveness (often of the implementation) of given policies and projects. This is particularly true in the the need to clarify the purposes of the investigation the need to operationalize purposes and areas of investigation the need to address principles of research design that include: formulating operational questions deciding appropriate methodologies deciding which instruments to use for data collection deciding on the sample for the investigation addressing reliability and validity in the investigation and instrumentation

42 THE NATURE OF INQUIRY addressing ethical issues in conducting the investigation deciding on data analysis techniques deciding on reporting and interpreting results. Indeed Norris (1990) argues that evaluation applies research methods to shed light on a problem of action (Norris 1990: 97); he suggests that evaluation can be viewed as an extension of research, because it shares its methodologies and methods, and because evaluators and researchers possess similar skills in conducting investigations (see http://www.routledge.com/textbooks/ 9780415368780 – Chapter 1, file 1.10. ppt). In many senses the eight features outlined above embrace many elements of the scientific method, which Smith and Glass (1987) set out in seven steps: 1 A theory about the phenomenon exists. 2 A research problem within the theory is detected and a research question is devised. 3 A research hypothesis is deduced (often about the relationship between constructs). 4 A research design is developed, operationalizing the research question and stating the null hypothesis. 5 The research is conducted. 6 The null hypothesis is tested based on the data gathered. 7 The original theory is revised or supported based on the results of the hypothesis testing. Indeed, if steps 1 and 7 were removed then there would be nothing to distinguish between research and evaluation. Both researchers and evaluators pose questions and hypotheses, select samples, manipulate and measure variables, compute statistics and data, and state conclusions. Nevertheless there are important differences between evaluation and research that are not always obvious simply by looking at publications. Publications do not always make clear the background events that gave rise to the investigation, nor do they always make clear the uses of the material that they report, nor do they always make clear what the dissemination rights (Sanday 1993) are and who holds them. Several commentators set out some of the differences between evaluation and research. For example Smith and Glass (1987) offer eight main differences (see http://www.routledge.com/textbooks/ 9780415368780 – Chapter 1, file 1.11. ppt): The intents and purposes of the investigation: the researcher wants to advance the frontiers of knowledge of phenomena, to contribute to theory and to be able to make generalizations; the evaluator is less interested in contributing to theory or the general body of knowledge. Evaluation is more parochial than universal (Smith and Glass 1987: 33–4). The scope of the investigation: evaluation studies tend to be more comprehensive than research in the number and variety of aspects of a programme that are being studied (p. 34). Values in the investigation: research aspires to value neutrality, evaluations must represent multiple sets of values and include data on these values. The origins of the study: researchhasitsorigins and motivation in the researcher’s curiosity and desire to know (p. 34). The researcher is answerable to colleagues and scientists (i.e. the research community) whereas the evaluator is answerable to the ‘client’. The researcher is autonomous whereas the evaluator is answerable to clients and stakeholders. The researcher is motivated by a search for knowledge, the evaluator is motivated by the need to solve problems, allocate resources and make decisions. Research studies are public, evaluations are for a restricted audience. The uses of the study: theresearchisusedto further knowledge, evaluations are used to inform decisions. The timeliness of the study: evaluationsmustbe timely, research need not be. Evaluators’ time scales are given, researchers’ time scales need not be given. Criteria for judging the study: evaluationsare judged by the criteria of utility and credibility, research is judged methodologically and by the contribution that it makes to the field (i.e. internal and external validity).

<strong>RESEARCH</strong> AND EVALUATION 41<br />

particular experiences. In response to this, some<br />

feminist researchers (p. 40) suggest that researchers<br />

only have the warrant to confine themselves to<br />

their own immediate communities, though this is<br />

acontentiousissue.Thereisvalueinspeakingfor<br />

others, not least for those who are silenced and<br />

marginalized, and in not speaking for others for<br />

fear of oppression and colonization. One has to<br />

question the acceptability and appropriateness of,<br />

and fidelity to, the feminist ethic, if one represents<br />

and uses others’ stories (p. 41).<br />

An example of a feminist approach to research<br />

is the Girls Into Science and Technology (GIST)<br />

action research project. This to<strong>ok</strong> place over three<br />

years, involving 2,000 students and their teachers<br />

in ten coeducational, comprehensive schools in<br />

one area of the United Kingdom, eight schools<br />

serving as the bases of the ‘action’, the remaining<br />

two acting as ‘controls’. Several publications have<br />

documented the methodologies and findings of<br />

the GIST study (Kelly 1986; 1989a; 1989b; Kelly<br />

and Smail 1986; Whyte 1986), described by<br />

its co-director as ‘simultaneous-integrated action<br />

research’ (Kelly 1987) (i.e. integrating action<br />

and research). Kelly is open about the feminist<br />

orientation of the GIST project team, seeking<br />

deliberately to change girls’ option choices and<br />

career aspirations, because the researchers saw<br />

that girls were disadvantaged by traditional sexstereotypes.<br />

The researchers’ actions, she suggests,<br />

were a small attempt to ameliorate women’s<br />

subordinate social position (Kelly 1987).<br />

case of ‘categorically funded’ and commissioned<br />

research – research which is funded by policymakers<br />

(e.g. governments, fund-awarding bodies)<br />

under any number of different headings that<br />

those policy-makers devise (Burgess 1993). On<br />

the one hand, this is laudable, for it targets<br />

research directly towards policy; on the other<br />

hand, it is dangerous in that it enables others<br />

to set the research agenda. Research ceases to<br />

become open-ended, pure research, and, instead,<br />

becomes the evaluation of given initiatives. Less<br />

politically charged, much research is evaluative,<br />

and indeed there are many similarities between<br />

research and evaluation. The two overlap but<br />

possess important differences. The problem of<br />

trying to identify differences between evaluation<br />

and research is compounded because not only do<br />

they share several of the same methodological<br />

characteristics but also one branch of research is<br />

called evaluative research or applied research. This<br />

is often kept separate from ‘blue skies’ research<br />

in that the latter is open-ended, exploratory,<br />

contributes something original to the substantive<br />

field and extends the frontiers of knowledge and<br />

theory whereas in the former the theory is given<br />

rather than interrogated or tested. Onecandetect<br />

many similarities between the two in that they<br />

both use methodologies and methods of social<br />

science research generally, covering, for example<br />

(see http://www.routledge.com/textbo<strong>ok</strong>s/<br />

9780415368780 – Chapter 1, file 1.9. ppt), the<br />

following:<br />

Chapter 1<br />

Research and evaluation<br />

The preceding discussion has suggested that<br />

research and politics are inextricably bound<br />

together. This can be taken further, as researchers<br />

in education will be advised to pay serious<br />

consideration to the politics of their research<br />

enterprise and the ways in which politics can<br />

steer research. For example, one can detect<br />

a trend in educational research towards more<br />

evaluative research, where, for example, a<br />

researcher’s task is to evaluate the effectiveness<br />

(often of the implementation) of given policies<br />

and projects. This is particularly true in the<br />

the need to clarify the purposes of the<br />

investigation<br />

the need to operationalize purposes and areas of<br />

investigation<br />

the need to address principles of research design<br />

that include:<br />

formulating operational questions<br />

deciding appropriate methodologies<br />

deciding which instruments to use for data<br />

collection<br />

deciding on the sample for the investigation<br />

addressing reliability and validity in the<br />

investigation and instrumentation

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!