RESEARCH METHOD COHEN ok

RESEARCH METHOD COHEN ok RESEARCH METHOD COHEN ok

12.01.2015 Views

COMPUTER USAGE IN CONTENT ANALYSIS 489 Diction. Some of these are reviewed by Prein et al. (1995: 190–209). These do not actually perform the analysis (in contrast to packages for quantitative data analysis) but facilitate and assist it. As Kelle (2004: 277) remarks, they do not analyse text so much as organize and structure text for subsequent analysis. These programs have the attraction of coping with large quantities of text-based material rapidly and without any risk of human error in computation and retrieval, and releasing researchers from some mechanical tasks. With respect to words, phrases, codes, nodes and categories they can: search for and return text, codes, nodes and categories filter text return counts present the grouped data according to the selection criterion desired, both within and across texts perform the qualitative equivalent of statistical analyses, such as: Boolean searches (intersections of text which have been coded by more than one code or node, using ‘and’, ‘not’ and ‘or’; looking for overlaps and co-occurrences) proximity searches (looking at clustering of data and related contextual data either side of a node or code) restrictions, trees, cross-tabs (including and excluding documents for searching, looking for codes subsumed by a particular node, and looking for nodes which subsume others) construct dendrograms (tree structures) of related nodes and codes present data in sequences and locate the text in surrounding material in order to provide the necessary context select text on combined criteria (e.g. joint occurrences, collocations) enable analyses of similarities, differences and relationships between texts and passages of text annotate text and enable memos to be written about text. Additionally, dictionaries and concordances of terms can be employed to facilitate coding, searching, retrieval and presentation. Since the rules for coding and categories are public and rule-governed, computer analysis can be particularly useful for searching, retrieving and grouping text, both in terms of specific words and in terms of words with similar meanings. Single words and word counts can overlook the importance of context. Hence computer software packages have been developed that look at Key-Words-In- Context. Most software packages have advanced functions for memoing, i.e. writing commentaries to accompany text that are not part of the original text but which may or may not be marked as incorporated material into the textual analysis. Additionally many software packages include an annotation function, which lets the researcher annotate and append text, and the annotation is kept in the text but marked as an annotation. Computers do not do away with ‘the human touch’, as humans are still needed to decide and generate the codes and categories, to verify and interpret the data. Similarly ‘there are strict limits to algorithmic interpretations of texts’ (Kelle 2004: 277), as texts contain more than that which can be examined mechanically. Further, Kelle (2004: 283) suggests that there may be problems where assumptions behind the software may not accord with those of the researchers or correspond to the researcher’s purposes, and that the software does not enable the range and richness of analytic techniques that are associated with qualitative research. Kelle (2004) argues that software may be more closely aligned to the technique of grounded theory than to other techniques (e.g. hermeneutics, discourse analysis) (Coffey et al. 1996), that it may drive the analysis rather than vice versa (Fielding and Lee 1998), and that it has a preoccupation with coding categories (Seidel and Kelle 1995). One could also argue that software does not give the same added value that one finds in quantitative data analysis, in that the textual input is a highly laborious process and that it does not perform the analysis but only supports the researcher doing the analysis by organizing data and recording codes and nodes etc. Chapter 23

490 CONTENT ANALYSIS AND GROUNDED THEORY Reliability in content analysis There are several issues to be addressed in considering the reliability of texts and their content analysis, indeed, in analysing qualitative data using a variety of means, for example: Witting and unwitting evidence (Robson 1993: 273): witting evidence is that which was intended to be imparted; unwitting evidence is that which can be inferred from the text, and which may not be intended by the imparter. The text may not have been written with the researcher in mind and may have been written for a very different purpose from that of the research (a common matter in documentary research); hence the researcher will need to know or be able to infer the intentions of the text. The documents may be limited, selective, partial, biased, non-neutral and incomplete because they were intended for a different purpose other than that of research (an issue of validity as well as of reliability). It may be difficult to infer the direction of causality in the documents – they may have been the cause or the consequence of a particular situation. Classification of text may be inconsistent (a problem sometimes mitigated by computer analysis), because of human error, coder variability (within and between coders), and ambiguity in the coding rules (Weber 1990: 17). Texts may not be corroborated or able to be corroborated. Words are inherently ambiguous and polyvalent (the problem of homographs): for example, what does the word ‘school’ mean: a building; a group of people; a particular movement of artists (e.g. the impressionist school); a department (a medical school); a noun; a verb (to drill, to induct, to educate, to train, to control, to attend an institution); a period of instructional time (‘they stayed after school to play sports’); a modifier (e.g. a school day); a sphere of activity (e.g. ‘the school of hard knocks’); a collection of people adhering to a particular set of principles (e.g. the utilitarian school); a style of life (e.g. ‘a gentleman from the old school’); a group assembled for a particular purpose (e.g. agamblingschool),andsoon.Thisisaparticular problem for computer programs which may analyse words devoid of their meaning. Coding and categorizing may lose the nuanced richness of specific words and their connotations. Category definitions and themes may be ambiguous, as they are inferential. Some words may be included in the same overall category but they may have more or less significance in that category (and a system of weighting the words may be unreliable). Words that are grouped together into a similar category may have different connotations and their usage may be more nuanced than the categories recognize. Categories may reflect the researcher’s agenda and imposition of meaning more than the text may sustain or the producers of the text (e.g. interviewees) may have intended. Aggregation may compromise reliability. Whereas sentences, phrases and words and whole documents may have the highest reliability in analysis, paragraphs and larger but incomplete portions of text have lower reliability (Weber 1990: 39). A document may deliberately exclude something for mention, overstate an issue or understate an issue (Weber 1990: 73). At a wider level, the limits of content analysis are suggested by Ezzy (2002: 84), who argues that, due to the pre-ordinate nature of coding and categorizing, content analysis is useful for testing or confirming a pre-existing theory rather than for building a new one, though this perhaps understates the ways in which content analysis can be used to generate new theory, not least through a grounded theory approach (discussed later). In many cases content analysts know in advance what they are looking for in text, and perhaps what the categories for analysis will be. Ezzy (2002: 85) suggests that this restricts the extent to which the analytical categories can be responsive to the

COMPUTER USAGE IN CONTENT ANALYSIS 489<br />

Diction. Some of these are reviewed by Prein et al.<br />

(1995: 190–209). These do not actually perform<br />

the analysis (in contrast to packages for quantitative<br />

data analysis) but facilitate and assist it. As<br />

Kelle (2004: 277) remarks, they do not analyse<br />

text so much as organize and structure text for<br />

subsequent analysis.<br />

These programs have the attraction of coping<br />

with large quantities of text-based material<br />

rapidly and without any risk of human error<br />

in computation and retrieval, and releasing<br />

researchers from some mechanical tasks. With<br />

respect to words, phrases, codes, nodes and<br />

categories they can:<br />

<br />

<br />

<br />

<br />

<br />

<br />

<br />

<br />

<br />

<br />

search for and return text, codes, nodes and<br />

categories<br />

filter text<br />

return counts<br />

present the grouped data according to the<br />

selection criterion desired, both within and<br />

across texts<br />

perform the qualitative equivalent of statistical<br />

analyses, such as:<br />

Boolean searches (intersections of text<br />

which have been coded by more than one<br />

code or node, using ‘and’, ‘not’ and ‘or’;<br />

lo<strong>ok</strong>ing for overlaps and co-occurrences)<br />

proximity searches (lo<strong>ok</strong>ing at clustering of<br />

data and related contextual data either side<br />

of a node or code)<br />

restrictions, trees, cross-tabs (including and<br />

excluding documents for searching, lo<strong>ok</strong>ing<br />

for codes subsumed by a particular node,<br />

and lo<strong>ok</strong>ing for nodes which subsume<br />

others)<br />

construct dendrograms (tree structures) of<br />

related nodes and codes<br />

present data in sequences and locate the text<br />

in surrounding material in order to provide the<br />

necessary context<br />

select text on combined criteria (e.g. joint<br />

occurrences, collocations)<br />

enable analyses of similarities, differences and<br />

relationships between texts and passages of text<br />

annotate text and enable memos to be written<br />

about text.<br />

Additionally, dictionaries and concordances of<br />

terms can be employed to facilitate coding,<br />

searching, retrieval and presentation.<br />

Since the rules for coding and categories are<br />

public and rule-governed, computer analysis can<br />

be particularly useful for searching, retrieving and<br />

grouping text, both in terms of specific words and in<br />

terms of words with similar meanings. Single words<br />

and word counts can overlo<strong>ok</strong> the importance<br />

of context. Hence computer software packages<br />

have been developed that lo<strong>ok</strong> at Key-Words-In-<br />

Context. Most software packages have advanced<br />

functions for memoing, i.e. writing commentaries<br />

to accompany text that are not part of the original<br />

text but which may or may not be marked<br />

as incorporated material into the textual analysis.<br />

Additionally many software packages include<br />

an annotation function, which lets the researcher<br />

annotate and append text, and the annotation is<br />

kept in the text but marked as an annotation.<br />

Computers do not do away with ‘the human<br />

touch’, as humans are still needed to decide and<br />

generate the codes and categories, to verify and<br />

interpret the data. Similarly ‘there are strict limits<br />

to algorithmic interpretations of texts’ (Kelle 2004:<br />

277), as texts contain more than that which can<br />

be examined mechanically. Further, Kelle (2004:<br />

283) suggests that there may be problems where<br />

assumptions behind the software may not accord<br />

with those of the researchers or correspond to<br />

the researcher’s purposes, and that the software<br />

does not enable the range and richness of analytic<br />

techniques that are associated with qualitative<br />

research. Kelle (2004) argues that software may<br />

be more closely aligned to the technique of<br />

grounded theory than to other techniques (e.g.<br />

hermeneutics, discourse analysis) (Coffey et al.<br />

1996), that it may drive the analysis rather than<br />

vice versa (Fielding and Lee 1998), and that it has<br />

a preoccupation with coding categories (Seidel<br />

and Kelle 1995). One could also argue that<br />

software does not give the same added value that<br />

one finds in quantitative data analysis, in that the<br />

textual input is a highly laborious process and that<br />

it does not perform the analysis but only supports<br />

the researcher doing the analysis by organizing<br />

data and recording codes and nodes etc.<br />

Chapter 23

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!