RESEARCH METHOD COHEN ok
RESEARCH METHOD COHEN ok RESEARCH METHOD COHEN ok
COMPUTER USAGE IN CONTENT ANALYSIS 489 Diction. Some of these are reviewed by Prein et al. (1995: 190–209). These do not actually perform the analysis (in contrast to packages for quantitative data analysis) but facilitate and assist it. As Kelle (2004: 277) remarks, they do not analyse text so much as organize and structure text for subsequent analysis. These programs have the attraction of coping with large quantities of text-based material rapidly and without any risk of human error in computation and retrieval, and releasing researchers from some mechanical tasks. With respect to words, phrases, codes, nodes and categories they can: search for and return text, codes, nodes and categories filter text return counts present the grouped data according to the selection criterion desired, both within and across texts perform the qualitative equivalent of statistical analyses, such as: Boolean searches (intersections of text which have been coded by more than one code or node, using ‘and’, ‘not’ and ‘or’; looking for overlaps and co-occurrences) proximity searches (looking at clustering of data and related contextual data either side of a node or code) restrictions, trees, cross-tabs (including and excluding documents for searching, looking for codes subsumed by a particular node, and looking for nodes which subsume others) construct dendrograms (tree structures) of related nodes and codes present data in sequences and locate the text in surrounding material in order to provide the necessary context select text on combined criteria (e.g. joint occurrences, collocations) enable analyses of similarities, differences and relationships between texts and passages of text annotate text and enable memos to be written about text. Additionally, dictionaries and concordances of terms can be employed to facilitate coding, searching, retrieval and presentation. Since the rules for coding and categories are public and rule-governed, computer analysis can be particularly useful for searching, retrieving and grouping text, both in terms of specific words and in terms of words with similar meanings. Single words and word counts can overlook the importance of context. Hence computer software packages have been developed that look at Key-Words-In- Context. Most software packages have advanced functions for memoing, i.e. writing commentaries to accompany text that are not part of the original text but which may or may not be marked as incorporated material into the textual analysis. Additionally many software packages include an annotation function, which lets the researcher annotate and append text, and the annotation is kept in the text but marked as an annotation. Computers do not do away with ‘the human touch’, as humans are still needed to decide and generate the codes and categories, to verify and interpret the data. Similarly ‘there are strict limits to algorithmic interpretations of texts’ (Kelle 2004: 277), as texts contain more than that which can be examined mechanically. Further, Kelle (2004: 283) suggests that there may be problems where assumptions behind the software may not accord with those of the researchers or correspond to the researcher’s purposes, and that the software does not enable the range and richness of analytic techniques that are associated with qualitative research. Kelle (2004) argues that software may be more closely aligned to the technique of grounded theory than to other techniques (e.g. hermeneutics, discourse analysis) (Coffey et al. 1996), that it may drive the analysis rather than vice versa (Fielding and Lee 1998), and that it has a preoccupation with coding categories (Seidel and Kelle 1995). One could also argue that software does not give the same added value that one finds in quantitative data analysis, in that the textual input is a highly laborious process and that it does not perform the analysis but only supports the researcher doing the analysis by organizing data and recording codes and nodes etc. Chapter 23
490 CONTENT ANALYSIS AND GROUNDED THEORY Reliability in content analysis There are several issues to be addressed in considering the reliability of texts and their content analysis, indeed, in analysing qualitative data using a variety of means, for example: Witting and unwitting evidence (Robson 1993: 273): witting evidence is that which was intended to be imparted; unwitting evidence is that which can be inferred from the text, and which may not be intended by the imparter. The text may not have been written with the researcher in mind and may have been written for a very different purpose from that of the research (a common matter in documentary research); hence the researcher will need to know or be able to infer the intentions of the text. The documents may be limited, selective, partial, biased, non-neutral and incomplete because they were intended for a different purpose other than that of research (an issue of validity as well as of reliability). It may be difficult to infer the direction of causality in the documents – they may have been the cause or the consequence of a particular situation. Classification of text may be inconsistent (a problem sometimes mitigated by computer analysis), because of human error, coder variability (within and between coders), and ambiguity in the coding rules (Weber 1990: 17). Texts may not be corroborated or able to be corroborated. Words are inherently ambiguous and polyvalent (the problem of homographs): for example, what does the word ‘school’ mean: a building; a group of people; a particular movement of artists (e.g. the impressionist school); a department (a medical school); a noun; a verb (to drill, to induct, to educate, to train, to control, to attend an institution); a period of instructional time (‘they stayed after school to play sports’); a modifier (e.g. a school day); a sphere of activity (e.g. ‘the school of hard knocks’); a collection of people adhering to a particular set of principles (e.g. the utilitarian school); a style of life (e.g. ‘a gentleman from the old school’); a group assembled for a particular purpose (e.g. agamblingschool),andsoon.Thisisaparticular problem for computer programs which may analyse words devoid of their meaning. Coding and categorizing may lose the nuanced richness of specific words and their connotations. Category definitions and themes may be ambiguous, as they are inferential. Some words may be included in the same overall category but they may have more or less significance in that category (and a system of weighting the words may be unreliable). Words that are grouped together into a similar category may have different connotations and their usage may be more nuanced than the categories recognize. Categories may reflect the researcher’s agenda and imposition of meaning more than the text may sustain or the producers of the text (e.g. interviewees) may have intended. Aggregation may compromise reliability. Whereas sentences, phrases and words and whole documents may have the highest reliability in analysis, paragraphs and larger but incomplete portions of text have lower reliability (Weber 1990: 39). A document may deliberately exclude something for mention, overstate an issue or understate an issue (Weber 1990: 73). At a wider level, the limits of content analysis are suggested by Ezzy (2002: 84), who argues that, due to the pre-ordinate nature of coding and categorizing, content analysis is useful for testing or confirming a pre-existing theory rather than for building a new one, though this perhaps understates the ways in which content analysis can be used to generate new theory, not least through a grounded theory approach (discussed later). In many cases content analysts know in advance what they are looking for in text, and perhaps what the categories for analysis will be. Ezzy (2002: 85) suggests that this restricts the extent to which the analytical categories can be responsive to the
- Page 458 and 459: PROCEDURES IN GRID ANALYSIS 439 inv
- Page 460 and 461: PROCEDURES IN GRID ANALYSIS 441 Box
- Page 462 and 463: SOME EXAMPLES OF THE USE OF REPERTO
- Page 464 and 465: GRID TECHNIQUE AND AUDIO/VIDEO LESS
- Page 466 and 467: FOCUSED GRIDS, NON-VERBAL GRIDS, EX
- Page 468 and 469: INTRODUCTION 449 Box 21.1 Dimension
- Page 470 and 471: ROLE-PLAYING VERSUS DECEPTION: THE
- Page 472 and 473: THE USES OF ROLE-PLAYING 453 of
- Page 474 and 475: ROLE-PLAYING IN AN EDUCATIONAL SETT
- Page 476: EVALUATING ROLE-PLAYING AND OTHER S
- Page 480 and 481: 22 Approaches to qualitative data a
- Page 482 and 483: TABULATING DATA 463 data set reprod
- Page 484 and 485: TABULATING DATA 465 Box 22.4 Studen
- Page 486 and 487: FIVE WAYS OF ORGANIZING AND PRESENT
- Page 488 and 489: SYSTEMATIC APPROACHES TO DATA ANALY
- Page 490 and 491: SYSTEMATIC APPROACHES TO DATA ANALY
- Page 492 and 493: METHODOLOGICAL TOOLS FOR ANALYSING
- Page 494 and 495: 23 Content analysis and grounded th
- Page 496 and 497: HOW DOES CONTENT ANALYSIS WORK 477
- Page 498 and 499: HOW DOES CONTENT ANALYSIS WORK 479
- Page 500 and 501: HOW DOES CONTENT ANALYSIS WORK 481
- Page 502 and 503: A WORKED EXAMPLE OF CONTENT ANALYSI
- Page 504 and 505: A WORKED EXAMPLE OF CONTENT ANALYSI
- Page 506 and 507: COMPUTER USAGE IN CONTENT ANALYSIS
- Page 510 and 511: GROUNDED THEORY 491 data, thereby c
- Page 512 and 513: GROUNDED THEORY 493 fragments are t
- Page 514 and 515: INTERPRETATION IN QUALITATIVE DATA
- Page 516 and 517: INTERPRETATION IN QUALITATIVE DATA
- Page 518 and 519: INTERPRETATION IN QUALITATIVE DATA
- Page 520 and 521: 24 Quantitative data analysis Intro
- Page 522 and 523: DESCRIPTIVE AND INFERENTIAL STATIST
- Page 524 and 525: DEPENDENT AND INDEPENDENT VARIABLES
- Page 526 and 527: EXPLORATORY DATA ANALYSIS: FREQUENC
- Page 528 and 529: EXPLORATORY DATA ANALYSIS: FREQUENC
- Page 530 and 531: EXPLORATORY DATA ANALYSIS: FREQUENC
- Page 532 and 533: EXPLORATORY DATA ANALYSIS: FREQUENC
- Page 534 and 535: STATISTICAL SIGNIFICANCE 515 alongt
- Page 536 and 537: STATISTICAL SIGNIFICANCE 517 hands
- Page 538 and 539: HYPOTHESIS TESTING 519 selection fr
- Page 540 and 541: EFFECT SIZE 521 differential measur
- Page 542 and 543: EFFECT SIZE 523 Box 24.17 The Leven
- Page 544 and 545: THE CHI-SQUARE TEST 525 The Effect
- Page 546 and 547: DEGREES OF FREEDOM 527 Box 24.22 A2
- Page 548 and 549: MEASURING ASSOCIATION 529 Box 24.23
- Page 550 and 551: MEASURING ASSOCIATION 531 found and
- Page 552 and 553: MEASURING ASSOCIATION 533 Box 24.26
- Page 554 and 555: MEASURING ASSOCIATION 535 Many usef
- Page 556 and 557: REGRESSION ANALYSIS 537 we know or
COMPUTER USAGE IN CONTENT ANALYSIS 489<br />
Diction. Some of these are reviewed by Prein et al.<br />
(1995: 190–209). These do not actually perform<br />
the analysis (in contrast to packages for quantitative<br />
data analysis) but facilitate and assist it. As<br />
Kelle (2004: 277) remarks, they do not analyse<br />
text so much as organize and structure text for<br />
subsequent analysis.<br />
These programs have the attraction of coping<br />
with large quantities of text-based material<br />
rapidly and without any risk of human error<br />
in computation and retrieval, and releasing<br />
researchers from some mechanical tasks. With<br />
respect to words, phrases, codes, nodes and<br />
categories they can:<br />
<br />
<br />
<br />
<br />
<br />
<br />
<br />
<br />
<br />
<br />
search for and return text, codes, nodes and<br />
categories<br />
filter text<br />
return counts<br />
present the grouped data according to the<br />
selection criterion desired, both within and<br />
across texts<br />
perform the qualitative equivalent of statistical<br />
analyses, such as:<br />
Boolean searches (intersections of text<br />
which have been coded by more than one<br />
code or node, using ‘and’, ‘not’ and ‘or’;<br />
lo<strong>ok</strong>ing for overlaps and co-occurrences)<br />
proximity searches (lo<strong>ok</strong>ing at clustering of<br />
data and related contextual data either side<br />
of a node or code)<br />
restrictions, trees, cross-tabs (including and<br />
excluding documents for searching, lo<strong>ok</strong>ing<br />
for codes subsumed by a particular node,<br />
and lo<strong>ok</strong>ing for nodes which subsume<br />
others)<br />
construct dendrograms (tree structures) of<br />
related nodes and codes<br />
present data in sequences and locate the text<br />
in surrounding material in order to provide the<br />
necessary context<br />
select text on combined criteria (e.g. joint<br />
occurrences, collocations)<br />
enable analyses of similarities, differences and<br />
relationships between texts and passages of text<br />
annotate text and enable memos to be written<br />
about text.<br />
Additionally, dictionaries and concordances of<br />
terms can be employed to facilitate coding,<br />
searching, retrieval and presentation.<br />
Since the rules for coding and categories are<br />
public and rule-governed, computer analysis can<br />
be particularly useful for searching, retrieving and<br />
grouping text, both in terms of specific words and in<br />
terms of words with similar meanings. Single words<br />
and word counts can overlo<strong>ok</strong> the importance<br />
of context. Hence computer software packages<br />
have been developed that lo<strong>ok</strong> at Key-Words-In-<br />
Context. Most software packages have advanced<br />
functions for memoing, i.e. writing commentaries<br />
to accompany text that are not part of the original<br />
text but which may or may not be marked<br />
as incorporated material into the textual analysis.<br />
Additionally many software packages include<br />
an annotation function, which lets the researcher<br />
annotate and append text, and the annotation is<br />
kept in the text but marked as an annotation.<br />
Computers do not do away with ‘the human<br />
touch’, as humans are still needed to decide and<br />
generate the codes and categories, to verify and<br />
interpret the data. Similarly ‘there are strict limits<br />
to algorithmic interpretations of texts’ (Kelle 2004:<br />
277), as texts contain more than that which can<br />
be examined mechanically. Further, Kelle (2004:<br />
283) suggests that there may be problems where<br />
assumptions behind the software may not accord<br />
with those of the researchers or correspond to<br />
the researcher’s purposes, and that the software<br />
does not enable the range and richness of analytic<br />
techniques that are associated with qualitative<br />
research. Kelle (2004) argues that software may<br />
be more closely aligned to the technique of<br />
grounded theory than to other techniques (e.g.<br />
hermeneutics, discourse analysis) (Coffey et al.<br />
1996), that it may drive the analysis rather than<br />
vice versa (Fielding and Lee 1998), and that it has<br />
a preoccupation with coding categories (Seidel<br />
and Kelle 1995). One could also argue that<br />
software does not give the same added value that<br />
one finds in quantitative data analysis, in that the<br />
textual input is a highly laborious process and that<br />
it does not perform the analysis but only supports<br />
the researcher doing the analysis by organizing<br />
data and recording codes and nodes etc.<br />
Chapter 23